Search Results

Search found 12625 results on 505 pages for 'ajax upload'.

Page 220/505 | < Previous Page | 216 217 218 219 220 221 222 223 224 225 226 227  | Next Page >

  • How can I upload and parse a vcard using ruby on rails?

    - by Angela
    I would like to be able to upload a vcard and have it parsed into my application's Model. I am thinking that Paperclip would be a way to upload, but I have no idea how to do the parse. I did a search on google, the only library that seems related is vpim, with mixed commentary. Wanted to see what brilliant ideas might be out there to solve this...thanks.

    Read the article

  • is it possible to continue a file upload after a browser has been quit?

    - by sprugman
    If a user starts uploading a file (in a web app), and they close their browser before the upload completes, what happens? My guess would be that the upload gets abandoned, and the server has some garbage cleaning mechanism that kicks in eventually, but I'm not sure. Any difference between servers or server languages? Any difference between quitting the browser and just closing a window?

    Read the article

  • How do I upload a file when I know the client's filepath?

    - by GJTorikian
    I'm interested in the way Sonic Living detects your iTunes library XML file. It prompts the user to approve its actions, then automatically uploads the XML file based on the client OS (and assuming the default installation path of iTunes). How the heck does it do that? None of the upload libraries I've looked at have this feature. They all assume that a user will click "Select file..." then browse around for the right thing to upload.

    Read the article

  • Convert raw IMAP server data into local folders, then upload partial dataset to new IMAP server?

    - by Manca Weeks
    I am transitioning a company with about 30 IMAP accounts, loaded with data (about 77GB total), to a new email host. The majority of the data will be converted into a local archive and distributed to the company computers as a static reference data set. The server side folders the users absolutely cannot do without being on the server will be uploaded back to the new server. I used Mac OS X Mail (Snow Leopard 10.6.6) to download the content. I notice some messages have the name [xxx].partial.emlx, which leads me to believe they have not been downloaded all the way. I have root access to the mail server data and could download the IMAP server data via FTP. I am not sure what utility to use to convert that data to local Mail.app mailboxes. Furthermore, I would appreciate any input on the best way to upload a portion of the data to the new server (GoDaddy), preserving the original dates of the messages. edit OK - forget the raw server data. I found a script that apparently does pretty good archiving IMAP folders to local mbx files. My main quest now is to batch upload a mailbox hierarchy to the new IMAP server without having to start-stop and deal with similar issues. Anyone know of a utility (hopefully for OS X, but if not, I'll fire up my XP virtual system...) that would be capable of this? Thanks, M

    Read the article

  • VSFTPD - FTP over TLS - Upload stops after exactly 82k?

    - by Redsandro
    I installed a VSFTP daemon on a CentOS server, using a RSA certificate for logging in using explicit TLS. Now, I cannot upload more than 82k. With files under that limit, there is no problem. The FTP works like a charm. But as soon as a file reaches 82k with FileZilla (81,952 bytes to be exact), the transfer will stop, and the FTP client hangs until time out is reached. FTP client console: 15:10:21 Command: STOR jquery-1.7.2.min.js 15:10:21 Response: 150 Ok to send data. 15:11:21 Error: Connection timed out 15:11:21 Error: File transfer failed after transferring 82 KB in 60 seconds /var/log/vsftpd.log FTP command: Client "x.x.x.x", "STOR jquery-1.7.2.min.js" FTP response: Client "x.x.x.x", "150 Ok to send data." OK UPLOAD: Client "x.x.x.x", "jquery-1.7.2.min.js", 81952 bytes, 1.32Kbyte/sec FTP response: Client "x.x.x.x", "226 File receive OK." // NOT okay, file is bigger // No mention of error here I cannot find relevant info about this problem, apart from a possible problem with trans_chunk_size (not mentioned in default config), but I tried different sizes and it has no impact on the problem. trans_chunk_size=4096 trans_chunk_size=8192 trans_chunk_size=9999 Ofcourse, after every configuration change, I restarted the server: /etc/init.d/vsftpd restart What else can cause this? It's not the latest version, but it's the latest update within the repositories that has been deemed fit for enterprise usage: Package info: $ yum info vsftpd Loaded plugins: fastestmirror Installed Packages Name : vsftpd Arch : x86_64 Version : 2.0.5 Release : 24.el5_8.1 Size : 286 k Repo : installed Summary : vsftpd - Very Secure Ftp Daemon URL : http://vsftpd.beasts.org/ License : GPL Description: vsftpd is a Very Secure FTP daemon. It was written completely from scratch.

    Read the article

  • File upload folder permission fastCGI - How to make it writeable?

    - by user6595
    I am using centos 5.7 with cPanel WHM running fastcgi/suEXEC I am trying to make a particular folder writable to allow a script to upload files but seem to be having problems. The folder (and all recursive folders) I want to be writable is: /home/mydomain/public_html/uploads And I want only scripts run by the user "songbanc" to be able to write to this directory. I have tried the following: chown -R songbanc /home/mydomain/public_html/uploads chmod -R 755 /home/mydomain/public_html/uploads But it still doesn't seem to work. The script will only upload files if I set the permissions manually via FTP client to 777. I assume I am misunderstanding how to set permission for users with fastcgi and hopefully someone can help me. Thanks in advance EDIT: Running getfacl on one of the scripts or folders gives the following: # file: home/mydomain/public_html/ripples/1.jpg # owner: songbanc # group: songbanc So it appears that the owner is correct? I'm now totally confused! EDIT 2: The plot thickens... lsattr and chattr are returning Inappropriate ioctl for device While reading flags on...

    Read the article

  • ASP.Net: Writing chunks of file..HTTP File upload with resume..

    - by Manish
    Please refer to question: http://stackoverflow.com/questions/2062852/resume-in-upload-file-control Now to find a solution for the above question, I want to work on it and develop a user control which can resume a HTTP File Upload process. For this, I'm going to create a temporary file on server until the upload is complete. Once the uploading is done completely then will just save it to the desired location. The procedure I have thought of is: Create a temporary file with a unique name (may be GUID) on the server. Read a chunk of file and append it to this temp file on the server. Continue step 1 to 3 until EOF is reached. Now if the connection is lost or user clicks on pause button, the writing of file is stopped and the temporary file is not deleted. Clicking on resume again or uploading the same file again will check on the server if the file exists and user whether to resume or to overwrite. (Not sure how to check if it's the same file. Also, how to step sending the chunks from client to server.) Clicking on resume will start from where it is required to be uploaded and will append that to the file on the server. (Again not sure how to do this.) Questions: Are these steps correct to achieve the goal? Or need some modifications? I'm not sure how to implement all these steps. :-( Need ideas, links... Any help appreciated...

    Read the article

  • how to display image in repeater after AJAX AsyncFileUpload ?

    - by Gordon Shamway
    Hello i am using the AsyncFileUpload from the ajaxtoolkit, upter the upload is complete , i want to display all the images that were uploaded (and their URL is inserted into the DB) with my repeater and binding method. but its not working! the images are uploaded, the DB insert works file, but i cant see the images after the repeater is dinded. (i have to mention that if i go step by step, i do see that the ITEMBOUND is being binded and executed) <ajax:TabContainer runat="server" ID="testTab"> <ajax:TabPanel runat="server" ID="testTabContainer"> <ContentTemplate> <ajax:AsyncFileUpload ID="uploadAsyncImage" runat="server" OnUploadedComplete="InsertToTemp" ThrobberID="Throbber"/> <asp:Label ID="Throbber" runat="server" Style="display: none"> <img src="Images/indicator.gif" align="absmiddle" alt="loading" /> </asp:Label> <br /> <asp:UpdatePanel runat="server" ID="pnlImages" UpdateMode="Conditional" ChildrenAsTriggers="true"> <ContentTemplate> <asp:Repeater runat="server" ID="rptImages" > <ItemTemplate> <a href="<%# DataBinder.Eval(Container.DataItem, "PicURL")%>" rel="lightbox"> <asp:Image ImageUrl='<%# DataBinder.Eval(Container.DataItem, "PicURL")%>' Width="50" Height="50" runat="server" ID="testPic" /> </a> </ItemTemplate> </asp:Repeater> </ContentTemplate> </asp:UpdatePanel> </ContentTemplate> </ajax:TabPanel> </ajax:TabContainer> </form> also adding my Code Behind, since it might be important or will help you get an idea of what i am trying to do.private void BindRepater() { ProjDataContext db = DbContext.DbContextProvider(); var pic = from p in db.TempProdCats where p.SessionID == Session.SessionID.ToString() select new { PicURL = p.PicURL }; rptImages.DataSource = pic; rptImages.DataBind(); pnlImages.Update(); } protected void rptImages_ItemDataBound(object sender, RepeaterItemEventArgs e) { if (e.Item.ItemType == ListItemType.Item || e.Item.ItemType == ListItemType.AlternatingItem) { Image lblCat = (Image)e.Item.FindControl("testPic"); lblCat.ImageUrl = (string)DataBinder.Eval(e.Item.DataItem, "PicURL"); pnlImages.Update(); } } your help will be highly appreciated , if any of you got any idea how to make it work...

    Read the article

  • Where to upload cgi script on Linux Debian VPS with Plesk ?

    - by Rebol Tutorial
    I'm newbie to Linux. I have uploaded rebol ( a script langage similar to perl from http://www.rebol.com/view-platforms.html ) the linux debian version to /usr/bin on my VPS Where do I upload this script : Code: #!/usr/bin/rebol -c REBOL [Title: "Cgi Test in Rebol"] print "content-type: text/html^/" print ["Date/time is:" now] as I have 2 domain names on this VPS, I can't find where the domain folder are to do so.

    Read the article

  • How to upload stuff to Amazon EC2 Windows instance?

    - by JohnIdol
    I've never used Amazon EC2 - I am thinking to test a few instances for running intensive computation processes I have a requirement for rather than buying real hardware. I am given to understand the thing is quite easy to setup - but I have no clue of how it actually works, in terms of transferring data to my ec2 instances. So the question is how can I upload stuff to my instance? Any help appreciated!

    Read the article

  • how do i upload virtual pc into a wesite using virtual server.

    - by ragu88
    dear friends, my school has given me a project on virtual pc and server. i am supposed to create a website with multiple virtual pcs in it.. students are supposed to go to the website and than to the virtual pc and do their lab experiments there... can i no how m i to do that and how do i upload a virtual pc into the website.. i m using ms virtual server 2005 r2.. Ragu

    Read the article

  • linksys WRT350N dd-wrt: how can i give upload priority to slingbox ?

    - by ufk
    Hi. i have the router WRT350N by Linksys I've installed dd-wrt v2.4 sp1 on it how can i configure it to give upload priority to slingbox ? The Slingbox is a TV streaming device that enables users to remotely view their home's cable. it seems that slingbox streams to amazon ec2 servers and from there to the client so i can't really give priority based on ip-address.

    Read the article

  • Do I need a Gigabit router with a 24Mbps down and 7Mbps upload speed cable modem?

    - by djangofan
    Do I need a Gigabit port capable wireless router with a 24Mbps down and 7Mbps upload speed cable modem? Does anyone know how to calculate this? FYI, I wont be using the wireless connection from my main computer system. My computer will connect via a hard wire into the router (of the wireless variety), which in turn is connected to the cable modem. My research suggests that the 100 Mbps port can easily handle it. Is that true?

    Read the article

  • Speaking at Tech-Ed Europe Next Week

    - by Stephen Walther
    I’m going to Berlin! Next week, I’m giving talks at Tech-Ed Europe on two of my favorite topics: What's New in Microsoft ASP.NET Model-View-Controller ASP.NET Model-View-Controller (MVC) 2 introduces new features to make you more productive when building an ASP.NET MVC application. Templated helpers allow automatically associatiating edit and display elements with data types. Areas provide a means of dividing a large Web application into multiple projects. Data annotations allow attaching metadata attributes on a model to control validation. Microsoft ASP.NET AJAX: Taking AJAX to the Next Level Hear how ASP.NET AJAX 4.0 makes building pure client-side AJAX Web applications even easier, and watch us build an entire data-driven ASP.NET AJAX application from start to finish by taking advantage of only JavaScript, HTML pages, and Windows Communication Foundation (WCF) services. Also learn about new ASP.NET AJAX features including the DataView control, declarative templates, live client-side data binding, WCF, and REST integration.   The conference has sold out, but you can register for the wait list: http://www.microsoft.com/europe/TechEd/

    Read the article

  • How does Ubuntu One work? Files seem to upload but can not be accessed on another system

    - by JQPublic
    I have two computers running 12.04, I sign into my Ubuntu 0ne account on each computer, "Hi JQPublic" both say. Do I understand correctly that placing a file (specifically an .ogg music file ... simply a song) into the "Ubuntu One" file folder (always synced) on one computer, I should be able to open the "Ubuntu One" file folder on the other computer and find said .ogg, right? That is how this is meant to work, yes? Because it's not what's happening. Nothing uploads (though the first computer claims such). Minutes pass, still nothing. A different experiment: I uploaded a file folder containing a song; the folder was available on both computers, but remained decidedly empty. Any suggestions? I'm otherwise LOVING my return to Ubuntu (left at 9.04). I really am inclined to think that I'm missing something HUGE, that this is all a user (me) error, not a bug ... is it me? All help appreciated. Peace, JQPublic

    Read the article

  • Can preventing directory listings in WordPress upload folders cause Google ranking drops when they cause 403 errors in Webmaster Tools?

    - by Kelly
    I recently moved to a new host that blocks crawling to my uploads folders but (hopefully) allows the files in the folder to be crawled. I now show many 403 errors for each folder in the uploads folder in my Webmaster Tools. For example, http://www.rewardcharts4kids.com/wp-content/uploads/2013/07/ shows a 403 error. For example, I can access this file: http://www.rewardcharts4kids.com/wp-content/uploads/2013/07/lunch-box-notes.jpg but I cannot access the folder it is in. My rankings went down after I moved to this host and I am wondering if: this could be the reason. is this how files/folders are supposed to be set up?

    Read the article

  • Concurrency with listeners and upload: is this solution sound?

    - by cbrulak
    I'm working on an Android app. We have listeners for position data and camera capture which is timed on a background thread (not a service). The timer thread dictates when the image is captured and also when the data is cached in a local sqlite db. So, my question is how to properly store the location data as it comes in based on the listeners and pull that data so that the database can be updated as the camera capture is executed. I can't put the location data into the database as it arrives because it is processed more frequently than the camera images and the camera images are what dominates the architecture at the moment. (Location data is supplement). However I need the location data to get a rough location of the where the image is captured. My first thought was to have singleton store the location data. And have a semaphore on the getInstance method so if the database update is happening (after an image is captured) we don't have an error. The location data can wait for the database update or it can be lost for that particular event it doesn't really matter. What are you thoughts? Am I on the right track? (And is this the right sub-site or would this be better on stackoverflow?)

    Read the article

< Previous Page | 216 217 218 219 220 221 222 223 224 225 226 227  | Next Page >