Search Results

Search found 1359 results on 55 pages for 'uploading'.

Page 32/55 | < Previous Page | 28 29 30 31 32 33 34 35 36 37 38 39  | Next Page >

  • Are some .ogv/.ogv files possible to be corrupted, to make it not possible to loop?

    - by C15R
    I discovered by trial and error that my one .ogv video will not loop, but if I interchange a random .ogv video from wikipedia that one does loop, and they are both called .ogv. Why is this? Is the not-looping file corrupted? EDIT! I just had a breakthrough! I discovered if I view the webpage using my computer's version of the file (not-uploaded version), then it does loop! (I open index.html on my computer, and it is a relative URL) But! The version which is uploaded onto my web server does not loop. (I changed the relative URL to an absolute URL which points to the file on my server. I tried re-uploading it, but still does not loop. If someone has an answer to why this is and how I can fix it, I will be very grateful!

    Read the article

  • Ownership/permissions of uploaded files

    - by Cudos
    Hello. I want to find out if I am on the right track. My script uploads files to the directory "images". The directory has this setup: owner/group = www-data Permissions = 700 Questions: Is this a good way to secure the directory from a hacker uploading files? Will the hacker be able to upload the files directly to the directory? Note: I have a bunch of other security measures in my upload script + an .htaccess script in the directory that disables script Execution. I just what to know if the permissions on the directory is sensible. I run apache 2.2

    Read the article

  • Odd FTP ASCII/BINARY transfer behavior after migrating to a new server

    - by Incognita Mundi
    I recently got a dedicated server and after migration to the new machine I started noticing problems with file transfers. My FTP client, despite being set to auto, keeps uploading php files in binary mode and the content of those files is messed up. Since I mostly upload files of different kinds it would be annoying to change from binary to ASCII every single time, beside, I never had such problems. What could be the cause of this behavior? My dedicated server runs CENTOS 6.4 and the ftp server is Pure-FTPd. I tried different FTP clients and they all have the same problem so I assume is soem misconfiguration on the server side. Thanks

    Read the article

  • Ubuntu 12.04 issue installing Webmin modules (file does not exist)

    - by Arthur
    I'm running ubuntu 12.04. On this machine I have Webmin 1.580 installed. When trying to add a new module to Webmin I receive an error: "Failed to install module from : File does not exist" This happens when i go to Webmin Configuration Webmin Modules Select "Third party module from" Select "OpenVPN admin tool" and clicking on install module After I selected "OpenVPN admin tool" this url is being added: http://www.openit.it/downloads/OpenVPNadmin/openvpn-2.5.wbm.gz This URL resolves to a correct file. When uploading the file directly, the same error occurs. In the log file: /var/webmin/webmin.log nothing special is reported. How can I fix this issue? P.s. I run this server in CLI mode

    Read the article

  • Upload from URL to FTP server

    - by Bibhas
    Ok, I have a tar.gz file somewhere in a web server. The link looks like http://abcd.com/abcd.tar.gz .. And I have an FTP server running somewhere. Now, to upload the file to the FTP server, Typically I need to download it from the web server and then upload it again to the FTP server. But I'm wondering if there is anyway, I can directly transfer the file to the FTP server over the web. Not by downloading and uploading again. Any help?

    Read the article

  • Auto-scaling EC2 Servers and Updating Code

    - by jstats
    We've come to the point where we need to set up autoscaling for our web server and I'm unsure how to go about the process of scaling servers and updating the the existing code without remaking a new AMI and changing the autoscale config to use it. I've read a bit about people bundling the new code and uploading it to s3 and having new servers grab the bundle on boot up but that doesn't seem all that pleasant either. Currently the web app's files live in a git repo, and when we update the code, we push it to github, ssh into the web app and run a hook to bring down the latest code. So I was thinking that another option could be to just run that hook on an hourly or daily cron task. Unfortunately that doesn't cover everything type of update (for example new blog posts' images and such which aren't included in the git repo) but it's something. Could anyone provide some advice on what a common solution is or anything as to why my proposed solution is a bad idea? Thanks all

    Read the article

  • How do I use a URL path instead of a file path in an Open File dialog in Mac OSX or ChromiumOS?

    - by Chris
    In Windows 7 (and perhaps earlier), the default "Open File" dialog box allows you to type a full URL into the "File name" section as if it were a file path, e.g. "http://www.example.com/pic.gif" instead of "C:/windows/pictures/pic.gif". When uploading a file to a website on the client side - say, an image - this allows the client to upload a picture located on a server accessible via the URL instead of downloading the image, saving it locally, then referencing the local image in the "Open File" dialog. It's a great option for Windows users. I have three separate questions: What is this procedure formally called? How do I describe this succinctly so that my searches for more information are fruitful? Can something similar be done in Mac OSX, Chromium OS, or a Linux environment? If so, how? Thanks!

    Read the article

  • How do I move an Amazon micro instance to a small instance?

    - by Navetz
    I want to move my instance to a micro instance to a small instance but when I try to launch a new AMI based on my Micro instance AMI it only gives me the option for 64 bit instances. My initial ami is based off an ubuntu 10.04 image. Is it not possible to move between 64 bit and 32 bit instance? Would it be possible to use a load balancer to have a 32bit instance and a 64bit instance work together? I have a website/web app that I will be uploading huge volumes of data to. I will be starting with 65gigs of images and then moving up to 100+ gigs of images. I am not sure which instance type would be best for this. I was going to use a load balancer and auto scaling to increase the number of instance when the load is high. Also when using a load balancer, does one of the AMI instance become the primary image and the rest act as clones of it?

    Read the article

  • WordPress and EverNote integration, any existing solutions?

    - by JXITC
    Actually the tag and category system in WordPress and EverNote are very similar, almost exactly the same. We can create a notebook/notebook group in EverNote, its counterpart in Wordpress is Category/Sub Category We can mutiple tag any note in EverNote, similarly we can tag any post in Wordpress as well! Since the two are so similar to each other in categorizing the article, I am wondering whether there is a existing solution to automatically convert between the two. Like one click to post my EverNote note to WordPress under the same category and same tag. Or pull down my WordPress article to my EverNote account under the same category and same tag. Preferably it can also handle image\video uploading issues as well. :) I saw a similar questions asked years ago: Posting from evernote to wordpress So I was wondering is there any new update for this idea right now?

    Read the article

  • One way sync with either Dropbox or Google drive on linux and windows

    - by sup
    I could not google this one. I would like to use Dropbox or Google drive only as backup, so I would like to ensure I only upload to those services and never download (unless done manually via the web interface). There would be several of us uploading stuff to one account, so we would need to ensure we don't accidentaly delete something on another guy's machine by doing some changes on our local machine. Is there a simple way to do this automatically - i.e. by running a daemon that will upload eveyuthing in a given folder but never download anything? I am on linux, so that complicates things a bit, but I am also interested in windows solutions.

    Read the article

  • Data take on with Drupal 6

    - by Robert MacLean
    We are migrating our current intranet to Drupal 6 and there is a lot of data within the current system which can be classified into: List data, general lists of fields. Common use is phone list of the employees phone numbers. Document repository. Just basically a web version of a file share for documents. I can easily get the data + meta infomation out, but how do I bulk upload the two types of data into Drupal, as uploading the hundred of thousands of items manually is just not acceptable.

    Read the article

  • How can I too many files upload more fast way to Cloud files in Rasckspace?

    - by andy kim
    I have a lot of image files, it's all I want to upload to RackSpace cloud files about a million in a single directory the fastest and most efficient way. but I'm use uploading python-cloudfiles script is very slow and I want to know different ways or python script code. because one by one connection upload is very slow. I think one files tar and uncompress directory is better way. but cloudfiles do not support this way. Who know any other way?

    Read the article

  • files appearing empty or only partially transferred on FTP server.

    - by james
    firstly, apologies if this question has been asked and answered before, but I have had a look through related queries and found nothing identical to this. I have had a website for a few years, and have never had any problem uploading any files. But today, when i went to transfer a new html file onto the server...i did so, the file arrived. so i browsed to the file in my browser to check the page as i always do, and the browser wouldn't acknowledge it. after repeated attempts to transfer it, it finally seemed to go over, but only 1/4 of the file size..4KB out of 16KB...so only the top of the page would be viewable in my browser...ive tried transferring on a number of ftp clients and no love... my expertise on this is limited and i cant really think of the next step, the server isnt full, so...im just stumped. any ideas? any and all feedback is great appreciated.

    Read the article

  • Unable to remove/hide icons from desktop folder

    - by Kyomu
    I wish to hide some windows icons from my desktop folder1, but cannot. When I click delete, it says they will be deleted from my desktop (which they already are), but they remain visible in the desktop folder, which is where I wish them removed. Is there any way I can remove these? (I always use Win+E which really lets me get to all of these easy enough). Offending icons: Libraries, Homegroup, User folder, Computer, Network, Control Panel, and Recycle Bin. For those who might be wondering why, I find myself in this folder a bit to often (including when uploading a file), and find these nothing but clutter, especially as these always default to the top. Thanks for your help! 1 By desktop folder, I mean the folder you would see by going to "My Computer" and clicking "Desktop" on the top left under Favorites. Note: Changing my "Desktop Icon Settings" under "Personalize" does not affect the desktop folder, but rather just the desktop view.

    Read the article

  • How to give write permissions to multiple users?

    - by Daniel Rikowski
    I have a web server and I'm uploading files using an FTP client. Because of that the owner and the group of the file are taken from the user used during the upload. Now I have to make this file writable by the web server (apache/apache). One way would be to just change the owner and the group of the uploaded file to apache/apache, but that way I cannot modify the file using the FTP account. Another way would be to give the file 777 permissions. Both approaches seem not very professional and a little bit risky. Are there any other options? In Windows I can just add another user to the file. Can something similar done with Linux?

    Read the article

  • Simple CMS (separate files for multiple users) [closed]

    - by Pentium100
    I need a simple CMS software (or maybe it's called something else). There are many users (adding new users has to be simple as it will be done manually, no "sign up" link) and those users have to access (just download, no uploading or modification) some files. A file can be shown just to one user or a group of users (a user can belong to multiple groups). Upload a file and tag which users/groups can access it. This can probably be done with FTP, but HTTP would be better. Any software (preferable open source) that can do this without much modification? I can modify PHP code a bit, but I am not a programmer.

    Read the article

  • How does dropbox version/upload large files?

    - by barfoon
    Hey everyone, I have a free dropbox account (2GB), and I was wondering how the versioning of large files works. I have a full backup of all my webfiles that sites @ just over 1GB. After the initial upload of 1GB, everytime it syncs will dropbox figure out the delta of the file, or will it have to upload the entire thing again to version it? It would be cool to always have an up to date version of a large file, but I dont want to kill my bandwidth uploading 1GB everytime. Is this possible? Thanks,

    Read the article

  • Migrating to a new server

    - by KThompson
    I was asked by a client to move a website to a new server. Normally I would make a local copy by downloading it and then uploading it to the new server after making any necessary configurations of course. This particular client has like 1 million images ranging up 1MB and I still haven't found a solution as to how to move these properly. I have tried zipping them on the server first but the file gets too big and it seems to just stop. Anyway, the client told me he can set up a "portal" so I don't have to download and re-upload the files. I have never heard of this no can I find what this is. Any of you know something that fits this description or solution to my image problem? Thanks in advance

    Read the article

  • Ultimate way to use Picasa in a home network

    - by luisfarzati
    I've been trying a lot of approaches but still didn't find any effective solution. I want gigs of photos in a network drive (a IOMega Home Media Network Drive, plugged to my wifi router). I'd like to do 2 things: Do a Picasa import process of all the photos in the drive, making Picasa organize all the files in a year/month folder structure physically. Ideally, the import target directory should be the same network drive, otherwise I should move all the imported files in my local computer back to the drive myself. Share the Picasa database over the network, by uploading it to the network drive. Have me and other members of the family point our Picasas to the network database, and see the photos as well as make changes (tag faces, create logical albums, etc) into it. Is ANY possibility to accomplish this? Or should I be looking for another photo management app, and in that case do you know such one? Thank you!

    Read the article

  • Best Wiki Software For Product Support

    - by Zapnologica
    Good day, I am looking for a wiki system which we are going to use at work for a form of product support. We manufacture multiple devices and now i want to make a wiki which contains all sorts of relevant and helpful information on that product which the users can look at before trying to contact us for support? Now immediately the 1st option that comes to my mind in Media wiki but I dont just want to jump on the band wagon. I thought I would ask around first. It should preferable be free. But obviously if its really worth paying for then thats not the end of the world. And the uploading of content and media is not of much importance as the end users will simply be reading the information which the company has published. Another nice to have but is not critical is if it where to run on asp.net as we have Microsoft server running anyway.

    Read the article

  • Hosing website on own server. What is Hardware requirement and Cost [closed]

    - by KuKu
    As i am about to finish my dream website, i need to host it on some server. I checked Amazon AWS Free Tier, i found it pretty complex. As i made full website in java(JSP + Servlet + mysql + Node.js), its been expensive to host. So i thought that why not to host on own server so that i will be fully dependent on my resources, not on any one else. And i know, in future i need to pay more and more to the other hosting company(because of uploading images and number of incoming users). So my question is, On initial stage what kind of hardware i will required. What can be the cost of that hardware? I already have 12MBPS broadband connection, will it be sufficient? It has static IP Address as well.

    Read the article

  • What are the different file permission codes and what do they mean?

    - by zeckdude
    I am working with a file upload script. I am currently uploading a file and then trying to echo out an anchor linking to that file, but since I used mkdir() with 0700 permissions to upload the file, it won't allow me access to view the file. I am pretty sure the problem I am experiencing is because of the file permission code I used. The problem is I just don't know what all the different file permission codes are and what they mean. Can somebody please list out all the different file permissions and what they each do?

    Read the article

  • Is is better to combine Apache for file manipulation and upload and Nginx for static file serving, or to use one of the two alone

    - by user1032393
    Based on my research, I've read that nginx is best and ideal for serving up static files and images. My application depends heavily on uploading of images and rewriting them, then serving them up. Given that I only have one VPS currently, it has been suggested that I use nginx to serve up the images and website, and reverse proxy to Apache (on the same VPS) to rewrite files with image magick and handle the file uploads. Which would be the best solution, Apache, Nginx, or Apache + Nginx? In terms of best solution, I'm looking at minimal average RAM consumption, while maintaining decent load speed of maybe sub 2 seconds?

    Read the article

  • cannot write to certain directory using CentOS distro

    - by fishcracker
    I have a script that will generate an image then upload it in a certain directory. The script works fine, however uploading it in the CentOS server, it won't work anymore. By debugging, I have found out that this is a permission denied issue. I am forced to chmod -R 777 certain_directory to make it work. I am not convince with this, I can't set the directory ownership to www-data I'm used to when I am using Debian. Any workaround for this?

    Read the article

  • all USB "randomly" disconnect and reconnect

    - by teuneboon
    I'm having a weird issue currently: at seemingly random times(although it seems to happen mostly when USB is used a lot, ie downloading + uploading over my WiFi stick or viewing an HD video from my external hard-drive) all my USB devices disconnect and reconnect after a few seconds. This wouldn't be a HUGE problem if I didn't have to rely on my USB for my internet, because now every time this happens my internet, and with it stuff like remote desktop, crashes. I'm running Windows 8 Pro 64 bit with a Gigabyte 990FXA-D3 motherboard. Now the weird thing: I had the same problem with my previous pc, which had a different motherboard(I think it was an Asus M4A785TD-V EVO). I googled a bit and didn't really find anything about this, only found out I could check if something shows up in the Event Viewer, and I get this error there when the USB crashes: "A timeout occurred while waiting for the EHCI host controller Interrupt on Async Advance Doorbell response." If you require any extra information I'm happy to provide it.

    Read the article

< Previous Page | 28 29 30 31 32 33 34 35 36 37 38 39  | Next Page >