Search Results

Search found 82718 results on 3309 pages for 'large file download'.

Page 253/3309 | < Previous Page | 249 250 251 252 253 254 255 256 257 258 259 260  | Next Page >

  • rsync command deletion error "IO error encountered -- skipping file deletion"

    - by Jam88
    I use rsync command to take backup of files from one of my ubuntu server to another ubuntu machine. Backup server trigger a script that use rysnc command. Here is the command I use rsync -rltvh --partial --stats --exclude=.beagle/ --exclude=.* --delete-after root@live_server:/home/ /home/live_server_backup/home /tmp/logfile.log 2&1 live_server is ssh-able without password. So it works. Now problem is with --delete-after option After all file synced .At the end I can see deletion procedure skipped.logfile error is like IO error encountered -- skipping file deletion When i tried to find log there were some error while file sync rsync: send_files failed to open "/home/xyz/Desktop/PPT_session_1_context.pdf": Permission denied (13) So my understanding is as rsync could not read all the files from target for safety reason it is skipping the file deletion. Is there any way to make --delete-after work even if there is some permission error? I do not want to use force deletion as it will be dangerous in some situation.

    Read the article

  • Decompress a PSC File

    - by Kevin Laity
    I need to restore a database that was backed up using navicat's compressed backup feature. But on the restore, there's an error coming up on one of the tables. I can't look at the SQL statements in the file directly because they're compressed, and I don't know what compression format it's in. Does anyone know of a way to decompress this file into a regular text file?

    Read the article

  • Bash delete file when variable = x

    - by twigg
    I'm creating a bash script which reboots the system at each reboot it adds a new line to a text file, I then read the text file before each reboot. Once the variable holding the number of lines reaches say 10 I want the script to delete the text file (at which point on the next reboot it will see the file isn't there, brake the loop and promote the user to start again). I tried this: exec < text.txt nol=0 while read line do nol=`expr $nol + 1` done reboot_count=10 if ["$nol" == "$reboot_count"]; then rm text.txt fi but this doesn't seem to be working, all help is appreciated :)

    Read the article

  • Creating an interactive "movie" from a Microsoft PowerPoint 2003 presentation (.ppt file) containing animation

    - by Andrew
    I have created a Microsoft PowerPoint 2003 presentation (a .ppt file) on Windows XP that contains a lot of animation, using PowerPoint's Custom Animation tools. I would like to share this presentation -- animations and all -- with a colleague who has a Mac computer and does not have PowerPoint or Keynote. Is it possible to make a "movie" from my .ppt file that contains a lot of animation? I recall that another colleague once made me a QuickTime "movie" from a PowerPoint file, I think (although I do not remember whether he made it using Windows or Mac). On this "movie," clicking advanced the animation/slide, just as clicking would do in an actual PowerPoint presentation. In that sense, it was not really strictly a movie, since I could interact with it by clicking. Do you know how to make such an interactive "movie" from a PowerPoint 2003 presentation (a .ppt file)? Thank you for your time!

    Read the article

  • Centralized sudo sudoers file?

    - by Stefan Thyberg
    I am the admin of several different servers and currently there is a different sudoers file on each one. This is getting slightly out of hand as quite often I need to give someone permissions to do something with sudo but it only gets done on one server. Is there an easy way of editing the sudoers file just on my central server and then distributing it by SFTP or something like that to the other servers in an easy way? Mostly wondering how other sysadmins solve this problem, since the sudoers file doesn't seem to be remotely accessible with NIS, for example. Operating system is SUSE Linux Enterprise Server 11 64-bit, but it shouldn't matter. EDIT: Every machine will, for now, have the same sudoers file. EDIT2: The accepted answer's comment was the closest to what I actually went ahead and did. I am right now using an SVN-supported puppet-installation and after a few headaches, it's working very well.

    Read the article

  • IIS 7 Website Bindings www.domain.com & domain.com

    - by asn187
    I am currently setting up a domain name on windows server 2008r2 iis7 In the local hosts file I have mapped Domain.com and www.Domain.com to the localhost The iis7 website has the bindings for Domain.com and www.Domain.com However, from the internet the website only loads up when browsing to www.Domain.com Is there something I am missing to make it work on Domain.com

    Read the article

  • Excel file growing huge (>150 MB)

    - by Josh
    There is one particular Excel file that is used by a number of employees at my company. It is edited from both Excel 2003 and 2007, with the "Sharing" feature turned on to allow multiple writers at once. The file has a decent amount of data on several sheets with some basic formatting, and used to be about 6MB, which seems reasonable for its content. But after a few weeks of editing, the file grew to 10, then 20 MB, and eventually skyrocketed to more than 150 MB, even though it still has about the same amount of data as before. It now takes 5-10 minutes to open it, and that much time again to save it. The first time this happened, I copied the content of each sheet into a new, blank workbook, and saved the new workbook; this brought it back down to about 6MB. Now, it has blown up again. The workbook uses the "Data Validation" feature to limit the values in certain columns to the contents of a few named ranges. Copying all the data into a new workbook means re-setting up all the data validation, which is a pain and not something that we want to do every month. As a troubleshooting step, I tried saving the file in "XML Spreadsheet 2003" format, hoping to get some insight into what was being stored. Sure enough, the file was almost a gig, and almost all of the 10 million lines look like this: <NamedCell ss:Name="Z_21D5114F_E50C_46AC_AA4F_C3FF540C717F_.wvu.FilterData"/> <NamedCell ss:Name="Z_1EE2BA5E_3011_4F9A_8ACD_E58835250FC4_.wvu.FilterData"/> <NamedCell ss:Name="Z_1E3BDCEA_6A72_4ECC_BF4F_7B03CC66181E_.wvu.FilterData"/> I've seen a few VBScripts online to manage and enumerate named cells that are hidden in Excel's built-in interface, though I wonder how they'd handle my 10 million named cells. What I really need, though, is an understanding of why this keeps happening. What actions in excel could be causing this?

    Read the article

  • IE 8 crashes when opening an html file from an autorun CD

    - by dave4351
    On Windows 7, when I insert a CD with a very simple HTML page referenced in the autorun.inf file, Internet Explorer 8 opens and then immediately crashes, with the error message "Internet Explorer has stopped working" and then checks for a solution. Then it says "A problem caused the program to stop working correctly. Windows will close the program and notify you if a solution is available." On Windows Vista, IE 8 just opens and then closes silently, with no error messages or help of any kind. I have since discovered this is due to IE 8's "protected mode" and turning this off fixes the issue so that the HTML file on the CD launches normally. What I'd like to know: is there any way to get around this, so that I can launch a simple HTML file in the user's default browser, using an autorun.inf file on a CD, without first having my users open IE 8 and turn off protected mode? Perhaps the most bizarre thing is that if IE is already open when you insert the CD, the HTML page just opens--no crash. It only crashes if IE 8 happens to be closed when you insert the CD. And so a possible workaround may be to get the default browser to launch and then open the HTML file. All suggestions welcome.

    Read the article

  • HOSTS File Edit in Windows 7 Not Effective - Pinging URL Still Shows Original IP Address

    - by Sootah
    I've edited my HOSTS file on my Windows 7 Ultimate PC to re-route a couple of URLs so that they point to 127.0.0.1, but after saving the file (and re-opening to verify the changes were written) and pinging them they still reply with the actual IP instead of being redirected to 127.0.0.1 as they should be. At least, that's how it worked in XP, Vista, etc. I even went so far as to restart my DNS service on the machine via services.msc; but no dice. So - I would imagine that Windows 7 keeps the HOSTS file there for legacy purposes and doesn't actually use it anymore. Is there a way to make W7 pay attention to the HOSTS file? In the event that you can't do that, where would I go to edit where these URLs point to? Thanks in advance! -Sootah

    Read the article

  • Iomega Home Media Network Hard Drive: Filesystem on the disk?

    - by JJarava
    Hi all! I've got to deal with a malfunctioning "Iomega Home Media Network Hard Drive", and I was wondering if anybody knew what file system format does Iomega use on the disk? I've been trying to find the answer online, but i've got nowhere, and checking an obviously malfucntioning unit is not going to give me any assurance. Thanks a lot

    Read the article

  • Timing issue with autohotscript, fails to dump or open destination file

    - by learnerforever
    I've created a autohotscript to quickly dump selected text into my jot file on the Desktop and I think I'm facing a timing error. The script works like thus: Select text when reading a text file, browsing internet, reading PDF, etc. Hit Ctrl + J Contents of selected text is dumped into my jot file. When I press Ctrl + J very quickly, it sometimes doesn't come up in my jot file and sometimes when I keep pressing Ctrl + J for a long time, many instances of the text appear. Could somebody please point out what's wrong with this script and how I can improve it. ^j:: Clipboard := "" ; clear Send, ^c ; simulate Ctrl+C (=selection in clipboard) selection = %Clipboard% ; save the content of the clipboard FileAppend, `n%selection%`n,C:\Users\jagrati\Desktop\jots.txt return

    Read the article

  • JDBCRealm can't find sqlite file

    - by Tom A
    My authentication fails with java.sql.SQLException: no such table: credentials where credentials is the name of the user/password table. I have checked the db file and the table is there. I think you also get this error when sqlite jdbc can't even find the file. I am specifying my realm in a META-INF/context.xml file. Is there any trick to getting the path right? I have tried just about everything I can think of.

    Read the article

  • RAID 50 24Port Fast Writes Slow Reads - Ubuntu

    - by James
    What is going on here?! I am baffled. serveradmin@FILESERVER:/Volumes/MercuryInternal/test$ sudo dd if=/dev/zero of=/Volumes/MercuryInternal/test/test.fs bs=4096k count=10000 10000+0 records in 10000+0 records out 41943040000 bytes (42 GB) copied, 57.0948 s, 735 MB/s serveradmin@FILESERVER:/Volumes/MercuryInternal/test$ sudo dd if=/Volumes/MercuryInternal/test/test.fs of=/dev/null bs=4096k count=10000 10000+0 records in 10000+0 records out 41943040000 bytes (42 GB) copied, 116.189 s, 361 MB/s OF NOTE: My RAID50 is 3 sets of 8 disks. - This might not be the best config for SPEED. OS: Ubuntu 12.04.1 x64 Hardware Raid: RocketRaid 2782 - 24 Port Controller HardDriveType: Seagate Barracuda ES.2 1TB Drivers: v1.1 Open Source Linux Drivers. So 24 x 1TB drives, partitioned using parted. Filesystem is ext4. I/O scheduler WAS noop but have changed it to deadline with no seemingly performance benefit/cost. serveradmin@FILESERVER:/Volumes/MercuryInternal/test$ sudo gdisk -l /dev/sdb GPT fdisk (gdisk) version 0.8.1 Partition table scan: MBR: protective BSD: not present APM: not present GPT: present Found valid GPT with protective MBR; using GPT. Disk /dev/sdb: 41020686336 sectors, 19.1 TiB Logical sector size: 512 bytes Disk identifier (GUID): 95045EC6-6EAF-4072-9969-AC46A32E38C8 Partition table holds up to 128 entries First usable sector is 34, last usable sector is 41020686302 Partitions will be aligned on 2048-sector boundaries Total free space is 5062589 sectors (2.4 GiB) Number Start (sector) End (sector) Size Code Name 1 2048 41015625727 19.1 TiB 0700 primary To me this should be working fine. I can't think of anything that would be causing this other then fundamental driver errors? I can't seem to get much/if any higher then the 361MB a second, is this hitting the "SATA2" link speed, which it shouldn't given it is a PCIe2.0 card. Or maybe some cacheing quirk - I do have Write Back enabled. Does anyone have any suggestions? Tests for me to perform? Or if you require more information, I am happy to provide it! This is a video fileserver for editing machines, so we have a preference for FAST reads over writes. I was just expected more from RAID 50 and 24 drives together... EDIT: (hdparm results) serveradmin@FILESERVER:/Volumes/MercuryInternal$ sudo hdparm -Tt /dev/sdb /dev/sdb: Timing cached reads: 17458 MB in 2.00 seconds = 8735.50 MB/sec Timing buffered disk reads: 884 MB in 3.00 seconds = 294.32 MB/sec EDIT2: (config details) Also, I am using a RAID block size of 256K. I was told a larger block size is better for larger (in my case large video) files. EDIT3: (Bonnie++ Results. Would love some guidance with this!)

    Read the article

  • Image file can't be opened by windows store app if the file is in the Ext4's external hard disk

    - by ? ?
    My question is described below: I have two disks, external hard disk and removable disk. They are formatted with EXT4. I put the same image in the disks, and use Ext2Fsd to mount the disks. I open the file in win8, and it will default open by the windows store app : photos. It will be success to open if the file is in the Ext4's removable disk. It will be failed to open if the file is in the Ext4's external hard disk. I don't know why it will be failed to open if the file is in the Ext4's external hard disk. Does anyone have similar experiences?

    Read the article

  • not a valid iso file error

    - by user23950
    I'm trying this windows 7 usb/dvd download tool that I discovered from ars technica: http://arstechnica.com/business/news/2009/12/-the-usb-flash-drive.ars But I don't know why my windows 7 iso file does not work and its saying that its not a valid iso file. What do I do, do you know of other tools that can burn / extract iso file contents to a flash drive? Except the tool that is always showing up when you search in google, the command prompt one: http://arstechnica.com/business/news/2009/12/-the-usb-flash-drive.ars/2 A free tool for windows xp or 7

    Read the article

  • File downloaded from IIS6/Win2003 server to a Mac (and not PCs) is incredibly slow

    - by Simon Swords
    We have a test zip file on our customers server that we host for him that when downloading to a Mac is incredibly slow. On a Mac - trying the download via Safari 5.0.3 and Chrome 8.0.552.231 results in a quick burst of normal download speed then plummets to almost no speed at all after 1 or 2 meg (between 1 and 5 Kb/s - yes, KiloBITS per second! According to the network monitor). Downloading via Windows was fine and speedy. Tested via; IE7 7.0.5730.13 and Chrome Portable 8.0.552.224 On Windows XP Pro, and; IE8 8.0.7600.16385 in a Windows 7 virtual machine running via VirtualBox 4.0.0 r69151 on the same Mac mention above Google hasn't helped us out on this occasion, possibly because the search terms I'm having to use are quite generic. Has anybody ever experienced this and if so how do we fix it? Thanks in advance

    Read the article

  • How does FTP servers handle rename to existing file

    - by Kristian
    How does FTP servers normaly handle rename to a file that aready exist? That is, will the existing to-file be overwritten or will I get an error? Is there some specification about how FTP servers should handle this, or is it OS or implementation dependent? vsftpd on Linux seems to overwrite, with FileZilla on Vista I get 550 file exists...

    Read the article

  • XAMPP pointing a file outside root folder

    - by Ravi
    I am using XAMPP for first time in Mac. Running out problems accessing other than root folder(htdocs).when I am placing my web application inside htdocs with default httpd.conf file it works when I try to point my web application url in httpd.conf it throws error I am aware that to modify the root folder I need to do changes to my XAMPP/etc/httpd.conf file With Default XAMPP MAC Settings, I am trying to change Server root,Document root and Directory in XAMPP/etc/httpd.conf file the following ServerRoot "/Users/ravi/Documents/Development/Backbone/backboneboilerplate" DocumentRoot "/Users/ravi/Documents/Development/Backbone/backboneboilerplate" <Directory /> Options FollowSymLinks AllowOverride All Order deny,allow Deny from all </Directory> <Directory "/Users/ravi/Documents/Development/Backbone/backboneboilerplate"> Options Indexes FollowSymLinks AllowOverride All Order allow,deny Allow from all </Directory> its throwing error when trying to start XAMPP httpd: Syntax error on line 54 of /Applications/XAMPP/xamppfiles/etc/httpd.conf: Cannot load /Users/ravi/Documents/Development/Backbone/backboneboilerplate/modules/mod_authn_file.so into server: cannot create object file image or add library

    Read the article

  • Create a zip file that doesn't create a folder when opened, Mac OS X

    - by MZimmerman6
    I am wondering if there is any way of easily creating a zip file that contains a bunch of files. However, when I double-click to open this zip file, I do not want it to create a subdirectory with the same name as the zip file, I just want all of the files to be extracted into the directory where the zip files is stored. Is there a certain flag I can add in terminal to do this, or any program that does this. Thanks for your help!

    Read the article

  • How to recover MySQL database from .mysql file?

    - by Brayn
    We had some problems with our MySQL server and somehow all I've got is a database.mysql file for the database I want to restore. I've done a bit of googleing but I didn't find anything about how should I handle this type of file. It's worth mentioning that the server was running Plesk and the database wasn't using InnoDB. Edit: I've forgot to mention that I don't know what application created the .mysql file and that it's in binary format. Thanks,

    Read the article

< Previous Page | 249 250 251 252 253 254 255 256 257 258 259 260  | Next Page >