Search Results

Search found 39784 results on 1592 pages for 'ignore files'.

Page 153/1592 | < Previous Page | 149 150 151 152 153 154 155 156 157 158 159 160  | Next Page >

  • md5sum of large files gives different results sometimes

    - by Emanuele
    I have an AMD quad core, 8 gb RAM, 1 SSD EXT2 (2 months old), 2 HDD EXT4, approximately 1 year old. I'm using Ubuntu 10.04 x86-64 and when I compute the md5sum of large files (9 GB) sometimes I get different values than the one stored on a reference file. Upon restarting and switching off the PC then I get the expected results no matter how many times I repeat it. But this is random. I've turn on ECC (the fastest possible settings) and the issue seems to be rarer, but I've run memtest86+ for 6+ hours without a glitch (and with ECC off!). Any idea? Should I update the BIOS of my motherboard (Asus EVO-something...don't remember it now)? I've tried all the rest apart this, but genuinely don't know what to do anymore... Any suggestion is appreciated!

    Read the article

  • Take a psd and export all the layers into separate image files without using Photoshop

    - by Darren
    This is my first question here and I'm not sure if this is the appropriate site for my question, but here it goes. If this is not the correct site, please direct me to the correct one. I have a psd file. I simply want to extract the layers from it into separate image files. I do not have Photoshop and I've already used up the free Photoshop trial. I am running Mountain Lion on a MacBook Pro. I have installed Gimp, but I can't figure out how to do what I want. I have searched using Google, but have been completely unable to find a tutorial showing me how to do this. I would really appreciate instructions or a link to an appropriate tutorial. I have seen some answers here, but they simply say use Gimp, which doesn't help me.

    Read the article

  • Automating the insertion of credits in video files on Mac OS X

    - by Roberto Aloi
    I have a bunch of video files (mp4). I need to insert some title information at the beginning and some credits at the end. I'm currently using iMovie. Since the title could be extracted from the filename and the credits are always the same, I'm wondering how could I make all this process as automatic as possible. I was thinking about Automator, but I'm open to any other solution. iMovie is the preferred tool so far, but I could use anything else (as far as it doesn't require additional licensing/cost). Any idea?

    Read the article

  • Painless way of consolidating files across multiple machines/OSes

    - by 5arx
    Just bought a NAS. So I thought I'd get all our photos, media files and pdfs consolidated, de-duplicated, de-junked and virus-checked and stick them all on it. We have 3 laptops, one running Windows, the others OSX. We have a file server running Windows - it was the result of an earlier attempt at a networked fileserver - and a Mac Pro that is also kind of a server (previous attempts at this job have resulted in most of our stuff being on it). Also memory cards/sticks, cd backups and so on. I would be grateful if anyone could suggest a strategy or, ideally, tool(s) I could use to solve this problem. It is probably no more than one or two terabytes of data in total, but I can imagine that going through it all manually, file-by-file may well drive me insane.

    Read the article

  • Explorer forgetting how to cut and paste files

    - by raimesh
    My Windows XP (SP3) machine (at work) occasionally "forgets" how to cut-and-paste files in Explorer - if I cut or copy a file then go to paste it elsewhere, the "paste" and "paste shortcut" menu items are greyed out. The keyboard shortcuts (Ctrl+X/Ctrl+C,Ctrl+V) don't work either, and neither does Drag-and-Drop - clicking and dragging a file/folder doesn't change the pointer and letting go doesn't drop the item anywhere. Restarting the explorer.exe process usually fixes the problem, but sometimes I need to do a full reboot. Note: this may be related to the following existing question, although that's Vista and it sounds like the asker has more permanent problems than I do: http://superuser.com/questions/27045/copy-paste-in-vista-explorer-broken-not-ms-vpc So, any idea what might be causing the problem and how to avoid it?

    Read the article

  • how to play MP3 files at fast speed

    - by Anil
    I have MP3 file and the contents are continuous and slow. Is there any tool, which converts them to fast speed. I am aware of the fact that with VLC, i can play fast. But, the problem is every time i have to fix the speed of the player. The Question is, i dont want to manipulate/tweak every time i play the file. I wish to have permanent solution to play the slow playing MP3 files to play fast (some thing like saving the file with fast speed etc).

    Read the article

  • Best way to use mod_rewrite to replace WordPress pages with static files

    - by David Moles
    Here's the situation: I've got an old WordPress installation that I'd like to archive as static files, but I'd also like to preserve old URLs. I've already created the static archive with wget and sorted out the filenames and links. Now I'd like to configure Apache to intercept requests for the old dynamic URL and replace them with the new static one, e.g.: http://www.example.org/log/?p=1234 or http://www.example.org/log/index.php?p=1234 should redirect to http://www.example.org/log/archives/1234.html I've tried adding the following to the VirtualHost config for example.org, but to no effect -- I just get the PHP page. RewriteCond %{REQUEST_URI} /log/ RewriteCond %{QUERY_STRING} p=([^&;]*) RewriteRule ^/$ http://%{SERVER_NAME}/log/archives/%1.html [R,L] I've enabled logging and I can see what look like other rules being applied, but not this one. None of my other guesses at match patterns for %{REQUEST_URI} seem to have any effect either (log, log/, log.*, even .*). I'm new to mod_rewrite and this is mostly cargo cult, so I'm pretty sure I've gotten it wrong. Anyone know what I should be doing here?

    Read the article

  • Does aria2 support write small files in batch?

    - by Jon
    I'm using aria2 to download 8 million jpg from flickr. Each image is about 100KB. I got a list of urls of these images in a txt file, the format is: http://farm2.staticflickr.com/1070/1151334893_5a8e7f77f4.jpg I'm wondering whether aria2 support writing small files in batch? Say write 100 image to disk when all of them are download in the memory, not just write every single file when the download is finished. Because I think writing in batch will better protect my hard disk. Or do you have other software or opensource code to recommend?

    Read the article

  • umask is being ignored on Gentoo while creating new files

    - by drcelus
    I have a server running Gentoo and hosting a drupal installation. Whenever a Drupal update is executed, the directory permissions of the updated module turn from 755 to 744 preventing the application from accessing the files. The umask is defined as 022 under /etc/profile and the Apache server is running under user and group nobody. I believe this has nothing to do with the drupal installation since if I create a directory as root, the same happens, it is created with 744 permissions, since the umask is 022 shouldn't it be created as 755 ? Why is the umask being ignored and how do I tell the server to create the directories with permission 755 ?

    Read the article

  • Certain selected files are not deleting when I use Disk Cleanup

    - by Stephanie Malouff
    I am using Windows 7 and have been trying to retrieve disk space, but when I run Disk Cleanup the Temporary Internet files (18.1KB) and System archived Windows Error Report (267KB) will not delete. It goes through the process as if they are being deleted, but my disk space remains the same; I then go back to Disk Cleanup to see if they have been removed and they are still there. I have tried removing them individually as well, but the outcome is still the same. Can someone please tell me why this is happening and what I can do to get them to delete properly?

    Read the article

  • Backup / Disaster Recovery, should I store RAR-compressed files?

    - by moraleida
    I'm in the process of recovering files from an accidentally formated Ext4 partition using Photorec. It had about 300Gb of data, of which I've already got hold of about 30Gb. So far, it seems to me that the recovery of RAR-compressed files has been much more successful than the recovery of individual uncompressed files and ZIP compressed files - in the sense that a lot of recovered files/zips were unreadable, and pretty much all of the RAR files were intact. Is there such a relation? Are RAR-compressed files really less prone to corruption and thus easier to recover?

    Read the article

  • Web log files analyzer

    - by Peter Štibraný
    I already use Google Analytics on my page, but I'd like to get additional info from log files. I've looked at various packages during last days, but nothing impressed me so far. Some requirements: must work on log file level (I use apache combined logs, but can configure apache to produce other types of logs) can generate static reports (windows/linux) or use GUI (windows only) should be easy to add custom user agents, and rerun analysis if it can recognize installation of eclipse plugins from log, that would be big plus understands google serp position referer should not require two days to setup (awstats, I am looking at you) should be still under active developement (i.e. analog isn't good answer) preferrably free, or at not very expensive :-) Any good analyzers programs out there?

    Read the article

  • Database or website of kernel config files ?

    - by Kami
    I've experienced some kernel panic after trying to compile gentoo kernel for a Sun UltraSPARC T5120 Server. The kernel panic came from a missing support for the SAS disk controller in the menu config. I've wasted so much time because I had no clue about the hardware I was using. I know that the kernel config depends on what you plan to do with your machine but I want to have a configuration file that at least match my hardware ! Is there a website or database that provides menuconfig's kernel configuration files for known or branded hardware like Dell Server or Apple computers ?

    Read the article

  • rsync to ONLY keep files in destination that have been removed from source

    - by David Corley
    We use rsync to copy filesystem contents from one machine to another as a backup. We first run MACHINE-X-MACHINE-Y rsync for a straight backup with the --delete and --delete-excluded switches We also run an internal Rsync between the MACHINE-Y destination, and another folder on MACHINE-Y with either of the delete flags. This maintains a non-destructive copy in the event someone inadvertently deletes a file on MACHINE-X. However, it also has the overhead of being a complete copy of what has already been synchronized. Ideally I want to be able to run the non-destructive rsync in such a way that the destination ONLY receives the deleted files and so avoids unnecessary duplication . Is there any way to do this?

    Read the article

  • Server unresponsive, messages shown on console but not in log files

    - by raistlin majere
    I'm using Ubuntu Server 10.04.4, and once in a while the server hangs (once a month) and is totally unresponsive. The tty is flooded with messages like these. The problem is that these messages are not in my log files after reboot. How to log these messages so that I can analyze them later? In the current logs I can't see anything that would tell me why this is happening. I would also appreciate if anybody can tell from those messages what's going on. This server is a guest virtual machine. The host server is also Ubuntu server 10.04 with KVM/QEMU.

    Read the article

  • Is there any way of preventing .csv files being converted into excel format

    - by Kevin Trainer
    I'm trying to work with an automated testing tool which can use .csv files as its data sourse. After saving a notepad file containing a number of fields and data seperated by commas as .csv it appears to have been converted to an excel file. When I run the test, only the first line of values is identified and can be run within the automated test. Not sure if this is expected with the testing product (www.badboy.co.au), but just wondered if there was a way of preventing excel from taking control of the .csv file? Any helpfull feedback would be great.

    Read the article

  • find files their name is smaller or greater than a given parameter

    - by Tzury Bar Yochay
    Say that in a given directory I got tzury@x200:~/Desktop/sandbox$ ls -l total 20 drwxr-xr-x 2 tzury tzury 4096 2011-03-09 10:19 N00.P000 drwxr-xr-x 2 tzury tzury 4096 2011-03-09 10:19 N00.P001 drwxr-xr-x 2 tzury tzury 4096 2011-03-09 10:19 N00.P002 drwxr-xr-x 2 tzury tzury 4096 2011-03-09 10:19 N00.P003 drwxr-xr-x 2 tzury tzury 4096 2011-03-09 10:19 N00.P004 drwxr-xr-x 2 tzury tzury 4096 2011-03-09 10:19 N01.P000 drwxr-xr-x 2 tzury tzury 4096 2011-03-09 10:19 N01.P001 drwxr-xr-x 2 tzury tzury 4096 2011-03-09 10:19 N01.P002 I seek for a bash way to grab the list of files which their name is either grater or smaller than a given parameter, for instance: $ my_finder lt N00.P003 shall return N00.P000, N00.P001 and N00.P002 $ my_finder gt N00.P003 shall return N00.P004, N01.P000, N01.P001 and N01.P002 I was thinking of iterating over for name in $(ls) and while $name != $2 but believe there are more elegant way of doing so

    Read the article

  • xcopy files and directory

    - by user1044937
    I have a folder named "C:\Jobs\job#1" , "C:\Jobs\job#2" "C:\Jobs\job#3" etc and a lot of directories and sub-directories under it. I want to get the all the directories under Jobs and xcopy them to C:\backup. Then I want to xcopy all the files under each Job#1, 2 ,3 etc. to C:\backup\job#1\month\\*.* To make it clearer. Source dir = C:\Jobs\job#1\"myfiles&dir" Destination dir = C:\Backup\job#1\month\"myfiles&dir" then do the next folder Source dir = C:\Jobs\job#2\"myfiles&dir" Destination dir = C:\Backup\job#2\month\"myfiles&dir" ...until all folders are back-up. Since the job folder keep increasing, by doing it this way I don't have to add extra code on this script except modify the month. Thank you.

    Read the article

  • Migrate to SSD - NTFS mount point for Program Files

    - by Icode4food
    Here is my thought. I have a new computer that I just built and am considering migrating to a SSD. I have Windows all setup and my Development environment configured so I want to avoid having to re-install a bunch of stuff. My thought is to clone my OS (win7) to the SSD and then mount a HDD partion to C:\Program Files (x86)\ with C being my SSD. This way as far as the programs are concerned they still live on the C drive but in reality they are physically located on the HDD. This seems to me like a good idea but after searching around a bit and not having found anyone else that had the same idea, I'm wondering why not. Maybe I am missing something that is obvious to everyone but me. Why is this a good or a bad idea?

    Read the article

  • "Keep iTune music files organized" no longer works?

    - by user22105
    I remember iTunes used to have the option to "keep music files organized" (i know it still does in its advanced Properties). When I used to edit music's artist, genre or album information. The mp3 file is automatically updated as well. But lately I've been updating some mp3s. And if I tried to open the updated mp3's Properties, under Details, the Artist, genre, and album information is not available. How can i fix this? Added: you can try editing this mp3 file with iTunes, it didn't work on my PC Here's a screenshot of what Im seeing after the edit.

    Read the article

  • Cannot access an application folder in Program files

    - by GiddyUpHorsey
    I recently installed Windows 7 Professional 64bit on a new machine. I installed an application using a ClickOnce installer. The application runs fine, but I cannot access the application folder it created in c:\Program files (x86). It bombs with access denied. I try to view the properties on the folder and it takes about 1 minute to display (other folders take 1 second). It says I cannot view any information because I'm not the owner. It doesn't say who the current owner is (instead - Unable to display current owner.) but says I can take ownership. When I try it fails again with Access Denied, even though I have administrative permissions. Why can't I access this folder nor take ownership?

    Read the article

  • Why the huge discrepancy in size between two similar zip files

    - by twpc
    I use WinZip to zip entire directories of code and send them to a fellow programmer. He makes changes and sends the directories of code back to me. Ignoring the fact that this is not a good way to keep the code clean when we are both working on it, I notice that his zip files are far smaller than mine, with basically the same data inside (mine range around 36,000 KB, his 2,000 KB). I believe he is also using WinZip. What's going on here, and how can I make mine "more compressed"?

    Read the article

  • Search for files which will open a certain application in Mac OS X

    - by Jacob Palme
    In Mac OS X, when you doubleclick on a file name, that file will open with the application which created the file. So there must be stored, somewhere in the file description on a Mac OS X file, information on which application created this file. Note that this is not the file extension, the file can have any extension, or no extension at all. Two questions regarding this information: (1) How can I search for all files which will open a specific application? (2) How can I see, and change, the application which a certain file will open?

    Read the article

  • nginx php-fpm keeps downloading files

    - by Sam Williams
    vhost: server { listen *:8080; location / { root /var/www/default/pub; index index.php; # if file exists return it right away if (-f $request_filename) { break; } if (!-e $request_filename) { rewrite ^(.+)$ /index.php$1 last; break; } } # serve static files directly location ~* \.(jpg|jpeg|gif|css|png|js|ico|html)$ { access_log off; expires max; } location ~* \.php$ { # By all means use a different server for the fcgi processes if you need to fastcgi_pass 127.0.0.1:9000; fastcgi_param SCRIPT_FILENAME $document_root$fastcgi_script_name; fastcgi_param PATH_INFO $fastcgi_script_name; include /usr/local/nginx/conf/fastcgi_params; } location ~ /\.ht { deny all; } } http://192.168.135.128/index.php loads just fine... http://192.168.135.128/public_/html/index.php downloads...

    Read the article

  • Apache 2.0 is showing only text and not php files

    - by denonth
    I have a web application written with PHP, html and JavaScript. On my pc I have installed a EasyPHP program which has Apache and everything installed. But I wanted to put this web app to my server and I have installed a Apache 2.0 but my php files are displayed as text or it starts to download them automatically. I have tried several things one of them is to add this to my conf file: AddType application/x-httpd-php .php AddType application/x-httpd-php-source .phps But it's still not working. What else can I do? Thank you

    Read the article

< Previous Page | 149 150 151 152 153 154 155 156 157 158 159 160  | Next Page >