Search Results

Search found 39784 results on 1592 pages for 'ignore files'.

Page 152/1592 | < Previous Page | 148 149 150 151 152 153 154 155 156 157 158 159  | Next Page >

  • A download manager for Linux which saves downloaded files in directories by date like 2012_06_29

    - by Gart
    I've been using Download Master on Windows for years and what I liked most about it is that this program can automatically put downloaded files into directories by download date: /Downloads | |--/2012_06_28 | | | |--a.zip | |--b.pdf | ... | |--/2012_06_29 | | | |--c.txt | ... ... I'm looking for something similar for Linux. Is there any free download manager that can do this? I have tried KGet and uGet but they both seem to lack this feature. If there is a way to configure them to do that, I'll be happy to know about it. Thank you.

    Read the article

  • Obtaining list of files from php script?

    - by SenorSputnik
    Naenara/KCCKP offers a catalog of hundreds of mp3 files that can only be downloaded in small amounts at a time. Clicking on a song title invokes mp3player.php and downloads a plaintext link to the mp3 file: http://www.kcckp.net/mp3player.php?e+8 Going directly to mp3player.php will display Korean error messages and send you back 1 page in your history. Is there any way to parse/coax mp3player.php into dumping a full list of mp3 links? I am sorry if this is a painfully easy or impossible task, I have not even begun to delve into php. </newbiewhining>

    Read the article

  • How do i get Safari to ignore the SSL Certificate error?

    - by Tangopop
    In IE 6, 7, 8 and Firefox 3.6.3 and 3.0.5 i have installed a local SSL Certificate on the machine i am testing on and i have gotten the browser to igonre the SSL error (which is off one of my Web Test servers) Now i am tryin to do the same thing within safari 4 and with no luck. Basically i am running some automated scripts to test my website before they go live and i need to be able to ignore these errors as they will all run autonomosly. This is the error screen i am trying to avoid: http://library.bowdoin.edu/news/images/ezproxy-err/safari.jpg As i say i have installed the certificate locally and the IE 7 browser on the same machine works fine.

    Read the article

  • Find words in many files

    - by ant2009
    Hello, I am looking for this struct messages_sdd_t and I need to search through a lot of *.c files to find it. However, I can't seen to find a match as I want to exclude all the words 'struct' and 'messages_sdd_t'. As I want to search on this only 'struct messages_sdd_t' The reason for this is, as struct is used many times and I keep getting pages or search results. I have been doing this without success: find . -type f -name '*.c' | xargs grep 'struct messages_sdd_t' and this find . -type f -name '*.c' | xargs egrep -w 'struct|messages_sdd_t' Many thanks for any suggestions,

    Read the article

  • Internet Explorer 8 Viewing PDF files vs saving them

    - by Andy Evans
    I have a user asking about viewing PDF files in Internet Explorer. When she clicks the file link on a website, IE prompts her to save the file, which she then has to open to read. What she's not seeing in the prompt is the ability to open the file without having to save it, or, just opening the PDF automatically. She and a few others are having the same problem, while a majority of the user do not have this problem. All of the users have IE 8 and Adobe Reader 9 installed. I've verified that Adobe Reader's "Display PDF in browser" option is enabled. What else should I check?

    Read the article

  • Apache - Same username in several .htpasswd files

    - by greydet
    In a virtual host, I setup two different <Location> blocks for which the access is restricted by two basic authentication htpasswd files. One htpasswd contains different usernames + a common user name. The other htpasswd file only contains the common user name. My problem is that once users connect a location with the common user name, they have immediate access to the other location without being asker for a different user name. Is there a way to restrict the username access only to the corresponding htpasswd file? Is there a way for users to ask to be re-prompted for another username/password?

    Read the article

  • Serving static web files off a non-standard port

    - by Nimmy Lebby
    I'm close to deploying a Django project to production. I'm looking over some infrastructure decisions. Something that came up was serving static files with a different server such as lighttpd. However, we're starting off with a single dedicated server so our only option would be to use a non-standard port for the static file webserver. Is there precedence for this? I.e. Does anyone "big" do this? Any particular port I should use or shy away from using? Can anyone thing of some downsides of going this route?

    Read the article

  • Serving Compressed Files Amazon vs Lightty

    - by tike
    We are currently using amazon CloudFront to serve css and according to Amazon itself, Amazon CloudFront can serve both compressed and uncompressed files from an origin server. But while i check compression it shows everything fine in origin server but it shows notcompressed checking in the link with cloudfront. e.g. http://www.port80software.com/tools/compresscheck.asp?url=http%3A%2F%2Fimgsrv.mydomain.com%2Fen-UK%2Fsomething.css it would result with Compression status: (gzip) while with cloudfront http://www.port80software.com/tools/compresscheck.asp?url=http%3A%2F%2hereisit.cloudfront.net%2F%2Fsomething.css Compression status: Uncompressed Origin server is running lighttpd with mod_deflate however, allowed config is: deflate.allowed_encodings = ("bzip2", "gzip", "deflate") [i would think, putting extra allowed encoding wont affect as such.] Here i am clueless, what is the real issue.

    Read the article

  • USB Flash drive corrupt files - Help needed

    - by Michael
    I have a 16gb usb flash drive with 8gb worth of data that I can't afford to lose. When I inserted it into my pc, inside the folder that I was storing the data I saw unintelligible characters and nothing in there would open. I ran, windows scandisk and the files (unfortunately) disappeared. I can see that the drive's space still appears to be taken up with data, about 8gb. What should I do to recover it? Is it possible? Thanks in advance... Michael

    Read the article

  • Unable to copy multiple files from remote sftp via ssh

    - by Kevin S
    I am trying to copy multiple (1000s) .JPG files from remote SFTP server to current server using mget command but it is not working. Here is the command that I used sftp [email protected] sftp> cd /images sftp> mget *.JPG Screenshot at http://i.stack.imgur.com/bSciQ.png Nothing happens so when I do CTRL + C it gives an error that says File "/images/*.JPG" not found. I have already referred all suggestions posted on the forum but nothing is working. Let me know if there is another way to copy.

    Read the article

  • Set nginx.conf to deny all connections except to certain files or directories

    - by Ben
    I am trying to set up Nginx so that all connections to my numeric ip are denied, with the exception of a few arbitrary directories and files. So if someone goes to my IP, they are allowed to access the index.php file, and the phpmyadmin directory for example, but should they try to access any other directories, they will be denied. This is my server block from nginx.conf: server { listen 80; server_name localhost; location / { root html; index index.html index.htm index.php; } location ~ \.php$ { root html; fastcgi_pass unix:/var/run/php-fpm/php-fpm.sock; fastcgi_index index.php; fastcgi_param SCRIPT_FILENAME /srv/http/nginx/$fastcgi_script_name; include fastcgi_params; } } How would I proceed? Thanks very much!

    Read the article

  • uploading large files (mp4) to IIS 7.5 gives 500 Internal Server Error

    - by dragon112
    I made a website on which i need to be able to upload video files and it has worked for quite a while. However after a while it just stopped working and now it will give me the following IIS error message when i upload a video. Images do work (possibly due to their smaller size). I use an html form with PHP server sided script to upload. I have already set the user permissions for the entire inetpub to allow all actions for the IIS user. If you have any idea what it could be PLEASE tell me, have been trying to fix this for weeks now. Thanks in advance!

    Read the article

  • How can I make Windows Media Player ignore global hotkeys in Windows 7?

    - by Schnapple
    I have Windows 7 and a Logitech G15 keyboard. One of the programs with the Logitech G15 allows you to control media players such as Winamp with the playback keys on the keyboard. Problem I'm having is that, even though I have told this program to not control Windows Media Player, every time I use it to pause Winamp, it then hits plays (or unpauses) Windows Media Player. Even more annoying given that Windows Media Player isn't even running as an active GUI program and instead as a background process, so I hear the sound of whatever the last video it was I playing. If I end-task wmplayer.exe it spins right back up but at least now it has no knowledge of a video to play, but this is annoying. How can I either a) Have Windows Media Player in Windows 7 completely unload when I close it, or b) Have Windows Media Player in Windows 7 ignore any sort of global hotkeys?

    Read the article

  • Dell Powervault MD3000 - Not sharing Files between servers

    - by Kevin
    I'm a developer who has to set up a Dell Powervault MD3000 due to lack of resources. I have connected the Powervault to 2 Dell 2950 servers via the SAS cables. I performed the setup using Dell's MD Storage Manager software (4 disks, RAID 5 with hot spare). Then I added the disks using Windows 2003 disk management (Basic, not dynamic disk and formatted with NTFS). When I add files to the array from one server, they are not visible on the other server (and vice-versa). Is the error in the windows disk management configuration?

    Read the article

  • RoboCopy Log Files missing whitespace?

    - by TwystNeko
    So I'm working on a script to use RoboCopy to copy a bunch of files, and log what it's copied. It works reasonably well, except for the logfiles. They tend to look like this: C:\Users\Tech\Documents\desktop.initechscan1.jpgtechscan2.jpgtechscan3.jpgtechscan4.jpgtechscan5.jpgwsus.jpgwsus2.jpgC:\Users\Tech\Documents\My Music\C:\Users\Tech\Documents\My Pictures\C:\Users\Tech\Documents\My Videos\C:\Users\Tech\Documents\My Digital Editions\ As you can see, the log seems to be missing all whitespace and separators between entries. Is there something I can do to fix this? It's kind of frustrating. The commandline I'm using is this: C:\Users\Tech\Desktop>Robocopy.exe C:\Users\Tech\Documents c:\Temp /e /l /b /xj /xf ntuser.* desktop.ini *.lnk /np /njh /log:migratedfiles.txt /v I have the /l in there since I'm debugging, and it's the easiest way to keep from copying everything a million times.

    Read the article

  • Live streaming external FLV files

    - by fandelost
    I want to broadcast live video from a playlist created by me, but the actual video files (.flv's) are stored in some external file hosting which I can't admin, but it's currently allowing the use of flash players for playing its contents (progressively). Which is the best solution for this? Can I use my current web server (Debian w/Apache) or should I use some specific server software? I've read some names such as Red5 and Wowza but I can't tell if those are solutions to my actual problem or how should I use them. Also, costs should be as low as possible for this project.

    Read the article

  • How can I edit files directly from find results

    - by Sergio Tulentsev
    Let's say, I want to replace one string with another. Normally I use TextMate, but today, I don't remember why, I decided to run the search in Sublime Text 2 (which I installed some time ago, but haven't used since). I was impressed with "Find Results" window, seems more useful that textmate's. Then I tried to edit text right inside that window and (I can swear on my macbook) it worked (I was able to commit those changes to git repo)! I thought then, "Wow, that's a really nice feature. Should I migrate from TM already?". Unfortunately, I wasn't able to reproduce this behaviour later. I can still edit text in "find results" window, but changes don't persist (that is, original files are not touched). Am I doing something wrong now, or was it caffeine-induced hallucination earlier in the day?

    Read the article

  • cmd files to hta

    - by Frode Eskil
    I have a lot of cmd files i use daily for example to add users to local groups, installing printers, run as admin tasks etc. I like to take the scripts i use most frequently and add them to a tabbed hta file, but i have trouble finding a good guide on how to easily do it. Anyone having a good site to share with me? Or do i finally have to start with vb scripting? I have done some but it's so much faster to do a cmd file for me.

    Read the article

  • Copy files with filter (XP)

    - by fire
    I have a huge folder (over 6GB) with multiple sub-folders that I want to copy onto an external hard drive, however I do not want it to copy any PDF, EXE or ZIP files across to save space. Is there any software that will help me achieve this? I have looked at TeraCopy but this doesn't seem to have any filter mechanism on it. I am using Windows XP (* sigh *). *edit: found the xcopy command, will this do it? Can anyone help me with the syntax?

    Read the article

  • include all vim files in a folder

    - by queueoverflow
    For my .bashrc I have a lot of small snippet files in .config/bash, like 10-prompt.sh and so on. In my actual .bashrc, I just have the following: configdir="$HOME/.config/bash" for file in "$configdir"/*.sh do source "$file" done I'd like to do the same for my .vimrc, but I am not that confident in VimL that I could write that. How would the snippet for .vimrc look like that includes all the snippets in a given subfolder? Ideally, I'd like to make a .vim/rc/ folder where I can put my snippets into.

    Read the article

  • Batch file to ZIP only files in directory or sub directory

    - by PaulJavier
    I wanted to know if possible how to create a command line to do the following - if a directory exist ZIP only the contents into a ZIP file. If a directory has sub-directories ZIP only the contents into another ZIP file. Example: C:\Directory\sample.txt ZIP only sample.txt C:\Directory\Directory1\sample1.txt ZIP only sample1.txt C:\Directory\Directory1\Directory2\sample2.txt ZIP only sample2.txt So it would have created 3 zip files in C:\Directory and sub-directories. I will not know the name of the sub-directories so can I also assign some sort of variable that says if there are directories or sub-directories in C:\Directory then start above ZIP(s)? Thank you, Paul

    Read the article

  • Automatically cycle numerous or large files to the trash

    - by minameismud
    I've been tasked with fixing a vendor's program that, under certain conditions, dumps gigs of junk files into a log directory. It ends up filling users' machines. My task is to figure out how to make it stop without any source code or additional running processes, and without making the program kasplode. In other words, I'm looking to use a feature of the file system to control the growth. One idea I had was to make a hard link from that folder to NUL, as you might with /dev/null in the linux world. However, my attempts to use the mklink program to create a junction result in a message that says Local volumes are required to complete the operation. Any ideas on how to complete the junction, or other ideas to solve the problem?

    Read the article

  • Rename multiple files as "Modified Date/Time" using cmd or Powershell

    - by Mehper C. Palavuzlar
    I have hundreds of JPG files in a folder. I want to rename each file so that the file name is replaced with "Modified Date/Time" of that file, namely DD.MM.RRRR.HH.MM.jpg. For example, Before After 001.jpg 11.01.2011.16.58.jpg 002.jpg 12.01.2011.09.32.jpg 003.jpg 14.01.2011.12.41.jpg ... ... Since colon (:) cannot be used in file names, the colon between HH and MM must be replaced with a period. I don't want to use a 3rd party tool. Can you provide me with the code to achieve this in Powershell or command line?

    Read the article

  • Converting DOCX files to PDF via SSH without losing formatting

    - by Reado
    I'm struggling to find a solution that will allow me to convert a DOCX file to a PDF without losing or malforming the formatting of the document on CentOS 5.7. I have tried CUPS-PDF but it doesn't work; spool files appear in the /var/spool folder but nothing happens after that. OpenOffice and LibreOffice converted a DOCX to PDF but the formatting was all wrong. However if I print the DOCX to a Windows PDF printer from my Windows 7 workstation, it outputs to PDF absolutely fine. So why can't Linux do the same? I tried to print via CUPS to the Windows PDF printer (shared) but the document appears in the queue as "Remote Downlevel Document" and doesn't print. This only happens when I print from Linux.

    Read the article

  • Restore SBS 2003 CompanyWeb calendar from flat files?

    - by BrandonS
    Ok so I have a job of recovering a calendar that was used for an event schedule. The situation is that there was never a backup done except for using Carbonite on the C drive. I have re-installed the server with the same server name and domain. I tried stopping the mssqlsharepoint service and over-righting the two DB files (STS_1.mdf and STS_1_log.LDF) and each one individually each time stopping and restarting the service. Now I knew before I started that this had a slim chance of working but I had to try because everything else i could find involved backing up before restoring and this just wasn't done. Please someone help it is driving me mad I tell you MADDDDDD. ps i am not the genius that set this up just the fool tasked to clean up the mess. :-)

    Read the article

< Previous Page | 148 149 150 151 152 153 154 155 156 157 158 159  | Next Page >