Search Results

Search found 39200 results on 1568 pages for 'zip files'.

Page 157/1568 | < Previous Page | 153 154 155 156 157 158 159 160 161 162 163 164  | Next Page >

  • Live streaming external FLV files

    - by fandelost
    I want to broadcast live video from a playlist created by me, but the actual video files (.flv's) are stored in some external file hosting which I can't admin, but it's currently allowing the use of flash players for playing its contents (progressively). Which is the best solution for this? Can I use my current web server (Debian w/Apache) or should I use some specific server software? I've read some names such as Red5 and Wowza but I can't tell if those are solutions to my actual problem or how should I use them. Also, costs should be as low as possible for this project.

    Read the article

  • RoboCopy Log Files missing whitespace?

    - by TwystNeko
    So I'm working on a script to use RoboCopy to copy a bunch of files, and log what it's copied. It works reasonably well, except for the logfiles. They tend to look like this: C:\Users\Tech\Documents\desktop.initechscan1.jpgtechscan2.jpgtechscan3.jpgtechscan4.jpgtechscan5.jpgwsus.jpgwsus2.jpgC:\Users\Tech\Documents\My Music\C:\Users\Tech\Documents\My Pictures\C:\Users\Tech\Documents\My Videos\C:\Users\Tech\Documents\My Digital Editions\ As you can see, the log seems to be missing all whitespace and separators between entries. Is there something I can do to fix this? It's kind of frustrating. The commandline I'm using is this: C:\Users\Tech\Desktop>Robocopy.exe C:\Users\Tech\Documents c:\Temp /e /l /b /xj /xf ntuser.* desktop.ini *.lnk /np /njh /log:migratedfiles.txt /v I have the /l in there since I'm debugging, and it's the easiest way to keep from copying everything a million times.

    Read the article

  • uploading large files (mp4) to IIS 7.5 gives 500 Internal Server Error

    - by dragon112
    I made a website on which i need to be able to upload video files and it has worked for quite a while. However after a while it just stopped working and now it will give me the following IIS error message when i upload a video. Images do work (possibly due to their smaller size). I use an html form with PHP server sided script to upload. I have already set the user permissions for the entire inetpub to allow all actions for the IIS user. If you have any idea what it could be PLEASE tell me, have been trying to fix this for weeks now. Thanks in advance!

    Read the article

  • How can I edit files directly from find results

    - by Sergio Tulentsev
    Let's say, I want to replace one string with another. Normally I use TextMate, but today, I don't remember why, I decided to run the search in Sublime Text 2 (which I installed some time ago, but haven't used since). I was impressed with "Find Results" window, seems more useful that textmate's. Then I tried to edit text right inside that window and (I can swear on my macbook) it worked (I was able to commit those changes to git repo)! I thought then, "Wow, that's a really nice feature. Should I migrate from TM already?". Unfortunately, I wasn't able to reproduce this behaviour later. I can still edit text in "find results" window, but changes don't persist (that is, original files are not touched). Am I doing something wrong now, or was it caffeine-induced hallucination earlier in the day?

    Read the article

  • cmd files to hta

    - by Frode Eskil
    I have a lot of cmd files i use daily for example to add users to local groups, installing printers, run as admin tasks etc. I like to take the scripts i use most frequently and add them to a tabbed hta file, but i have trouble finding a good guide on how to easily do it. Anyone having a good site to share with me? Or do i finally have to start with vb scripting? I have done some but it's so much faster to do a cmd file for me.

    Read the article

  • include all vim files in a folder

    - by queueoverflow
    For my .bashrc I have a lot of small snippet files in .config/bash, like 10-prompt.sh and so on. In my actual .bashrc, I just have the following: configdir="$HOME/.config/bash" for file in "$configdir"/*.sh do source "$file" done I'd like to do the same for my .vimrc, but I am not that confident in VimL that I could write that. How would the snippet for .vimrc look like that includes all the snippets in a given subfolder? Ideally, I'd like to make a .vim/rc/ folder where I can put my snippets into.

    Read the article

  • Rename multiple files as "Modified Date/Time" using cmd or Powershell

    - by Mehper C. Palavuzlar
    I have hundreds of JPG files in a folder. I want to rename each file so that the file name is replaced with "Modified Date/Time" of that file, namely DD.MM.RRRR.HH.MM.jpg. For example, Before After 001.jpg 11.01.2011.16.58.jpg 002.jpg 12.01.2011.09.32.jpg 003.jpg 14.01.2011.12.41.jpg ... ... Since colon (:) cannot be used in file names, the colon between HH and MM must be replaced with a period. I don't want to use a 3rd party tool. Can you provide me with the code to achieve this in Powershell or command line?

    Read the article

  • Automatically cycle numerous or large files to the trash

    - by minameismud
    I've been tasked with fixing a vendor's program that, under certain conditions, dumps gigs of junk files into a log directory. It ends up filling users' machines. My task is to figure out how to make it stop without any source code or additional running processes, and without making the program kasplode. In other words, I'm looking to use a feature of the file system to control the growth. One idea I had was to make a hard link from that folder to NUL, as you might with /dev/null in the linux world. However, my attempts to use the mklink program to create a junction result in a message that says Local volumes are required to complete the operation. Any ideas on how to complete the junction, or other ideas to solve the problem?

    Read the article

  • Restore SBS 2003 CompanyWeb calendar from flat files?

    - by BrandonS
    Ok so I have a job of recovering a calendar that was used for an event schedule. The situation is that there was never a backup done except for using Carbonite on the C drive. I have re-installed the server with the same server name and domain. I tried stopping the mssqlsharepoint service and over-righting the two DB files (STS_1.mdf and STS_1_log.LDF) and each one individually each time stopping and restarting the service. Now I knew before I started that this had a slim chance of working but I had to try because everything else i could find involved backing up before restoring and this just wasn't done. Please someone help it is driving me mad I tell you MADDDDDD. ps i am not the genius that set this up just the fool tasked to clean up the mess. :-)

    Read the article

  • Converting DOCX files to PDF via SSH without losing formatting

    - by Reado
    I'm struggling to find a solution that will allow me to convert a DOCX file to a PDF without losing or malforming the formatting of the document on CentOS 5.7. I have tried CUPS-PDF but it doesn't work; spool files appear in the /var/spool folder but nothing happens after that. OpenOffice and LibreOffice converted a DOCX to PDF but the formatting was all wrong. However if I print the DOCX to a Windows PDF printer from my Windows 7 workstation, it outputs to PDF absolutely fine. So why can't Linux do the same? I tried to print via CUPS to the Windows PDF printer (shared) but the document appears in the queue as "Remote Downlevel Document" and doesn't print. This only happens when I print from Linux.

    Read the article

  • Take a psd and export all the layers into separate image files without using Photoshop

    - by Darren
    This is my first question here and I'm not sure if this is the appropriate site for my question, but here it goes. If this is not the correct site, please direct me to the correct one. I have a psd file. I simply want to extract the layers from it into separate image files. I do not have Photoshop and I've already used up the free Photoshop trial. I am running Mountain Lion on a MacBook Pro. I have installed Gimp, but I can't figure out how to do what I want. I have searched using Google, but have been completely unable to find a tutorial showing me how to do this. I would really appreciate instructions or a link to an appropriate tutorial. I have seen some answers here, but they simply say use Gimp, which doesn't help me.

    Read the article

  • md5sum of large files gives different results sometimes

    - by Emanuele
    I have an AMD quad core, 8 gb RAM, 1 SSD EXT2 (2 months old), 2 HDD EXT4, approximately 1 year old. I'm using Ubuntu 10.04 x86-64 and when I compute the md5sum of large files (9 GB) sometimes I get different values than the one stored on a reference file. Upon restarting and switching off the PC then I get the expected results no matter how many times I repeat it. But this is random. I've turn on ECC (the fastest possible settings) and the issue seems to be rarer, but I've run memtest86+ for 6+ hours without a glitch (and with ECC off!). Any idea? Should I update the BIOS of my motherboard (Asus EVO-something...don't remember it now)? I've tried all the rest apart this, but genuinely don't know what to do anymore... Any suggestion is appreciated!

    Read the article

  • Automating the insertion of credits in video files on Mac OS X

    - by Roberto Aloi
    I have a bunch of video files (mp4). I need to insert some title information at the beginning and some credits at the end. I'm currently using iMovie. Since the title could be extracted from the filename and the credits are always the same, I'm wondering how could I make all this process as automatic as possible. I was thinking about Automator, but I'm open to any other solution. iMovie is the preferred tool so far, but I could use anything else (as far as it doesn't require additional licensing/cost). Any idea?

    Read the article

  • Painless way of consolidating files across multiple machines/OSes

    - by 5arx
    Just bought a NAS. So I thought I'd get all our photos, media files and pdfs consolidated, de-duplicated, de-junked and virus-checked and stick them all on it. We have 3 laptops, one running Windows, the others OSX. We have a file server running Windows - it was the result of an earlier attempt at a networked fileserver - and a Mac Pro that is also kind of a server (previous attempts at this job have resulted in most of our stuff being on it). Also memory cards/sticks, cd backups and so on. I would be grateful if anyone could suggest a strategy or, ideally, tool(s) I could use to solve this problem. It is probably no more than one or two terabytes of data in total, but I can imagine that going through it all manually, file-by-file may well drive me insane.

    Read the article

  • Explorer forgetting how to cut and paste files

    - by raimesh
    My Windows XP (SP3) machine (at work) occasionally "forgets" how to cut-and-paste files in Explorer - if I cut or copy a file then go to paste it elsewhere, the "paste" and "paste shortcut" menu items are greyed out. The keyboard shortcuts (Ctrl+X/Ctrl+C,Ctrl+V) don't work either, and neither does Drag-and-Drop - clicking and dragging a file/folder doesn't change the pointer and letting go doesn't drop the item anywhere. Restarting the explorer.exe process usually fixes the problem, but sometimes I need to do a full reboot. Note: this may be related to the following existing question, although that's Vista and it sounds like the asker has more permanent problems than I do: http://superuser.com/questions/27045/copy-paste-in-vista-explorer-broken-not-ms-vpc So, any idea what might be causing the problem and how to avoid it?

    Read the article

  • how to play MP3 files at fast speed

    - by Anil
    I have MP3 file and the contents are continuous and slow. Is there any tool, which converts them to fast speed. I am aware of the fact that with VLC, i can play fast. But, the problem is every time i have to fix the speed of the player. The Question is, i dont want to manipulate/tweak every time i play the file. I wish to have permanent solution to play the slow playing MP3 files to play fast (some thing like saving the file with fast speed etc).

    Read the article

  • Best way to use mod_rewrite to replace WordPress pages with static files

    - by David Moles
    Here's the situation: I've got an old WordPress installation that I'd like to archive as static files, but I'd also like to preserve old URLs. I've already created the static archive with wget and sorted out the filenames and links. Now I'd like to configure Apache to intercept requests for the old dynamic URL and replace them with the new static one, e.g.: http://www.example.org/log/?p=1234 or http://www.example.org/log/index.php?p=1234 should redirect to http://www.example.org/log/archives/1234.html I've tried adding the following to the VirtualHost config for example.org, but to no effect -- I just get the PHP page. RewriteCond %{REQUEST_URI} /log/ RewriteCond %{QUERY_STRING} p=([^&;]*) RewriteRule ^/$ http://%{SERVER_NAME}/log/archives/%1.html [R,L] I've enabled logging and I can see what look like other rules being applied, but not this one. None of my other guesses at match patterns for %{REQUEST_URI} seem to have any effect either (log, log/, log.*, even .*). I'm new to mod_rewrite and this is mostly cargo cult, so I'm pretty sure I've gotten it wrong. Anyone know what I should be doing here?

    Read the article

  • Does aria2 support write small files in batch?

    - by Jon
    I'm using aria2 to download 8 million jpg from flickr. Each image is about 100KB. I got a list of urls of these images in a txt file, the format is: http://farm2.staticflickr.com/1070/1151334893_5a8e7f77f4.jpg I'm wondering whether aria2 support writing small files in batch? Say write 100 image to disk when all of them are download in the memory, not just write every single file when the download is finished. Because I think writing in batch will better protect my hard disk. Or do you have other software or opensource code to recommend?

    Read the article

  • Backup / Disaster Recovery, should I store RAR-compressed files?

    - by moraleida
    I'm in the process of recovering files from an accidentally formated Ext4 partition using Photorec. It had about 300Gb of data, of which I've already got hold of about 30Gb. So far, it seems to me that the recovery of RAR-compressed files has been much more successful than the recovery of individual uncompressed files and ZIP compressed files - in the sense that a lot of recovered files/zips were unreadable, and pretty much all of the RAR files were intact. Is there such a relation? Are RAR-compressed files really less prone to corruption and thus easier to recover?

    Read the article

  • Certain selected files are not deleting when I use Disk Cleanup

    - by Stephanie Malouff
    I am using Windows 7 and have been trying to retrieve disk space, but when I run Disk Cleanup the Temporary Internet files (18.1KB) and System archived Windows Error Report (267KB) will not delete. It goes through the process as if they are being deleted, but my disk space remains the same; I then go back to Disk Cleanup to see if they have been removed and they are still there. I have tried removing them individually as well, but the outcome is still the same. Can someone please tell me why this is happening and what I can do to get them to delete properly?

    Read the article

  • Database or website of kernel config files ?

    - by Kami
    I've experienced some kernel panic after trying to compile gentoo kernel for a Sun UltraSPARC T5120 Server. The kernel panic came from a missing support for the SAS disk controller in the menu config. I've wasted so much time because I had no clue about the hardware I was using. I know that the kernel config depends on what you plan to do with your machine but I want to have a configuration file that at least match my hardware ! Is there a website or database that provides menuconfig's kernel configuration files for known or branded hardware like Dell Server or Apple computers ?

    Read the article

  • umask is being ignored on Gentoo while creating new files

    - by drcelus
    I have a server running Gentoo and hosting a drupal installation. Whenever a Drupal update is executed, the directory permissions of the updated module turn from 755 to 744 preventing the application from accessing the files. The umask is defined as 022 under /etc/profile and the Apache server is running under user and group nobody. I believe this has nothing to do with the drupal installation since if I create a directory as root, the same happens, it is created with 744 permissions, since the umask is 022 shouldn't it be created as 755 ? Why is the umask being ignored and how do I tell the server to create the directories with permission 755 ?

    Read the article

  • Web log files analyzer

    - by Peter Štibraný
    I already use Google Analytics on my page, but I'd like to get additional info from log files. I've looked at various packages during last days, but nothing impressed me so far. Some requirements: must work on log file level (I use apache combined logs, but can configure apache to produce other types of logs) can generate static reports (windows/linux) or use GUI (windows only) should be easy to add custom user agents, and rerun analysis if it can recognize installation of eclipse plugins from log, that would be big plus understands google serp position referer should not require two days to setup (awstats, I am looking at you) should be still under active developement (i.e. analog isn't good answer) preferrably free, or at not very expensive :-) Any good analyzers programs out there?

    Read the article

  • Is there any way of preventing .csv files being converted into excel format

    - by Kevin Trainer
    I'm trying to work with an automated testing tool which can use .csv files as its data sourse. After saving a notepad file containing a number of fields and data seperated by commas as .csv it appears to have been converted to an excel file. When I run the test, only the first line of values is identified and can be run within the automated test. Not sure if this is expected with the testing product (www.badboy.co.au), but just wondered if there was a way of preventing excel from taking control of the .csv file? Any helpfull feedback would be great.

    Read the article

  • Search for files which will open a certain application in Mac OS X

    - by Jacob Palme
    In Mac OS X, when you doubleclick on a file name, that file will open with the application which created the file. So there must be stored, somewhere in the file description on a Mac OS X file, information on which application created this file. Note that this is not the file extension, the file can have any extension, or no extension at all. Two questions regarding this information: (1) How can I search for all files which will open a specific application? (2) How can I see, and change, the application which a certain file will open?

    Read the article

< Previous Page | 153 154 155 156 157 158 159 160 161 162 163 164  | Next Page >