Search Results

Search found 89530 results on 3582 pages for 'file read'.

Page 173/3582 | < Previous Page | 169 170 171 172 173 174 175 176 177 178 179 180  | Next Page >

  • ffmpeg: cut multiple input files with seeking to one output file

    - by Josef Kufner
    I have list of video files (loaded from database), each with start and end time of requested interval: # file begin end v1.mp4 1:01 2:01 v2.mp4 3:02 3:32 v3.mp4 2:03 5:23 And I need to create single video file containing these intervals: [0:00]---v1---[2:00]---v2---[2:30]---v3---[5:50] I preffer usig ffmpeg, since it is installed on server. Caller program is written in PHP. It is easy to cut one input to one output (argument escaping removed for clarity): exec("ffmpeg -ss $begin -i $input_file -ss $begin -c copy $output_file"); I there any easier way than executing ffmpeg for each interval and then execute it once more to concatenate prepared clips together? I really do not like to have a lot of temporary files or dealing with complex process handling.

    Read the article

  • batch copy files with error log on missing permissions

    - by sc911
    Hi *, I'm searching for a tool to batch-copy files, that should support the following points: copy files from a net-share report any errors show errors only or filter log on errors don't stop on an error also report if a file or a folder could not be copied due to missing permissions if possible it should have a queue where new job can be added while copying I tried the following tools: TerraCopy: takes a lot time to just calculate the time and the size of the job and does not report errors due to missing permissions (it doesn't even add those files to the copy-queue) Karne's replicator: does not report errors due to missing permissions xcopy: does a great job when using the right parameters and piping the output to a file (in the German localization xcopy /k /r /e /i /s /c /h SOURCE TARGET>LOGFILE 2>&1 will do the job. opening the logfile in IE will give you a great monitor). but quing jobs it not possible (ok, you can join them all in a batch-file, but you can not queue jobs while another one is running (hm, thinking of a batch-script that loops through a file with the source-target-config...)) to be continued Which tools do you use? Tell me! Thx sc911

    Read the article

  • How to convert an XFS file system to HFS+

    - by user219350
    I have repeatedly convinced of the reliability of the XFS file system , and I was more than satisfied . I was happy with everything in Ubuntu 14.04 ( great software) , but there is a little "but ! " Basically, I work on OSX-Mavericks 10.9.3, which sees very Windows 8.1 and works wonders with NTFS, but does not see Ubuntu! Briefly describe the equipment: ASRock B75 Pro3-M i5 3330 GeForce GTX 650 Ti SATA 500GB running OS X Mavericks + Clover - a boot disk Toshiba 2TB running Windows 8.1 (x64) and Ubuntu 14.04 (amd64) If you boot from the Toshiba (where there is Ubuntu and boot Windows + GRUB) after restart boot from Clover, it is impossible. Tried a lot of options - as Clover installation and boot priority, and various settings for GRUB, but have not found an acceptable option and have no desire to reinstall again Clover (Mavericks reboots 20 seconds - excellent!) So please help on the file system - how to convert from XFS to HFS+ journaled. Mavericks to saw it all synced on Mac. Thank you for the sensible answer and help! Originally in Russian.

    Read the article

  • Mavericks permission issues with Windows Server deduplicated shares

    - by dmohlmaster
    We have a number of 10.9-10.9.3 - Mavericks - machines installed throughout our facility. Much of the user content is pulled from shares stored on our Windows Server 2012 fileservers with deduplication enabled. I have found that files newly written or unoptimized are able to be accessed without issue - read, written, modified, etc. Once the file gets optomized/deduplicated and Windows adds the P & L attributes - sparse and symlink - the Macs running Mavericks begin to have access issues. Once the files get deduplicated, users begin receiving read access errors when copying files (see error1 below). This happens when copying to folders within the current folder tree or copying somewhere to the local system. If you 'stop' the copy operation and retry a few more times, it may eventually work for the specific instance but fail again later. I am however, able to copy these files without issue via the terminal. Other systems running 10.7 do not experience the same issues and are able to access file server resources without issue. Many of the systems having issues are newer and thus not able to be downgraded to 10.8 or 10.7. I have tried finder replacements such as Pathfinder but the results are the same. I know this is at least similar to the issues many Mac users are already experiencing and posting about but I haven't seen it directly linked to deduplication and the attributes written by Windows server. Has anyone seen this issue? Have any solutions been found? Error 1: When copying files after the PL attributes have been set by deduplication. "One or more items can't be copied to "Foler" because you don't have permissions to read them. ******************************************' Via the system.log, I am also seeing the following error when accessing these deduplicated file shares. The reparse point tag listed below is "IO_REPARSE_TAG_DEDUP" Reported error: "smbfs_nget: filename.ext - unknown reparse point tag 0x80000013"

    Read the article

  • Windows Server - share files without access for administrator

    - by Pawel
    We have a MS Windows Server 2008 R8 based server that is administrated by our IT department. We would like to achieve two things simultaneously: A folder on the server, containing several thousand files (new files added frequently) that is accessible to some ActiveDirectory users (e.g. board of directors) but is not accessible by IT department employees IT department employees still maintain rights to administrate the server, including installing new software and services We already checked some solutions: Using NTFS access rights. Unfortunately IT (members of "Administrators" group) can set themselves as new owners of the files and change the permissions so that they gain access to the files. Enabling EFS. Unfortunately even if you do not allow IT to access files, they still can disable EFS completely because they have administrative rights. Moreover as far as I know you have to manually add permissions for all users but the owner for each new file - very inconvenient. Creating a new role for the IT department that has all the privileges apart from taking ownership of files. Unfortunately if you're not a member of the Administrators group, you cannot install new software, no matter what privileges you add to the role. TrueCrypt - nice free encryption software, but with poor sharing capabilities. You can either mount an encryption container on the server (and then IT has access to its contents) or you mount them locally but only one user can mount it for writing. AxCrypt - free encryption software that enables file-by-file encryption on the server. There are some disadvantages though - you have to manually encrypt each new file added. The files have their extensions changes. You can only set one password for all files (so all users have to know this one password). Any other ideas? Our budget is limited so enterprise-class software from Symantec or PGP would probably be not an option.

    Read the article

  • Locate a user's bashrc file

    - by Starkers
    Really confused. Upon running cat etc/passwd I have found this: postgres:x:117:126:PostgreSQL administrator,,,:/var/lib/postgresql:/bin/bash meaning I have a postgres user, right? I want to change the bashrc environment file of this user to make commands available to it. /var/lib/postgresql doesn't contain a bashrc file, and /bin/bash doesn't contain it either, so I don't really know what's going on. All I know is a created postgres using the useradd command, so why do I have some weird user with no home directory? So confused :(

    Read the article

  • Converting PDF eBooks into a Kindle format

    - by Ender
    Over the past couple of years I've amassed quite a collection of guides, tutorials and ebooks in PDF format. A lot of these are quite useful for work, especially PDF documentation, and rather than have to be at a computer every time I want to read how to do something in Sitecore or to read through a software testing ebook I'd like to do it on my brand-spanking-new Kindle. However, even though there is now a native PDF reader on the Kindle due to the nature of PDF's they are practically unreadable. The text doesn't wrap due to how PDF's are sized and so far after a bunch of Google searches I've yet to find a viable solution to get my PDF's converted into a readable Kindle format. Sometimes these books have code or pictures/tables in them, but most of the time they're text-heavy and to be honest I'd be surprised if there wasn't a free tool to handle the converting of PDF to one of the (seemingly many) Kindle formats. So, can anyone help me out with this? EDIT: I've tried Calibre, and have checked their forums to play with some of the advanced settings, yet the solutions available seem to be extremely poor, especially if the book you're attempting to read contains equations, code, or anything outside of plain text. I've also tried Amazon's conversion service, which wasn't much help with such documents. The best way I have found so far is to build the entire thing over again in ePub or RTF format and convert to MOBI from there. This works for text-heavy books with tables, but anything technical still isn't covered. Can anyone help with this?

    Read the article

  • Where is the start up file located?

    - by starcorn
    Hello, I want to add some lines which should execute every time Ubuntu boots up, so I don't have to change them manually everytime. I've read in some place that you should edit this file /etc/rc.local. However when I add the lines I want to execute at start up it doesn't run those lines. So I wonder where the start up file is located in ubuntu? Those lines I want to add is to change the sensitivity for the trackpoint One of the lines I want to add: echo -n 250 > /sys/devices/platform/i8042/serio1/serio2/sensitivity

    Read the article

  • Best alternative of Property file in Java

    - by Ranna
    Hey I an working on the product which is live at multiple portals. The product is developed in GWT, JAVA, Hibernate. My question is : Whether there is any alternative of using property file in java. My Requirement : For one property key there are multiple values in live portal for each different portal. Each time I change property file, I need to make the war again. The loading of any of the property should not be time-consuming. Any help or suggession would be apprecialble !!!

    Read the article

  • Synchronize two directories on linux pc

    - by Gab
    I need a distributed filesystem (or a synchronization tool) that is capable of keeping a directory synchronized across 4 pc. My requirements are: offline access (data must be available offline on each pc) preserve execution rights: some files are marked executable on a linux partition. This flag should be replicated. efficient sync strategy: some of my files are 20GB, they are changed quite often, but in very little parts (Virtualbox images). Delta transmissions are welcome. efficient handling of space: no history for files, files shouldn't be copied to temp directories "just in case you break it". it must propagate deletions of files modification can happen in any of the 4 pcs, they should be propagated when other pc are connected. Other specs of my solution are: Sync is over a lan, the total amount of data to be synced is around 180GB, in some ten thousand files. Changes are small, but can happen in big files. At the moment i'm interested in a linux only solution. conflicts either don't happen or are solved with "last one wins" I haven't found any good solution. I've been trying: unison: it is the only one working at the moment, but during the hashing phase it hangs my pc for some minute, disk light steady on. Sparkleshare doesn't handle large files nicely. It keeps an history of all your changes that grows up indefinitely. They promise it will be fixed in next releases, but at the moment it still doesn't fit my needs. owncloud (keeps history of each file i change) coda ? (help! i couldn't set it up correctly!) git-annex assistant transforms all your files in symlinks and mark the original file as read only ("just in case you make a mistake while you modify it"!). Before you edit a file you have to issue a special command "git-annex unlock", that creates a local copy of the file, and you have to remember to lock it again if you want it synchronized. What to try next?

    Read the article

  • File access forbidden in htpasswd

    - by Nerd-Herd
    I have been using the htpasswd generated in this question and it seemed to have been working well until recently. Since yesterday, I am not able to access the newest file created in the folder ChatLogs(named 10_07_2012.txt). The server returns a 403 Forbidden error saying: Forbidden You don't have permission to access /ChatLogs/2012/07/10_07_2012.txt on this server. I am still able to access older files(until 09 July, 2012). At first I thought it might be because of file permissions, but they are the same as on other 9 files in the folder. What could be the problem? Please Help.

    Read the article

  • Strategies for removing register_globals from a file

    - by Jonathan Rich
    I have a file (or rather, a list of about 100 files) in my website's repository that is still requiring the use of register_globals and other nastiness (like custom error reporting, etc) because the code is so bad, throws notices, and is 100% procedural with few subroutines. We want to move to PHP 5.4 (and eventually 5.5) this year, but can't until we can port these files over, clean them up, etc. The average file length is about 1000 lines. I've already cleaned up a few of the low-hanging fruit, however the job took almost an entire day for 2 300-500 line files. I am in a quagmire here (giggity). Anyway, has anyone else dealt with this in the past? Are there any strategies besides tracing backwards through the code? Most static analysis tools don't look at code outside of functions - are there any that will look at the procedural code and help find at least some of the problems?

    Read the article

  • Guide to installing a fully encrypted file system?

    - by Michael Stum
    I have a little Netbook on which I want to install Ubuntu 10.10 (32-Bit) on. However, since it is a portable PC I want to completely encrypt the file system (in case of theft). Currently it runs Windows 7 Starter and I use TrueCrypt which installs a custom boot loader that asks for the password. I remember from the past that Linux can do that as well by putting /boot on it's own, unencrypted partition. Since it's been ages since I last worked with file system encryption (I remember setting up LVM and a custom patched grub to ask for the password) I wonder how that would work nowadays and if there is a step-by-step how-to for it?

    Read the article

  • Website .htaccess file for Wordpress sub folder

    - by ubique
    I developed a Flash website for a client and added the following .htaccess file in the root directory and the non-www to www redirect works perfectly. RewriteEngine On RewriteCond %{HTTP_HOST} ^website.com [NC] RewriteRule ^(.*)$ http://www.website.com/$1 [L,R=301] I was also asked to add a Wordpress blog so I put it in a new directory folder (as opposed to a sub domain) with so the URL is www.website.com/blog Does Google now see the main site and blog as two different websites? Do I need to link them together using another .htaccess file in the Wordpress Root so Google automatically crawls the whole domain? Any help appreciated....

    Read the article

  • HFS+/How to convert an XFS file system to HFS+ [on hold]

    - by user219350
    I have repeatedly convinced of the reliability of the XFS file system , and I was more than satisfied . I was happy with everything in Ubuntu 14.04 ( great software) , but there is a little "but ! " Basically, I work on OSX-Mavericks 10.9.3, which sees very Windows 8.1 and works wonders with NTFS, but Ubuntu does not see! Briefly describe the equipment: ASRock B75 Pro3-M i5 3330 GeForce GTX 650 Ti SATA 500GB running OS X Mavericks + Clover - a boot disk Toshiba 2TB running Windows 8.1 (x64) and Ubuntu 14.04 (amd64) If you boot from the Toshiba (where there is Ubuntu and boot Windows + GRAB) after restart boot from Clover, it is impossible. Tried a lot of options - as Clover installation and boot priority, and various settings Grab, but have not found an acceptable option and reinstall again Clover - no desire ( Mavericks 20 seconds reboots - excellent !) So please help on the file system - how to convert to XFS HFS + magazine . Mavericks to saw it all synced on Mac . Thank you for the sensible answer and help! Originally in Russian.

    Read the article

  • Using linq to parse file [closed]

    - by Emaan Abdul majeed
    i am working parsing textfile using LINQ but got struc on it,its going outof range exception string[] lines = File.ReadAllLines(input); var t1 = lines .Where(l => !l.StartsWith("#")) .Select(l => l.Split(' ')) .Select(items => String.Format("{0}{1}{2}", items[1].PadRight(32), //items[1].PadRight(16) items[2].PadRight(32), items[3].PadRight(32))); var t2 = t1 .Select(l => l.ToUpper()); foreach (var t in t2) Console.WriteLine(t); and file is about 200 to 500 lines and i want to extract specific information so i need to split that information to different structure so how to do it this..

    Read the article

  • Getting data from csv file and returning objects in collections

    - by Jacob
    I have very simple class of person as below: public class Person { int ID; Gender gender; Date dateOfBirth; public Person(final int iD, final Gender gender,final Date dateOfBirth) { ID = iD; this.gender = gender; this.dateOfBirth = dateOfBirth; } } Gender is enum : public enum Gender { Male, Female } In CSV file i will have data, for example: 1;Male;23-02-2001 2;Female;11-06-1989 3;Male;02-12-1999 Is in java any simple way to get all persons from csv file and return it as ArrayList<Person> persons ?

    Read the article

  • How to find what files / directories are not copied yet?

    - by user8676
    Hi all, I found the following 'nice' situation: An archive of few disks (actually three disks) which has a bunch of photos (more or less) organized. Well, this is good. A big disk shared on a network which has a bunch of photos which has another folder structure (even if is somewhat recognizable for a human being) than the archive described above, but some of the files on this big network share are the same with the files from the archive. Well, this is bad. What we need is to move the different (new) files from the network share in the archive (perhaps we'll use for this a new disk added to archive). The program that we need is different from a regular File Duplicate Finder program because usually the File Duplicate Finder finds the duplicates from all sources comparing each file with another. We want to find the differences between the two sources. It is fine for us to have a report generated in text file which after this we'll use to do our move. A Windows solution will be preferred. Any ideas? TIA

    Read the article

  • Address search from large text file

    - by user73809
    Basically I want to develop a Address lookup(part of my project) using C# (and I can use SQL if necessary). I have a very large text file which have all the UK address and postcodes. Addresses needs to be looked up from this text file. For example, as soon as a user type in the first letter of the postcode it should list all the postcodes starting with that letter and as they type in further, the list should narrow down. Could you guys please how I can do this search very fast and list the results in a efficient way. If you guys know any methods and technologies pls do advice.

    Read the article

  • Configure DNSMASQ to use /etc/hosts file

    - by casey_miller
    I have installed DNSMASQ but it was not starting as 53 port was busy. I found out that ubuntu already had dnsmasq package and it's working. Now the problem is...I just want to be able to resolve my hosts in /etc/hosts through it i.e: nslookup somehostonlan localhost to be resolved to certain IP taken from /etc/hosts file. but this is not happening. Why? BTW, as the caching DNS server it's working fine. I just want it to resolve hosts from /etc/hosts file that's all

    Read the article

  • How to optirun .sh file?

    - by Min
    I managed to get as far as optirun glxgear, which worked perfectly, but I am unable to define settings for a specific shell script. In its wiki it says this: optirun [options] <application> [application-parameters] and I don't know what option has to be nor the application-parameter. (I'm really new) If my file is /home/hon/hon.sh, how can I execute it via terminal? If I just type optirun hon or optirun hon.sh it shows this error: /usr/bin/vglrun: 296: exec: hon: not found there I try to find the file /usr/bin/vglrun. How can I run the shell scripts I need to run?

    Read the article

  • apt-get could not open lock file

    - by user114373
    I am trying to get an nfs client running on a Sheeva-plug running debian 2.6.22. The host is Ubuntu 12.04 and claims (from showmount -e) to be exporting the desired directory. There is no showmount binary in the sheeva-plug, so I'm trying to install it from the nfs-common package: # apt-get install nfs-common The response ends with E: could not open lock file /var/cache/apt/archives/lock - open (no such file or directory) E: Unable to lock the download directory. I am root while doing this. Similar errors arise when trying to install other packages. How do I correct these errors so apt-get will do its work?

    Read the article

< Previous Page | 169 170 171 172 173 174 175 176 177 178 179 180  | Next Page >