Search Results

Search found 38500 results on 1540 pages for 'ini files'.

Page 243/1540 | < Previous Page | 239 240 241 242 243 244 245 246 247 248 249 250  | Next Page >

  • Is there a good, free way to fix broken/corrupt .wmv files?

    - by chbtn
    I've recovered some files from an hdd that weren't supposed to be deleted in the first place, but they have seeking problems/crash the players. Since they have the right size, I'm thinking it might be a problem of corrupt index/header, so I'm trying to find a way to fix them. It's easy to find examples on how to fix corrupt .avi files with mencoder, but .wmv seems trickier. Also, I realize there might not be a way to fix these files, but I figure I might as well as try. As far as players go, I've tried opening it with vlc/mplayer/windows media player. I can use anything on Windows XP/7 and Ubuntu, as long as it's free. Since the files are 200mb+ and there are quite a few, I don't think trial software would work.

    Read the article

  • How to recover disk and files after 10.04 boot failure?

    - by K R Jawaharlal
    I have a 1TB HDD with four Windows XP partitions and a 120GB HDD with 10.04. While working in Ubuntu, due to delay and failure to shutdown, I switched off the system. Next it failed to boot in Ubuntu and stopped at initramfs. After that, I tried to repair from the booting stage. By mistake instead of hdd no I used partition no. This damaged the Windows also. Then Windows XP was reloaded and is running. When I boot with 12.04, it is able to detect the 120GB HDD, but, it is unable to mount. I am unable to access the files. I would like to revive the disk and recover files. Would appreciate any help.

    Read the article

  • In Windows XP Professional, is there a limit on the number of files that can be contained in a single folder? [duplicate]

    - by Andrew
    This question already has an answer here: How many files can a windows folder contain? 1 answer I am running Windows XP Professional, service pack 3. Right now I have 4,398 files in a single folder, and Windows XP seems to read it fine. How many more files can I place in this same folder, either theoretically or practically? Thanks for your time.

    Read the article

  • How do I prevent my swf files being hotlinked, downloaded etc.

    - by undefined
    I have swf files that are embedded in a PHP page using SWFObject. These swf files are in the same directory as my PHP files. for example www.myurl.com/index.php embeds www.myurl.com/flashfile.swf, index.php and flashfile.swf are in the same directory. However I want to prevent people from being able to type in www.myurl.com/flashfile.swf and viewing the swf. I want the browser to deny access to this file unless it has been embedded by the PHP file. Should I move my swfs to another folder and protect this folder somehow - is this with the .htaccess file? I am running Apache on a linux machine. While my main concern is for swf files I would like to protect graphics used on the site too. all help appreciated thanks

    Read the article

  • When I move or delete files, they occasionally appear right back where they were a day later

    - by Shane Nault
    When I move or delete files, they occasionally appear right back where they were a day later. Also, when I move files, they not only re-appear where I moved them from, they are also where I moved them to as well. Is my system duplicating files? It seems so. This problem has been occurring randomly, but with increasing regularity. Most of the issues are with downloaded music and image files, but it has happened occasionally with Word documents as well. Also, some of my desktop icons have reappeared after I have moved or deleted them. I have run extensive system scans for malware, viruses and the sort but nothing pops up. Is there something wrong with my settings or is there another problem? I use Windows 7, SP 1 with all current updates, and have a few other issues. I was worried that perhaps some of my settings were incorrect.

    Read the article

  • Is there any way to open more than one folders(Home folder, videps etc..) by clicking the 'files' icon on the unity bar?

    - by 001neeraj
    Why only one nautilus window manager is opens when clicking on the 'Files' icon on the unity bar? I need to open two or more HomeFolder windows by clicking on the 'files' icon on the unity pane. I cannot create more than one nautilus window manager unless i tap superkey(windows key) and type homefolder,videos etc or typing the same in gnome-do. Is there any way to open more than one folders(Home folder, videos etc..) by clicking the 'files' icon on the unity bar??

    Read the article

  • Can create and delete files through samba, but not modify?

    - by PoppySeeds
    I'm connecting from Windows 7 to Ubuntu Server 10.04 using samba. I can create new files in the share, delete those files, but if I try to edit a file, and then save it, I get an access denied. My User in Windows is the same as the one I'm connecting as and as setup in Ubuntu. This is very odd. I also have trouble copying files to Ubuntu as well. It errors but lets me copy them anyway.

    Read the article

  • Is there a limit of max. number of files in external hard drive folder?

    - by tfs
    I have a FAT32 external hard drive where I keep backups downloaded from webserver. I have a directory with 30 subdirectories. One of the subdirectories contains 21381 files and when I try to copy more files into this directory I get 0x80070052 error. However,it's possible to copy one more file in this directory (only one) if I make it's name shorter (8 characters instead of 22 as it's original name). How do I solve this problem? Now I can not synchronize external hard disk files with server files which is very important for me.

    Read the article

  • Do any database "styles" use discrete files for their tables?

    - by Brad
    I've been talking to some people at work who believe some versions of a database store their data in discrete tables. That is to say you might open up a folder and see one file for each table in the database then several other supporting files. They do not have a lot of experience with databases but I have only been working with them for a little over a half year so I am not a canonical source of info either. I've been touting the benefits of SQL Server over Access (and before this, Access over Excel. Great strides have been made :) ). But, other people were of the impression that the/one of the the benefit(s) of using SQL Server over Access was that all the data was not consolidated down into one file. Yet, SQL Server packs everything into a single .mdf file (plus the log file). My question is, is there an RDBMS which holds it's data in multiple discrete files instead of one master file? And if the answer is yes, why do it one way over the other?

    Read the article

  • OS X - Automatically Set Execute Permissions for New Files?

    - by i help X u
    I'm using OS X 10.6.4 and am trying to set a folder to automatically enable execute permissions on new script files copied or created in a directory. I have used Sandbox 2 to set every permission for the folder to enabled with sticky bits and the inherit flag set but I still have to manually set the execute flag using chmod for every new flag. I've done: chmod -R a+rwxs ~/scripts I've done: chmod 7777 ~/scripts And the permissions for the folder show as: drwsrwsrwt+ for the folder. But if I add a new script file it's set to "-rw-r--r--+" (the default) I looked at setting "unmask 000" in the .profile file but the default value for files is 666 with an unmask of 022 so that's not relevant since I would need a default value of 777 for files. I have figure out how to use chmod in an AppleScript triggered by a folder action to automate this but I'm wondering if there is a simple ACL or chmod setting I'm missing. So, is there a way to automatically set execute permission for new files? (Without using a folder action and AppleScript?)

    Read the article

  • Can't boot ubuntu 12.04, stuck in busybox. Can't view files from ubuntu trial disc, or windows partition

    - by Maura
    So, I'm slightly computer literate, and find myself frustrated and overwhelmed. My computer is a acer laptop, extensa 5620-6572. I have a dual boot with windows vista and ubuntu 12.04. The ubuntu 12.04 I got was from an upgrade, not a disc. I tried to load ubuntu 12.04, and it gets stuck in the "busybox", and I don't know how to proceed from there. I went to my windows partition and downloaded Ex2 from http://sourceforge.net/projects/ext2fsd/ and thought I'd try access my files and save them to a external HD. Then when I restarted my computer and went to windows, it always freezes after it loads the OS. So then, I downloaded and burned a ubuntu 12.04 boot disk, and the disk works fine. But I still can't figure out how to view files on my harddrive.

    Read the article

  • How do I prevent lighttpd from caching static files, even when modified on disk?

    - by Pixelastic
    I am using lighttpd to serve static files. I have a bunch of images in a dir that I regularly update. This will change the file content (and filesize) as well as the modification date, but not their filename. When I access the files through http, the updates are not taken into account and lighty serves the old file. I can manually rename the file to something different, then lighttpd will return a 404 error, and if I rename my file back, I will get the correct updated version. Seems like lightty is using some kind of cache mechanism of its own (which is fine) to return static files. Unfortunatly, it seems that this mechanism doesn't update itself when files are modified. I checked through Wireshark, and my browser is really doing a request to the file, this is not a browser caching issue. It returns a 200 OK when requesting it from an empty cache, and a 304 Not Modified otherwise, as expected. But the file is returned with a wrong Last-Modified header that do not reflect the real last modification date. Maybe there is some config directive that I am not aware of ? I would like the files returned by lighty to reflect the changes made on disk directly, or at least being able to invalidate its cache.

    Read the article

  • Join multiple consecutive SQLite database dump files into 1 common database? Purpose: Search through ENTIRE Chrome Browsing History

    - by porg
    Google Chrome 's default web browsing history search engine only lets you access the records of the recent 100 days. Nevertheless in your application data, Chrome keeps your entire browsing history in SQLite database files, with the file naming scheme of "History Index YYYY-MM". I am looking for a way to search… …through my entire browsing history, …with sophisticated filters (limit search terms to certain fields such as URL, domain, title, body text; wildcard or regex terms, date ranges). … in … …either some ready-made software. eHistory came close, as it can limit terms to fields, but it lacks wildcards/regexes, and has the same limited time horizon as the default search. Beyond that, I could not find any suited Chrome extension or standalone (Mac) app. …or a command line to join multiple SQLite database files into one database, which I can then query (with the full syntax power). In the spirit of the pseudo code below: Preferred this way: sqlite --targetDatabase ChromeHistoryAll --importFiles /path/to/ChromeAppData/History\ Index* --importOnlyYetUnknownFiles Or if my desired feature --importOnlyYetUnknownFiles is not possible (feature could also be called "avoid duplicate imports by checking UIDs"), then by explicitly only importing files, of which I know, that they have yet not been imported into the ChromeHistoryAll database: cd ChromeAppData; sqlite --databaseTarget ChromeHistoryAll --importFiles YetNotImported1 YetNotImported2 YetNotImported3 All my queries I would then perform in the database "ChromeHistoryAll" P.S.: Additional question of general interest: Is there a way to perform a database query in a temporary database which was created on-the-fly from multiple files? Like: sqlite --query="SQL query" --targetDatabase DbAll --DBtemporaryInRAM --importFiles db1 db2 db3 This is surely not applicable for my Chrome question, as these History Index files have a combined file size of 500MB together, thus such a query would be of bad performance. But it could come handy in other situations.

    Read the article

  • How do I fully clear Firefox's cache of CSS and JS files?

    - by Mike Webb
    I work on a website at my work. The issue is that if I visit the site, which uses the cached versions of the CSS and JS files, and then upload an updated copy of a CSS/JS file, Firefox will still use the cached version. I can go to 'Tools-Clear Recent History' and clear the Cache of "Everything" and it still uses the cached version of the files. It will eventually updated and use the new files, but it can takes hours for this change to occur. So, how do I completely clear Firefox's cache of these files?

    Read the article

  • How do I configure Gnome 3 so that it doesn't pop up a dialog for 'open with files' when I mount a drive?

    - by michael
    I am running Gnome 3 on Ubuntu 11.10. In the file manager, when I click a drive under 'Devices', Gnome 3 always pops up a dialog with the choices 'open with files' and 'eject' and then I need to click 'open with files' to get rid of that dialog. Is there a way to configure Gnome 3 not to do that? I am in file manager already, clicking a drive should show the content in the right pane. Why does it still ask me to 'open with files'?

    Read the article

  • What editor/viewer to use to inspect large text based files?

    - by Turismo
    Are there any text editors/viewers (preferably on windows but other platforms are also ok) that can handle files of 500 MB or more? The editors I checked so far (Notepad++, Notepad, Eclipse) all choked on files of that size. Edit: Many thanks for the great suggestions. I tried gvim as it was the top voted and was available on Windows. I opened the file in a reasonable time. After that scrolling and searching was very smooth as long as syntax highlighting was turned off. From the other editors mentioned TextPad and EmEditor both claim to be able to handle large files very well. EmEditor seems to be built exactly for editing large files. I'll probably try both and report back.

    Read the article

  • I installed XAMPP on Ubuntu, and its running without any of its files.

    - by CDeanMartin
    I installed XAMPP on Ubuntu, and its running without any of its files. Well, not really, but it sure seems that way. I followed the directions on the Apache Friends website to the letter. XAMPP works fine, it's sample apps run like a charm. But I have no idea how it is running. The installation should have created an otf folder, or a XAMPP folder, or an htdoc file according to the tutorials. But there is no trace of any of these files or folders anywhere in the file browser. Are there any linux utilities as good as the Windows Explorer that can find my missing files? I particularly need to find the var/www folder to put my .php files in.

    Read the article

  • (How) does deleting open files on Linux and a FAT file system work?

    - by lxgr
    It's clear to me how deleting open files works on filesystems that use inodes - unlink() just decreases the link count to zero, and when the last file handle to the file is closed, the inode will be removed. But how does it work when using a file system that doesn't use inodes, like FAT32, with Linux? Some experiments suggest that deleting open files is still possible (unlike on Windows, where the unlink call wouldn't succeed), but what happens when the file system is uncleanly unmounted? How does Linux mark the files as unlinked, when the file system itself doesn't support such an operation? Is the directory entry just deleted, but retained in memory (that would guarantee deletion after unmounting in any case, but would leave the file system in an inconsistent state), or will the deletion only be marked in memory, and written at the time the last file handle is closed, avoiding possible corruption, but restoring the deleted files after an unclean unmount?

    Read the article

  • Add closure after concatenating files or keep one for each separate file?

    - by sdespolit
    At the moment every javasrcipt file in our project includes closure wrapper. Before deploying all files are concatenated and minimized. And i'm asking myself, do we really need all that closures OR one closure added after concatenation will do the job. I can see only one reason not to do that - the variable we are sharing between files will get in the global namespace. But there is no conflict so is don't see any problem. The real question is what other cons may be? Or more generally what approach is better - one closure or many?

    Read the article

  • Files: Name column basically useless in list view. Feature or bug?

    - by Luksurious
    So the following is happening in Files whenever the width of the window is somewhat smaller than all the content of the list: In list view the name column is cropped unto a point where no name is visible at all! And it is even not possible to change the column size manually. Funnily, in some situations it quickly flickers from a larger column width to the small one back and forth before settling on the small size. Unnecessarily to say, it is extremely annoying. Is there a way around this or is this just "bad design"? Oh yeah, Ubuntu 13.04 & 13.10, Files version 3.8.2.

    Read the article

  • How to copy a folder with many files with integrity check?

    - by RafaelM
    I just got a new hard drive and I want to move many of the folders from the old hard drive to the new one, but I want to make sure everything is copied over correctly. I tried using md5summer to generate md5 sums of the original files, copied files over and then tried to compare md5 sums of both sets of files. This took ages because there are many large video files. Is there any software I can use to make this process as painless as possible? I just need basic file integrity checking. Thanks in advance

    Read the article

  • How can I decrypt encrypted files using a PEM private key?

    - by Phil Cole
    I have files which have either been encrypted with a public key and the Blowfish algorithm, or a public key and the AES-256 algorithm. I'm looking to put together a Perl script that would be able to use the private keys (which I do have) to decrypt the files. The public and private key files are all in PEM format, and while I can find ways of reading the PEM files, and ways of decrypting data with a key, I haven't yet found a way of going from PEM - key. Any suggestions?

    Read the article

  • How do I convert Word files to PDF programmatically?

    - by Shaul
    I have found several open-source/freeware programs that allow you to convert .doc files to .pdf files, but they're all of the application/printer driver variety, with no SDK attached. I have found several programs that do have an SDK allowing you to convert .doc files to .pdf files, but they're all of the proprietary type, $2,000 a license or thereabouts. Does anyone know of any clean, inexpensive (preferably free) programmatic solution to my problem, using C# or VB.NET? Thanks!

    Read the article

< Previous Page | 239 240 241 242 243 244 245 246 247 248 249 250  | Next Page >