Search Results

Search found 74849 results on 2994 pages for 'file folder'.

Page 47/2994 | < Previous Page | 43 44 45 46 47 48 49 50 51 52 53 54  | Next Page >

  • The /usr/bin/x11 folder

    - by WrecklessT
    Why does the /usr/bin/x11 folder hold another x11 folder and when you open that === x11 you get another x11 and then another and another??? I did it about 6 times and got frustrated so I have no idea how deep this goes but What is the purpose (or is it a glich?)? Is this eating disk space more than it should? (I was going to delete one or more but thought I better ask first before I end up !fucked ) and finally Is it just me and I am going crazy maybe the pills are wearing off or maybe they are kicking in.. I am not sure but an answer would help. I can see no reason why this should be happening at all.

    Read the article

  • Cannot see shared folder in /mnt/hgfs

    - by blasto
    I am trying to share a folder between Lubuntu 13.04 (in VMware player) and Windows 7 64 bit. I followed a tutorial till step 16. I typed a command and saw nothing. I also went into the /mnt/hgfs folder and saw nothing there. How do I fix this ? http://theholmesoffice.com/how-to-share-folders-between-windows-and-ubuntu-using-vmware-player/ Command - dir /mnt/hgfs EXTRAS - By the way, this is how I actually reached step 16. Step 12 - sudo apt-get install hgfsclient Step 14 - If it does not work, then follow this tutorial - http://www.liberiangeek.net/2013/03/how-to-quickly-install-vmware-tools-in-ubuntu-13-04-raring-ringtail/ Step 16 - STUCK !!!

    Read the article

  • Move Joomla website to new folder

    - by Jon
    I currently have a website. I have created a new folder on the website called V2. Under this folder I have installed Joomla and configured my new looking site. I now want to make V2 the default website. I could point the website to that V2 directory however I have other folders under the current root website that I need to keep. How can I transfer V2 to the root of my website? Is it just a case of copying all the files?

    Read the article

  • Hide folder names or such?

    - by Miller
    Okay, I have a cpanel account with unmetered everything (pay a bit per month), so I wanna host my forum on it etc I have the domains as lets say money.com wordpress.com forum.com As I'll have to put everything in different folders for instance money will be in /m/ wordpress /w/ and forum in /forum/ or something. What I'm saying is, how do I hide the file so it'll look like money.com/m/ is actually money.com ?? I need to hide the folder name the contents are in so I can host multiple sites, therefore the site will look like its the only site on the host so I don't have to add a redirect for it to direct it to the folder? Thanks guys, been trying for a while!

    Read the article

  • Hiding images in folder without renaming or moving the files

    - by Marcus
    I'm dual booting Ubuntu with Windows and I have all my music on a separate harddrive. In Windows the album art is hidden by default but when I access the folder in Ubuntu there is two album artwork files for every mp3, one small and one large. I would like to hide those images without having to rename them with a dot before or moving them to some other folder becuase then the album artwork would dissapear in Windows. Is there a way to make a .hidden file which hides all images or any other way which hides all images in nautilus?

    Read the article

  • Unity: Nautilus folder opens in background

    - by JJD
    I am used to open a folder from the keyboard with a shortcut. The alternatives are: Use Unity Dash with the shortcut SUPER + F to open the Dash File Lens Open Kupfer with the configured shortcut (CTRL + ALT + SPACE, in my case) Then I type the name of the folder and press ENTER if it is selected. However, the window that opens up is not active but in the background. I have to press ALT + TAB to bring it into foreground. I use Ubuntu Precise (12.04.). Question: How can I configure Ubuntu, Unity, Nautilus so newly opened windows become foreground immediately? Edit: This behavior only happens when another application is already open such as GEdit. It does not happen when the only other selectable entity is the Desktop.

    Read the article

  • Issues With IIS Hosting Two Domains From Same Folder [closed]

    - by Bob Mc
    I have two different domain names that resolve to the same ASP.Net site. Both domains are hosted on the same server, which runs Windows Server 2003 and IIS6. The sites are differentiated in IIS Manager using host headers. However, both of the sites point to the same folder on the local drive for the site's page files. I am occasionally experiencing an ASP.Net error that says "The state information is invalid for this page and might be corrupted." I'm the site developer so I've addressed all the relevant code-related causes for this issue. However, I was wondering whether having two domains/sites sharing the same folder for an ASP.Net application might be causing this intermittent error. Also, is this generally a bad practice? Should I make separate, duplicate folders for each of the domains? Seems like that can become a maintenance headache.

    Read the article

  • How to Create a Folder from Selected Files in Windows

    - by Lori Kaufman
    We’ve previously written about a tool that allows you to create a bunch of folders at one time from a list of words or phrases. However, what if you want to create one or more folders from a bunch of selected text files? There’s a simple, free tool, called Files 2 Folder, that allows you to do that. Installing Files 2 Folder adds an option to the context menu for Windows Explorer. Simply extract the .zip file you downloaded (see the link at the end of this article). Right-click on the Files2Folder.exe file and select Run as administrator. If the User Account Control dialog box displays, click Yes to continue. 8 Deadly Commands You Should Never Run on Linux 14 Special Google Searches That Show Instant Answers How To Create a Customized Windows 7 Installation Disc With Integrated Updates

    Read the article

  • Clone User Profile with Encrypted Home Folder

    - by doublerainbow64
    I'd like to clone my user profile on my local machine. I'd like to do so, just to keep all the preferences and not to have to redefine all of them manually. I've tried this, but when I login, it will just return to KDM login screen. BTW I'm using Kubuntu 14.04. What might be interesting: I have encrypted my home folder. Does this make any difference? I've also tried to copy the contents of .kde folder into /etc/skel, but this had no effect. Thank you very much in advance, doublerainbow64

    Read the article

  • share folder in server torrent daemon like in DC(Direct Connect )

    - by alex_bubuker
    So my question: is it possible to share on a server torrent daemon/application a whole folder with films? I had expirience with DC(Direct Connect) applications and it worked well. I selected folder and exec command share == other people stared to download files from my computer. In my expirience with torrent application you should first create torrent file and then people can download file from you. So is there any torrent client/plugin for torrent client that DC-like behaviour become possible? If it is possible at all ... Thank you

    Read the article

  • Run Tests in Folder

    - by Tomas Mysik
    Hi all, today we would like to show you another minor improvement we have prepared for NetBeans 7.2. Today, let's talk a little bit about testing. This minor improvement will be useful especially for users who have a lot of unit tests (it means all of us, right? ;) - just right click on any folder underneath Test Files node and you will notice: The result is as expected - all the tests from the given folder are run: That's all for today, as always, please test it and report all the issues or enhancements you find in NetBeans BugZilla (component php, subcomponent PHPUnit).

    Read the article

  • Copying first 1000 PDF files having single, double quotes in their name to another folder

    - by racer_ace
    I am having this folder with PDFs into it and I need to process 1000 at a time. So I need to move them into another folder, process them and delete them. For this I tried using $ find . -maxdepth 1 -type f |head -1000|xargs cp -t $destdir It gives error on single and double quotes in filename. There are thousands of files and I have no idea how many of them has these quotes in them. Can anyone help me find a solution? And I tried with the -0 option, it did not work

    Read the article

  • What is the ideal file size for a web page? [closed]

    - by Rob
    Possible Duplicate: Is there a maximum size that web pages should be kept under? What is the ideal file size for a web page? Specifically when it comes to image sizes, what's the total file size for a webpage which includes several images. I tend to compress images down as much as possible before it starts to visually lose quality. We run several CMS website's and the clients tend to ask this question a lot! I'd love to hear another view on it.

    Read the article

  • Uploading multiple files asynchronously by blueimp jquery-fileupload

    - by Ryo
    I'm using jQuery File Upload library (http://github.com/blueimp/jQuery-File-Upload), and I've been stuck figuring out how to use the library satisfying the following conditions. The page has multiple file input fields surrounded by a form tag. Users can attach multiple files to each input field All files are sent to a server when the button is clicked, not when files are attached to the input fields. Upload is done asynchronously Say the page has 3 input fields with their name attributes being "file1[]", "file2[]" and "file3[]", the request payload should be like {file1: [ array of files on file1[] ], file2: [ array of files on file2[] ], ...} Here's jsFiddle, it's behaving weird so far in that it sends post request twice and the first one is cancelled. http://jsfiddle.net/BAQtG/24/ The core part of js code looks like this. $(document).ready(function(){ var filesList = [] var elem = $("form") file_upload = elem.fileupload({ formData:{extra:1}, autoUpload: false, fileInput: $("input:file"), }).on("fileuploadadd", function(e, data){ filesList.push(data.files[0]) }); $("button").click(function(){ file_upload.fileupload('send', {files:filesList} ) }) }) Anybody have idea how to get this to work? Updates Now thanks to @CBroe 's comment, the issue that request is sent twice is fixed. However the keys of request parameter is not correctly set. Here's updated jsFiddle. http://jsfiddle.net/BAQtG/27/

    Read the article

  • Download a File in a Batch File

    - by Cristian
    I've never done any scripting on Windows but now I need to write a batch file that downloads a file off the internet (amongst other things). If it was linux I would use wget. Is there a builtin executable that will download a file to a given directory? This needs to run on XP.

    Read the article

  • File transfer from MP3 player to computer

    - by JP
    I own an old 20GB Creative Zen jukebox (all external screws are gone on it LOL) and I would like to transfer files back to my PC to get them on my new iPod. Problem is, no matter what software I use (WinAMP, Creative Media Source or Windows Media Player), it just stops transferring files on my HD with an error message that says there is no more space on the destination folder. Problem is, there is still 320GB free. I tried lot of things like installing newer driver, latest Zen plug-in for Media Source, latest WinAMP version. Sometimes, it just works and then again, it stops working and I get this non-sense error. Restarting my PC sometime solves the issue by giving me enough time to transfer 10 or 15 more files and then I get the error again and again. Yesterday though I managed to transfer up to 3GB of MP3s on my computer before getting the error. Seems like I'm having a driver issue or a weird behavior from the player and/or the software I'm using. 3 different software can't reproduce the exact same issue by themselves so it must be something related to the driver. I can't find any post of any sort concerning such issue on old forums. Any idea?

    Read the article

  • In search of a network file system with extended caching to speed up file access

    - by Brecht Machiels
    I'm running a small home server that stores my documents. The disks in this server are in a RAID 1 configuration (using Linux md) and it's also periodically being backup up to an external hard drive to make sure I don't lose them. However, I'm always accessing the files from other computers on the home network using an SMB share, and this results in a considerable speed penalty (especially when connected over WLAN). This is quite annoying when editing large files, such as digital camera RAWs, for example. I've been looking for a solution to this problem. It would have to offer some kind of local caching to speed up the file access. The client would preferably not keep a copy of all data on the server, as it consists of a very large collection of photographs, most of which I will not access frequently. Instead, it should only cache the accessed files and sync the changes back in the background. Ideally, it would also do some smart read-ahead (cache the files that are in the same directory as the currently opened file, for examples), but I suppose that's asking a bit much. Synchronization should be automatic (on file change). Conflicting file changes (at the same time on different clients) are unlikely to happen in my use case, but I would prefer if they are handled properly (notification to the user). I've come across the following options, so far: something similar to Dropbox. iFolder seems to be the only thing that comes close, but its reputation (stability) and requirements put me off. A distributed file system such as OpenAFS. I'm not sure this will speed up file access. It is probably overkill for what I need. Maybe NFS or even Samba offer these possibilities. I read a bit about Windows' Offline Files, but its operation seems limited (at least on Windows XP). As this is just for personal use, I'm not willing to spend a lot of money. A free solution would be preferred. Also, the server needs to run on Linux, and I need a client for at least Windows.

    Read the article

  • How can I erase the traces of Folder Redirection from the Default Domain Policy

    - by bruor
    I've taken over from an IT outsourcer and have found a struggle now that we're starting a migration to windows 7. Someone decided that they would setup Folder redirection in the Default Domain Policy. I've since configured redirection in another policy at an OU level. No matter what I do, the windows 7 systems pick up the Default Domain Policy folder redirection settings only. I keep getting entries in the event log showing that the previously redirected folders "need to be redirected" with a status of 0x80000004. From what I can tell this just means that it's redirecting them locally. Is there a way I can wipe that section of the GPO clean so it's no longer there? I'm hesitant to try to reset the default domain policy to complete defaults. ***UPDATE 6-26 I found that the following condition occurred and was causing the grief here. I've already implemented the new policies for clients, and for some reason, XP was working great, 7 was refusing to process. The DDP was enforced. Because of this, and the fact that the folder redirection policies were set to redirect back to the local profile upon removal, it was forcing clients to pick up it's "redirect to local" settings. Requirements for to recreate the issue. -Create a new test OU and policy. -Create some folder redirection settings, set them to redirect to local upon removal -Remove settings on that GPO -Refresh your view of the GPO and check the settings. -You'll notice that the settings show "not configured" entries for folder redirection. -Enforce this GPO -Create another sub-OU -Create a GPO linked to this sub-ou and configure some folder redirection settings. -Watch as the enforced GPOs "not configured" setting overrides the policy you just defined. I've had to relink the DDP to all OU's that have "block inheritance" enabled, and disable the "enforced" option on the DDP as a workaround. I'd love to re-enable enforcement of the DDP, but until I can erase the traces of folder redirection settings from the DDP, I think I'm stuck.

    Read the article

  • Windows: File copy/move with filename regular expressions?

    - by Ian Boyd
    i basically want to run: C:\>xcopy [0-9]{13}\.(gif|jpg|png) s:\TargetFolder /s i know xcopy doesn't support regular-express filename searches. i can't find out how to find out if PowerShell has a Cmdlet to copy files; and if it does, how to find out if it supports regular expression filename matching. Can anyone think of a way to perform a recursive file copy/move with regex filename matching?

    Read the article

  • Duplicate file finder

    - by Andrija
    I need free duplicate file finder/remover app, with ability to find duplicate files/folders by name and/or by size and to remove one of duplicates. Can you please recommend any? and why? Thanks EDIT: Changed to CW. Please add more apps to list if you know any.

    Read the article

< Previous Page | 43 44 45 46 47 48 49 50 51 52 53 54  | Next Page >