Search Results

Search found 40294 results on 1612 pages for 'dot files'.

Page 283/1612 | < Previous Page | 279 280 281 282 283 284 285 286 287 288 289 290  | Next Page >

  • Is there a maximum of open files per process in Linux?

    - by Malax
    My question is pretty simple and is actually stated in the title. One of my applications throws errors regarding "too many open files" at me, even tho the limit for the user the application runs with is higher than the default of 1024 (lsof -u $USER reports 3000 open fds). Because I cannot imagine why this happens, I guess there might be a maximum per process. Any idea is very appreciated! Edit: Some values that might help... root@Debian-60-squeeze-64-minimal ~ # ulimit -n 100000 root@Debian-60-squeeze-64-minimal ~ # tail -n 4 /etc/security/limits.conf myapp soft nofile 100000 myapp hard nofile 1000000 root soft nofile 100000 root hard nofile 1000000 root@Debian-60-squeeze-64-minimal ~ # lsof -n -u myapp | wc -l 2708

    Read the article

  • Using the Drop Box folder while sharing on Mac OS X, files are locked.

    - by vgm64
    While using the Drop Box in the Shared folder on two MacBook Pros running Mac OS X 10.6, I noticed that when files were dropped into the drop box they were read only (locked?) when going from one of the laptops to the other (but not necessarily the other direction). Is there a way to change this? My girlfriend nearly had an aneurism because she couldn't rename a file. I duplicated it and deleted the original, but this seems unnecessary. I remember the permissions on the file were: nobody: Read Only everyone: Read Only

    Read the article

  • What linux permissions are need for www?

    - by Xeoncross
    I know that 777 is full read/write/execute permission for owner/group/other. So this doesn't seem to be needed as it leaves random users full permissions. What permissions are need to be used on /var/www so that... Source control like git or svn Normal users in a group like "websites" or added to "www-data" Servers like apache or lighthttpd And PHP/Perl/Ruby can all read, create, and run files there? If I'm correct, Ruby and PHP scripts are not "executed" directly - but passed to an interpreter. So there is no need for execute permission on files in /var/www. Therefore, it seems like the correct permission would be chmod -R 1660 which would make all files shareable by these four entities all files non-executable by mistake block everyone else from the directory entirely set the permission mode to "sticky" for all future files Is this correct? Update: I just realized that files and directories might need different permissions - I was talking about files above so i'm not sure what the directory permissions would need to be.

    Read the article

  • How can I access shared XP files over wifi with an Archos 5 Internet Tablet with Android?

    - by Fred
    I have an Archos 5 Internet tablet with Android. I have been using it to stream video over my XP wifi network. Yesterday the Archos stopped seeing any of my computers. I have found some programs that allow me to see the network, so I know the hardware works. However, this program does not allow me to pick the correct program to play the files I want. I suspect the problem lies somewhere on the XP side. This would probably work better if the Archos 5 allowed me to manually input the IP address. Help?

    Read the article

  • How can I access shared XP files over wifi with an Archos 5 Internet Tablet with Android?

    - by Fred
    I have an Archos 5 Internet tablet with Android. I have been using it to stream video over my XP wifi network. Yesterday the Archos stopped seeing any of my computers. I have found some programs that allow me to see the network, so I know the hardware works. However, this program does not allow me to pick the correct program to play the files I want. I suspect the problem lies somewhere on the XP side. This would probably work better if the Archos 5 allowed me to manually input the IP address. Help?

    Read the article

  • Can the Windows Media Player COM control play AVI files from memory (instead of from a file)?

    - by MusiGenesis
    I have a C# app, and I'm looking at using the Windows Media Player COM control to play animation and audio. So far, the only way I see of programatically controlling what the control is playing is to set its URL property to point to some file (I assume there's some way to pass in a playlist). Is there any way the WMP can render an AVI that is entirely in-memory, like a MemoryStream or something? If so, can WMP skip from one AVI to the next seamlessly (i.e. no glitch in either the audio or the video as it transitions from one to the next)? If WMP only plays files, is there some way to cue up a list of the files in advance of play start? If so, can WMP be made to skip from one file to the next without a brief interruption? Any knowledge or links to knowledge would be much appreciated.

    Read the article

  • Use inotifywait and lftp to synchronize servers

    - by KBoek
    I have two servers: Server A (CentOS), where people can upload files to (upload root is /files) Server B (Win 2008), with FileZilla FTP Server (FTP root is C:\content) I want that whenever a file is uploaded to Server A, to any subfolder under /files, the file is automatically copied to the exact same subfolder on Server B. Thus, if a user uploads "flowers.jpg" to /files/photos/12345/ then the file must be copied over FTP to C:\content\photos\12345 So far I have this bash script, it does copy the files to server B, but all files are placed in C:\content, and not in the corresponding subfolders. Who can help me find the correct syntax? #!/bin/bash cd /files inotifywait -q -r -m -e close_write,moved_to . --format %w%f | while read FILE; do lftp -e "put $FILE; exit" -u user,password -p 2121 ftp.server-a.com done

    Read the article

  • How do I upload huge files across the internet without using P2P?

    - by Brien Malone
    I work remotely and have 44GB of media files that I need to send back to my office. There are lots of free services out there that can handle up to 2GB, but I haven't seen talk of anything larger. We both have 50mbps+ connections, so I would rather not mail physical media (though, that is an option). Bittorrent is blocked at my corporate headquarters. We have an FTP server, but the per-user cap is 10GB. I use Citrix, but throughput is throttled to 3mbps. (44gb @ 50mbps = 4 to 5 hours... @3mbps = 5 or 6 days.) Any suggestions appreciated. Windows 7 and Windows 2003 Server are the OSes Involved I have tried JetBytes and it is blocked by our content filter

    Read the article

  • Looking for a component (.NET or COM/ActiveX) that can play AVI files in a WinForms app

    - by MusiGenesis
    I'm looking for something like the Windows Media Player control that can be hosted on a form. The WMP doesn't work for me because I need a control that can play a continuously-appended playlist of AVI files in sequence, so that the transition from one file to the next happens seamlessly (i.e. without any glitches or pauses in the video and audio). With WMP, there's always a delay between files of half a second or so. Does anyone know of a control (it can be either commercial or open-source) that can do this? I assume anything like this wraps DirectX, and that's OK too.

    Read the article

  • Encrypting Files with AES, Encrypting Key with RSA - Am I on the right track?

    - by Shawn Steward
    Overview: I'm trying to design an application that will encrypt files to safely send through the mail. I'm planning on using AES/RijndaelManaged encryption from .Net to encrypt the files initially, using a randomly generated key using RNGCryptoServiceProvider. I'm then encrypting this random AES key with a RSA Public key. The receiver of the data is the only one with the RSA Private key to decrypt it. My question: Is this the proper way to do something like this? If so, is it safe to send this RSA-Encrypted key with the data since it requires the private key to decrypt? Also - when having the end user generate their Public/Private key pair, what is the best way to save the Private key? I do not want it to be only accessible from one machine, so I am trying to avoid using the user's key store. But MSDN says it is not safe to save the key to a file, so how else can you accomplish this?

    Read the article

  • How to open embedded Visio files with only Visio Viewer?

    - by Christopher Galpin
    For some bizarre reason Visio Viewer seems incapable of opening vsd files when they're embedded within a (2010) Excel document: However if I open the Excel document with 7-zip, browse to xl\embeddings, extract oleObject1.bin and rename it to a .vsd file extension it opens perfectly fine with Visio Viewer in Internet Explorer. Since this opens fine on a computer with the full Visio installed, my guess is it's trying to launch Visio rather than IE. Is there something I can do to fix this? I've been scanning with Process Monitor and searching with RegScanner comparing a Visio-installed system to a VisioViewer-installed system and there doesn't actually seem to be any sort of registry value giving direction here... maybe it's just the nature of OLE. (If it can't be fixed I'm okay with converting the extracted one to a .png or so to replace the embedded object, but the "solutions" I've found for vsd to image conversion are very poor or non-working, i.e. buggy code, code with Visio dependencies, or online services.)

    Read the article

  • Opening Office 2007 files using a Vista or Win7 client on a server 2008 file share causes lockups an

    - by DrZaiusApeLord
    I think this mostly happens when trying to open files opened by other users. In the XP/2003 days you would get some kind of warning about a locked/read only file. With 7/Vista/2008 I'm just seeing clients hang (Word just sits there) and if I go into the file share and attempt to right-click on the file, explorer hangs for several minutes. I tried disabling AV on the file server as well as locally. No luck. I've read that SMB2.0 might be the culprit here, but even testing that solution means disabling it on both the client and server, and requires a server reboot. Does this sound like an SMB2 issue? The server is 2008 SP1. The clients are Win7 vanilla and Vista SP2 with all the current updates. Office 2007 SP2 with all updates. Thanks.

    Read the article

  • How to transfer files between a remote Linux machine and a local Windows machine?

    - by Lazer
    I need to transfer files/folders between a remote machine running Linux and my local machine running Windows XP. I usually access the remote machine through ssh, and sometimes use VNC sessions. What is the easiest way to transfer a file from the remote machine to my local machine and vice versa? I think ftp should be the solution. Is there a better way? If ftp is a good option, how do I initiate the transfer? Do I need to setup ftp servers on both the machines and then transfer? What exactly needs to be done?

    Read the article

  • Reading the *.sqlite files stored in the firefox profile.

    - by celil
    How would one go about reading the data in the *.sqlite files in one's profile? Trying to read them with sqlite3 was unsuccessful. $ sqlite3 SQLite version 3.7.4 Enter ".help" for instructions Enter SQL statements terminated with a ";" sqlite> .load ./downloads.sqlite Error: dlopen(./downloads.sqlite, 10): no suitable image found. Did find: ./downloads.sqlite: unknown file type, first eight bytes: 0x53 0x51 0x4C 0x69 0x74 0x65 0x20 0x66 sqlite> .exit Is this due to Firefox using an older version of sqlite? I am using Firefox v3.6.13

    Read the article

  • How do I extract all the files in a VHD to a hard disk including permissions?

    - by Middletone
    I'd like to know wha thte best way is to make an exact copy of a vhd image and pu tit onto my hard disk. I've tried xcopy but there seems to be a number of issues rlated to permissions when doing this. Ideally I'd like to copy the bits so that they match exactly on the new drive. I encountered this when trying to restore a vista backup only to discover the idiots work who decided to not let me restore a 400 gig image to a 1 TB drive size. I've sucessfully mounted the drive in Win 7 which is the environment in which I'm trying ot copy these files.

    Read the article

  • Is it possible to add files to the "Wordpress Media Library" using the command line?

    - by Tom
    Wordpress has it's own "Media Library" which is used when you upload images and other media for use in blog posts and pages. The advantage of the media library is that it automatically produces thumbnails of the images and the web interface gives you extra info such as who uploaded the image, which articles use the image, etc. My question is, does anyone have any tips on interacting with the media library via the command line instead of using the Wordpress web interface? For example, any ideas on how to add a image to the media library from the command line? If I copy files to the media library directory (usually .../wp-content/uploads/YYYY/MM/) from the command line they do not show up in the Wordpress dashboard - I guess because there needs to be an associated database entry for the media to be registered with Wordpress.

    Read the article

  • How do you handle files that can't support concurrent edits in Mercurial?

    - by Scott Whitlock
    I'm using Mercurial with TortoiseHg. Each developer has their own repositories, and there's one central repository on the server for synchronizing our changes. (This will sound lame, but we're using it to manage the source for a legacy VB6 project. Nothing we can do about that...) As has been pointed out elsewhere, there is a big problem in VB6 with merging the .frx (form resources) files. So code changes seem to merge fine, but if two developers both make changes at the same time in the form design view, we can't merge. I'm ok with disallowing concurrent edits, but of course the whole point of Mercurial is that it's distributed so there is no option to force a file to be locked before editing. I don't believe there's a Mercurial solution for this, so I'm wondering: other developers who are using Mercurial for version control, do you have some 3rd party tool that assists with locking files for editing in the cases where it's necessary? Did we make a mistake using Mercurial instead of something like SVN?

    Read the article

  • For large files compress first then transfer or rsync -z? which would be fastest?

    I have a ton of relativity small data files but they take up about 50 GB and I need them transferred to a different machine. I was trying to think of the most efficient way to do this. Thoughts I had were to gzip the whole thing then rsync it and decompress it, rely on rsync -z for compression, gzip then use rsync -z. I am not sure which would be most efficient since I am not sure how exactly rsync -z is implemented. Any ideas on which option would be the fastest?

    Read the article

  • Saving Types generated via Reflection.Emit as code file (.cs) instead of saving it in .dll files

    - by Manish Sinha
    Before start let me tell my experience: I am experienced with C#.NET, web services, XML part and few more. Reflection is something new to me, though I have read extensively on it and tried out some experimental code, but haven't made anything great using reflection I checked out many examples of how we can create Type at runtime and then which can be saved in an assembly (.dll) files. Of all the examples I have seen is about saving the created types in the .dll files instead of code file. Isn't there any way to create the code file out of reflection? I need to create code file since I want to distribute code instead of compiled assemblies. What I want to do is something like xsd.exe does, either spit out a .dll or the code file(in any language). Isn't there any way to create a code file, since most of the place I can find is AssemblyBuilder ab = System.AppDomain.CurrentDomain.DefineDynamicAssembly(an, AssemblyBuilderAccess.Save); and then lastly ab.Save("QuoteOfTheDay.dll");

    Read the article

  • Best and Proper Permissions Settings for Directory

    - by Dr. DOT
    I am interested in knowing the proper, yet security-conscious settings for a directory. Here's my scenario: I have a username for FTP access to my server called "user". For the purpose of the scenario, PHP runs as "nobody" on my server. I have a directory off the document root called "sample". The "sample" directory is chmod'd at 0755 (drwxr-xr-x) "Sample" is owned by "user" and the group is set to "user" The above is all very straight forward and standard. So I want to have a script be able to create (mkdir) and delete (rmdir) directories under "sample". Yet, I don't want to obviously overly expose my server by opening up the permissions (I could easily chmod sample to 0777 and make it world write-able). What is the best combination of permissions, owner settings and/or group settings to allow my script to create and delete directories under "sample" while retaining the ability for "user" to continue to FTP into the directory? Thanks.

    Read the article

  • Way to check files in and out with multiple developers.

    - by Roeland
    I am a web developer working with 2-3 people including me. Our current setup is very simplistic. We try to let each other know when we are working on a specific file. We use FTP to edit our files. Recently we have run into the problem of 2 people accidentally editing one file, or working on a local file then uploading when another person just did the same. From what I have read I need some sort of control system. I have heard of subversion and mercurial. It seems that these systems may not be what I need though, since its just giving me different versions of the files. I don't know if it solves the issue of two people working on a file and overwriting each others work. What are your suggestions for solving my problem?

    Read the article

  • How to set the path to "context path" for uploaded files using Apache Common fileupload?

    - by BugKiller
    Hi, I'm using Apache common fileupload library with Netbeans 6.8 + Glassfish.I'm trying to change the current upload path to be in the current context path of the servlet , something like this: WEB-INF/upload so I wrote : File uploadedFile = new File("WEB-INF/upload/"+fileName); session.setAttribute("path",uploadedFile.getAbsolutePath()); item.write(uploadedFile); but I notice that the library save the uploaded files into glassfish folder , here what I get when I print the absolute path of the uploaded file : C:\Program Files\sges-v3\glassfish\domains\domain1\WEB-INF\upload\xx.rar My Question : How can I force the common fileupload to save the uploaded file in a path relative to the current servlet path , so I don't need to specify the whole path ? is this possible ?

    Read the article

< Previous Page | 279 280 281 282 283 284 285 286 287 288 289 290  | Next Page >