Search Results

Search found 30764 results on 1231 pages for 'copy control'.

Page 367/1231 | < Previous Page | 363 364 365 366 367 368 369 370 371 372 373 374  | Next Page >

  • Clipboard app that automatically saves to Web share folder ?

    - by cyberpine
    I currently use SnagIt and have used Windows 7 Snipping tools. These tools allow you to copy a piece of your screen the clipboard and then paste it into other applications on your desktop. They work with Outlook and it's really useful as you don't have to save the image to do a write-up or send an email. Problem is - you can't paste clipboard images into webforms. Gmail web does have image insert feature, but it only works on web images that you copied as it extracts the full url from your copy and uses that. It does not work on with images in your local clipboard. Does anybody know of an application that would allow to clip and save directly to a web folder, maybe something that replaces your clipboard with the URL of the save image location? It would be awesome if the picasa web client or evernote allowed actual clipboard pasting. Instead they ask for a file to upload.

    Read the article

  • Explorer forgetting how to cut and paste files

    - by raimesh
    My Windows XP (SP3) machine (at work) occasionally "forgets" how to cut-and-paste files in Explorer - if I cut or copy a file then go to paste it elsewhere, the "paste" and "paste shortcut" menu items are greyed out. The keyboard shortcuts (Ctrl+X/Ctrl+C,Ctrl+V) don't work either, and neither does Drag-and-Drop - clicking and dragging a file/folder doesn't change the pointer and letting go doesn't drop the item anywhere. Restarting the explorer.exe process usually fixes the problem, but sometimes I need to do a full reboot. Note: this may be related to the following existing question, although that's Vista and it sounds like the asker has more permanent problems than I do: http://superuser.com/questions/27045/copy-paste-in-vista-explorer-broken-not-ms-vpc So, any idea what might be causing the problem and how to avoid it?

    Read the article

  • Pasting into Vista cmd.exe broke, why?

    - by Michel de Ruiter
    I use Windows Vista x64 and regularly use the command line window CMD.EXE. I have enabled QuickEdit Mode (and Insert Mode and AutoComplete), to be able to quickly copy and paste text. Copying (select block, Enter) works fine. Pasting (right click) text also works, as long as it has been copied inside a CMD.EXE window. When I copied the text somewhere else (in an editor, browser or whatever) however, pasting into CMD.EXE does not work! :-( Using the menu to Edit, Paste does not do anything either, so it's not a mouse thing. I also tried elevating CMD.EXE. I can copy/paste freely between CMD.EXE instances of all sorts: elevated/normal, x64/x86... I'm sure it did work on this machine until relatively recently. What could have happened? Some Windows Update perhaps? The problem has been reported by others, but without a solution.

    Read the article

  • Outlook 2010 - Downloads same messages multiple times when opening

    - by AaronM
    I run multiple POP mailboxes, and have the 'Leave copy of messages on server' set for 20 days. I have been working like this for several years. However, I have just this week upgraded to Outlook 2010, and for some reason a couple of the mailboxes are downloading all the messages again, each time I open outlook. I have gone through all the settings but I can't see anything obvious, and apart from the upgrade to 2010, I havent touched any settings (2010 imported everything from 2007). I use webmail when away from my PC, hence why I leave a copy of emails on the server. Any ideas? It only downloads when Outlook first opens, thankfully not every time I do a send/receive.

    Read the article

  • Windows Boot disc to save files?

    - by acidzombie24
    Somehow after updating my legit windows 7 OS with no pirated or mod software on my PC i was welcome to a black screen. I popped in my ghost disc and copied the files i need to an external HD. IIRC windows 7 disc can do that too. Problem with the way i did it on ghost was it excepted me to select 1 file (an HD disc image) so i couldnt select multiple folders to move. Also when i did move i had no idea if it finished or how long it would take. My linux live cd couldnt access the HD. Anyways, is there a disc i can use to easily copy files from my laptop to my external HD? I think ghost, windows 7 and windows server all allow me but is there one that is better suited to copy files?

    Read the article

  • NFS Issues in Gnome

    - by Alex
    I mount NFSv4 export via /etc/fstab and mount and use the shared folder in nautilus. There are two issues: When I copy a large file (around 4 GB) to the NFS server, the progress bar rapidly goes to 2 GB and then basically stops moving. But the copy s still in progress - it is just not displayed well When I disconnect from the network without unmounting the nfs share, nautilus freezes. How can I work around that? /etc/export on the server /export/share 192.168.0.0/24(rw,sync,insecure,no_subtree_check,anonuid=1000,anongid=1000) /etc/fstab on the client: server:/share /mnt nfs4 soft,tcp

    Read the article

  • Home Server Restore

    - by Bryan Avery
    I have had to reinstall Home Server on my server and I would now like to restore it back to the state it was the last moment it stopped. I have the hard drive in a state it was last in, which is a small 250 gb hard disk. I have now installed 1.5tb hard disks, and installed a full licenced copy, as the original copy was a trial version. So I'm in a state where I have a new install, I have one of the old drives plugged in, but I can't transfer the old backups across, how do I do this?

    Read the article

  • Home Server Restore

    - by Bryan Avery
    I have had to reinstall Home Server on my server and I would now like to restore it back to the state it was the last moment it stopped. I have the hard drive in a state it was last in, which is a small 250 gb hard disk. I have now installed 1.5tb hard disks, and installed a full licenced copy, as the original copy was a trial version. So I'm in a state where I have a new install, I have one of the old drives plugged in, but I can't transfer the old backups across, how do I do this?

    Read the article

  • How to avoid copying corrupted files with rsync

    - by Roberto Aloi
    I have an HDD with plenty of files, some of which are unfortunately corrupted. I'm now trying to copy the good files into a new HDD. I'm using: rsync -azP SRC TGT When rsync comes to one of the corrupted files, I can see a message in the console: rsync: read errors mapping XXX: Input/output error (5) In the target folder, I still see the corrupted file, which I'm not able to open and which I have to delete manually. Is there any option to tell rsync not to copy files after a i/o error?

    Read the article

  • How to setup VM in KVM? Qcow or LVM etc.

    - by JohnAdams
    Finally, after quite a bit of this vs that, I have chosen to virtualize a couple of my servers with KVM. I did do a test setup as well, but I have a few questions about setting VM's in KVM. Would appreciate pointers. What is the best storage to use - Qcow2 or LVM? I like the fact that I can copy the VM file easily with a Qcow2 but what about LVM, how do I take a backup or make copy on a development server to play with? I know I can clone a LVM, but how do I bring to my development server? How do I setup the guest partitioning? For example, when setting up Ubuntu inside Ubuntu, do I choose LVM for that VM or regular fdisk partitioning? Can I increase the partition size then later, if I need a bigger disk?

    Read the article

  • VirtualBox cloned Ubuntu Server network error

    - by Luke
    I run a number of virtual servers on my network and I want to be able to easily clone base installations of Ubuntu Server. I use the VBoxManage command to clone the actual hard disk and I then create a new profile for my VM and copy over the settings of the original VM. However, when I boot into the cloned VM, there seems to be a network problem. When I issue a PING I get the message "Network unreachable". I traced it down to the fact that the virtual network card of the cloned VM has a different MAC address then the original VM. When I copy the MAC address the clone seems to work fine. How can I have the cloned VM have it's own MAC address?

    Read the article

  • Open file in local text editor from within an SSH connection

    - by Sam
    I'm not a vim guy. I'd like to be able to open log files in Sublime Text when in an SSH connection from within Terminal. Is there a way I could do this? I'm thinking there must be a command or something that could copy the file over to a temporary directory in OS X and then open it in Sublime Text, and when I save it, it'll copy back to the original location through SSH; similar to how FileZilla does it. I'm on Mac OS X MT. The server I SSH into is running Ubuntu. I'm using Terminal.

    Read the article

  • Is it possible to continue torrent downloads from other client

    - by Nrew
    I have downloaded a torrent file but I downloaded it using bitlord. And because bit lord doesnt have the ability like initial seeding. I decided to use utorrent. Ive already tried to continue it. What I did was to: Opened the original torrent file in utorrent. Target where the incomplete bitlord incomplete file was. But it seems that utorrent isn't detecting the file and it makes another copy of it and start to download from the beginning. Im at about 80 percent and there are only 18 seeds but there's 17 leech. And my connection is slow so I really have to initial seed. Other details: I've reformatted the original machine where I downloaded it. Also didn't copy the original application.

    Read the article

  • Where are Wireless Profiles stored in Ubuntu

    - by LonnieBest
    Where does Ubuntu store profiles that allow it to remember the credentials to private wireless networks that it has previously authenticate to and used? I just replaced my Uncle's hard drive with a new one and installed Ubuntu 10.04 on it (he had Ubuntu 9.10 on his old hard drive. He is at my house right now, and I want him to be able to access his private wireless network when he gets home. Usually, when I upgrade Ubuntu, I have his /home directory on another partition, so his wireless profile to his own network persists. However, right now, I'm trying to figure out which .folder I need to copy from his /home/user folder on the old hard drive, to the new hard drive, so that he will be able to have wireless Internet when he gets home. Does anyone know with certainty, exactly which folder I need to copy to the new hard drive to achieve this?

    Read the article

  • Can't change read only folder in windows 7

    - by James Drinkard
    I'm trying to run a Spring MVC 2.5 tutorial and when I run the ant script for a deploy, I get this error: deploy: [copy] Copying 2 files to C:\apache-tomcat-7.0.8\webapps\c:\projects\workspace\springapp BUILD FAILED C:\projects\workspace\springapp\build.xml:46: Failed to copy C:\projects\workspace\springapp\war\WEB-INF\web.xml to C:\apache-tomcat-7.0.8\webapps\c:\projects\workspace\springapp\WEB-INF\web.xml due to failed to create the parent directory for C:\apache-tomcat-7.0.8\webapps\c:\projects\workspace\springapp\WEB-INF\web.xml After reviewing the directory: springapp I saw the properties as read-only. No problem I thought as I'm logging in as administrator. However, changing the uac settings, going to a command prompt as admin and then trying to change the properties of the folder with attrib, making me the owner of the folder, changing the security settings etc... did nothing. I can't seem to change this folder to anything. So my question is, how do I change the settings on that folder so ANT can make changes to that folder?

    Read the article

  • Advice on off-site backup of Hyper-V Failover Cluster

    - by Paul McCowat
    We are currently setting up a Server 2008 R2 which will be off-site over a leased line with VPN. At the main site is 2 x Hyper-V hosts in a failover cluster with PowerVault M3000i iSCSI SAN. We are using BackupAssist for local backups and each host backups up itself and it's guests nightly creating a 500GB backup each which is copied to a 2TB rotated NAS drive. Files and SQL DB's are also backed up / log shipped etc. Looking for the best way to backup the Hyper-V VM's and copy them off-site so that the OS's are only a month old and the data is a day old. The main backups are too large to transfer between backups so options discussed so far are: Take rotating individual backups of the VM's each day and copy over, Day 1 SQL VM, Day 2 Exchange VM etc, would require more storage. Look in to Hyper-V snapshots, however don't believe these are supported in clustering. 3rd party replication tools

    Read the article

  • CMD file time not always matching windows explorer file time

    - by skyrail
    I have a set of file I need to set the created, modified and last access date to exif date taken value, after a copy between 2 folders (might be fat32 on memory card or ntfs on fixed or usb disk). When I copy a file, the date and time switch to the current date. Then I change all 3 dates manually, either with change attributes in windows explorer or far manager on the command line. To make it faster I wrote a batch script getting original file dates (with php and function stat), building a batch script that invoke nircmd setfiletime for each file. Then I apply this batch to the copied version. The operation is relatively fast and reliable. Unfortunately, a bunch of files have last access and created time different in cmd and windows explorer (1H difference). Very strangely, it happens with dates between november and february, which make the operation unreliable. Why is this happening, and how can I fix it?

    Read the article

  • Accessing a storage-side snapshot of a cluster-shared volume

    - by syneticon-dj
    From time to time I am in the situation where I need to get data back from storage-side snapshots of cluster shared volumes. I suppose I just never figured out a way to do it right, so I always needed to: expose the shadow copy as a separate LUN offline the original CSV in the cluster un-expose the LUN carrying the original CSV make sure my cluster nodes have detected the new LUN and no longer list the original one add the volume to the list of cluster volumes, promote it to be a CSV copy off the data I need undo steps 5. - 1. to revert to the original configuration This is quite tedious and requires downtime for the original volume. Is there a better way to do this without involving a separate host outside of the cluster?

    Read the article

  • Why is one specific file not showing up in my document library?

    - by Jay Bazuzi
    In my "Documents" library on Windows 7, one file is not showing up in Windows Explorer. When I look in C:\Users\%USERNAME%\Documents\blah\blah all 24 files appear. But when I look in Libraries > Documents > blah > blah only 23 show up. I made a copy of the file and the copy appears. Refresh doesn't help. The "Arrange by" setting defaults to "Name". When I change it to "Folder" the extra file appears, but changing it back to "Name" the file disappears again. How can I make the file appear in all views? Why would it disappear? EDIT: I deleted the Windows Search Index and things seem to be working again. I say it's a bug in the Search Service.

    Read the article

  • How to tell nginx to honor backend's cache?

    - by ChocoDeveloper
    I'm using php-fpm with nginx as http server (I don't know much about reverse proxies, I just installed it and didn't touch anything), without Apache nor Varnish. I need nginx to understand and honor the http headers I send. I tried with this config (taken from the docs) but didn't work: /etc/nginx/nginx.conf: fastcgi_cache_path /var/lib/nginx/cache levels=1:2 keys_zone=website:10m inactive=10m; fastcgi_cache_key "$scheme$request_method$host$request_uri"; /etc/nginx/sites-available/website: server { fastcgi_cache website; #fastcgi_cache_valid 200 302 1h; #fastcgi_cache_valid 301 1d; #fastcgi_cache_valid any 1m; #fastcgi_cache_min_uses 1; #fastcgi_cache_use_stale error timeout invalid_header http_503; add_header X-Cache $upstream_cache_status; } I always get "MISS" and the cache dir is empty. If I uncomment the other directives, I get hit, but I don't want those "dumb" settings, I need to control them within my backend. For example, if my backend says "public, s-maxage=10", the cache should be considered stale after 10 secs. Instead, nginx will store it for 1h, because of these directives. I was thinking whether I should try proxy_cache, not sure what's the difference. In both fastcgi and proxy modules docs it says this: The cache honors backend's Cache-Control, Expires, and etc. since version 0.7.48, Cache-Control: private and no-store only since 0.7.66, though. Vary handling is not implemented. nginx version: nginx/1.1.19 Any thoughts? pd: I also have the reverse proxy that is offered by Symfony2 (which I turn off to use nginx's). The headers are interpreted correctly by it, so I think I'm doing it right.

    Read the article

  • Prevent Java application from accessing/monitoring/altering clipboard contents

    - by mcstrother
    I'm a student using a service that provides practice questions for standardized tests. The service requires that I access the questions by downloading and running a Java application. If I try to copy anything from any window of my computer (including applications unrelated to the question bank) while the application is running, the copied item is replaced with an obnoxious message asking me to not pirate their copyrighted material. I find this obnoxious, and I also really don't like the idea that any application can slurp up any and all potentially sensitive information that I happen to copy while it's running. Is there are a way to limit the privileges of this application to stop it from doing this? Thank you!

    Read the article

  • Why does my drive show as 180GB used, but copies only 40GB?

    - by Manuel
    My computer crashed, so I removed the drive (with Windows XP Professional 32-Bit) and put this drive in another computer running Windows 7 64-bit. Booting with Win7, when going to My Computer it is showing the XP drive as more than 180GB used. So as a quick backup, I'm trying to copy all the disk's content to a folder, but when I start copying, it's showing only 40 GB as the total size of the files to copy. I enabled showing system and hidden files in View options. What could be the problem?

    Read the article

  • Hyper-V Ubuntu Networking Problems Copying Large Amounts of Data

    - by Anonymous
    I am trying to copy a large amount (about 50 GB) of data over my network from a Hyper-V-hosted virtual machine running Ubuntu 11.04 (Natty Narwhal) to another (non-virtual) Ubuntu host that I plan to use for testing upgrades to one of our web applications. The problem I am having is with the virtual machine, which I shall refer to in what follows as "source.host". This machine is running 64-bit Ubuntu Server with the 2.6.38-8-server kernel and the Microsoft Linux Integration Components for Hyper-V kernel modules (hv_utils, hv_timesource, hv_netvsc, hv_blkvsc, hv_storvsc, and hv_vmbus) loaded. It uses a Hyper-V "synthetic network adapter" for its networking interface. To do the copy, I log on to the machine with the data and run the following commands (Call the remote machine "destination.host".): $ cd /path/to/data $ tar -cvf - datafolder/ | ssh [email protected] "cat > ~/data.tar" This runs for a while and then suddenly stops after transferring somewhere from 2-6 GB. The terminal on the source.host machine displays a Write failed: broken pipe error. The odd part is this: after this occurs, the "source.host" machine is no longer able to talk to the rest of the network. I cannot ping any other hosts on the network from the "source.host" machine, and I cannot ping the "source.host" machine from any other host on the network. I am equally unable to access the any of the web services hosted on "source.host". Running ifconfig on "source.host" shows the network adapter to be up and running as usual with the correct IP address and everything. I tried restarting the networking service with $ /etc/init.d/networking restart but the problem does not go away. Restarting the machine makes it capable of talking to the network again -- it can ping and be pinged by other hosts, and the web services are also accessible and usable as normal -- but attempting the copy operation again results in the same failure, requiring another restart. As an experiment, I tried replacing the tar -- ssh pipeline above with a straight scp: $ scp -r datafolder/ [email protected]:~ but to no avail Thinking that the issue might have to do with the kernel packet-send buffers filling up, I tried increasing the buffer size to 12 MB (up from the 128 KB default) with # echo 12582911 > /proc/sys/net/core/wmem_max but this also had no effect. I'm guessing at this point that it might be a problem with the Microsoft synthetic network driver, but I don't really know. Does anyone have any suggestions? Thank you very much in advance!

    Read the article

  • Dying SanDisk Memory Stick Pro Duo

    - by Different55
    I have a Memory Stick Pro Duo and after attempting to delete the largest file from a Mac the stick has become unusable. I can almost access it. When I put it in my PC I can open/delete/copy/paste/rename/modify one file/folder, then it can't detect the card. If I reinsert the card I can move on to the next file, but this is really annoying and my PSP won't read it at all. The memory card access light will flash for a really long time before it says that every file is corrupted. When I have tried to format it with either the PC, PSP, or a camera that uses a memory stick pro duo, it fails. I've tried with all the different options on windows, I tried formatting it through CMD, but nothing I have tried works. Should I copy every file off one by one or is there a way to fix it?

    Read the article

< Previous Page | 363 364 365 366 367 368 369 370 371 372 373 374  | Next Page >