Search Results

Search found 13068 results on 523 pages for 'copy and paste'.

Page 303/523 | < Previous Page | 299 300 301 302 303 304 305 306 307 308 309 310  | Next Page >

  • Making a hidden truecrypt volume with existing data

    - by Bill Grey
    I have a 1TB hdd, which I would like to encrypt. I would like to make a hidden volume, with almost nothing within but some decoy data, and the rest in a hidden volume. However, my driver is over 95% full. Is it still possible to do this, or would it have to be done on an empty drive, and then copy the data over? I could not find the answer to this question in the documentation. Also, how easy would it be to undo, or unencrypt the drive? Would it again need another empty drive to begin with?

    Read the article

  • How to push to github from a server account with multiple users?

    - by kirdie
    We have a web server which contains a web application stored as a github project. Now all of us can push from our local machines to github and then pull on the server but sometimes we want to make small changes and immediately see the effect so it would be great to be able to push at the server too. Now I created an ssh key for the server but I don't want to add the servers ssh key to my github account because then all github actions done from the server are counted to my account. Is it possible to add the ssh key to the github web application project without creating a new user for the server and what is the best practice for this situation? I also don't want to copy my private key to the server obviously.

    Read the article

  • ExtX file system on my usb key

    - by yves Baumes
    Hi all, if I format my usb key with an extX file system, copy some files on it and then give it to a friend for him to add files or modify existing one on this key, then he is rejected by its own system. Because its User ID (UID) nor GID are the same as mine on my machine. How to get rid of this limitation? Is it possible to disable user rights on a ext2/ext3 partition? Of course, I would really like not to rely on any other file system.

    Read the article

  • what am I doing wrong? Trying to reinstall Windows 7 starter onto my Acer Aspire One Netbook?

    - by Robbie Roberts
    I have been having some issues with my Netbook so I figured I would reformat it. I downloaded a copy of windows 7 starter, inserted it into my usb dvd drive and started my netbook. I made it as far as, "where do you want to install windows?" and it seems like the computer just freezes. it shows, Disk 0 Partition 1: PQSERVICE 13.0 GB OEM (Reserved) Disk 0 Partition 2: SYSTEM RESERVED 101.0 MB System Disk 0 Partition 3 218.8 GB Primary I cannot click on either of them, What am I doing wrong?

    Read the article

  • Upload a directory recursively to an FTP server

    - by Nicolas Raoul
    I am writing a Linux shell script to copy a local directory to a remote server (removing any existing files). Local server: ftp and lftp commands are available, no ncftp or any graphical tools. Remote server: only accessible via FTP. No rsync nor SSH nor FXP. I am thinking about listing local and remote files to generate a lftp script and then run it. Is there a better way? Note: Uploading only modified files would be a plus, but not required

    Read the article

  • Automatically distinguish difference between multiple HDDs in linux?

    - by Jakobud
    I'm running Ubuntu Server 9.10. I have two external USB HDDs. I use them each for different backup reasons. So certain data gets stored on one HDD, and different information gets stored on the other HDD. I want to make a script that can look at the external HDD can determine which HDD it is, so that it can copy the proper information to it. Is there a way for Linux to determine this? Like if I see one HDD as /dev/sdc1, then unplug it and plug in the other HDD, should Linux see it as /dev/sdd1 or will it be /dev/sdc1? I'm a bit of a Linux newb and I don't quite understand how it determines the /dev/sdxx values that it assigns to drives.

    Read the article

  • How to set my Ubuntu account to super user at all times?

    - by iaddesign
    I have the latest Ubuntu installed and I'll be the only one using it off the network. My question is: how can I make myself super user at all times? Because when I try to delete a file it says I don't have privileges to do so. I know you are going to say it's a security risk but I'm off the network and want to learn all that I can. I don't want to delete the files through the terminal but want to do it through the user interface/explorer. I've installed LAMP and can't copy my site to the www directory. I've tried to remove the preinstalled index file and it won't let me.

    Read the article

  • Unable to access a registry key

    - by gix
    I wondered why some file associations were broken and stumnled upon a nasty problem: I cannot access the key HKEY_CURRENT_USER\Software\Classes. As you can see above, I cannot view its current permissions, nor its owner (which should by SYSTEM). And no matter what I tried I'm unable to fix it. Trying to change the owner gives me a permission denied (also when trying from the Administrator account). RegDacl, which seemed to help another user, also aborts with a permission denied error. Any suggestions how to repair this? If I cannot repair this I guess I have to dump that user account and create a new one. Are there any "diff-like" tools out there to help copy over registry keys?

    Read the article

  • Login authentication vanished from MongoDB install

    - by Robert Oschler
    A few months ago I enabled password protection on my MongoDB install. Today I ran the Mongo client and forgot to use my login details. Instead of rejecting nearly everything I try to do from the shell, like it should, I had complete access to all the databases and collections. Fortunately this instance is only running a few test apps, so I quickly shutdown the MongoD instance until I figure this out. Has anybody ever seen this kind of behavior before and knows what is going on? The MongoD instance is running on a Linux VM hosted by Azure. The only thing I can think of is that perhaps Azure restored an old copy of the VM, but I received no E-mails to that effect and everything else on the server seems to be proper, including new daemon processes that I added after I enabled password protection on MongoD.

    Read the article

  • Shared storage solution for our sql server backups

    - by Gokhan
    We have 3 clustered sql servers. We have 5+ multi terrabyte databases and their backup files (compressed using quest litespeed) are hitting over 600gb each, We are required to keep at least a week or two weeks (if we can) of weekly full backups and then 6 days differential backups, and a week or 2 weeks worth of log backups local. We are currently limited to 2TB volumes from our san team, we can have multiple volumes but they are expensive ($200 per raw TB per month) and having to deal with many backup volumes instead of a single big volume is difficult. I think if we could have a shared network storage of 20TB+ raid 10 or so for all our servers for keeping the backups and another department will copy them to tape from the network storage and delete files according to the retention period would be good, if this box would be a build in operating system (even unix a complete file storage system) that would be good. What do you guys think, does this make sense to you, is there any manufacturer that sells a storage product like that which that work in a clustered environment? Thank you

    Read the article

  • VMware Fusion / New Machine / verify VMs?

    - by drewk
    I just got a 15" MacBook Pro with 8GB RAM, 2.66 i7 and 500GB SSD. Very sweet and fast! My old machine was 2.66 Core 2 Duo. I used Migration Assistant to migrate my old MBP to the new. When I started VMware to run Windows and Ubuntu I got the message first This VM has been moved or copied to a new machine. Did you move or copy it? I answered Copied. Then I got the message this CPU is a different configuration than the machine that created the VM. You may get unpredictable results. I answered Open Anyway. Windows XP, Windows 7, and Ubuntu all seem to work OK, but how can I verify? Is there some form of test harness for a VM? I have had a VM degrade and become unusable with very catastrophic data loss (on Parallels...), so I am a bit paranoid. Thanks,

    Read the article

  • Why does diskpart set the volume attributes on all volumes?

    - by Nick
    I was trying to migrate a Win7 OS from a HDD to a SSD. I've created 2 partition with 1024KB offset, with diskpart: 100MB System Reserved and a 60GB for C:. I've cloned their contents using Easeus Disk Copy. I've loaded the Windows 7 Boot DVD, and wanted to use diskpart to drop the letter for the System Reserved partition and make it hidden. select volume 0 detail volume attribute volume set nodefaultdriveletter attribute volume set hidden These 2 attribute set commands actioned on both volumes (0 and 1, MSR and C:) instead of the selected one, and viceversa. I've tried to clear these attributes from volume 1, but it cleared them also from volume 0. Why does DiskPart have this behaviour?

    Read the article

  • Script to gather all the files ending in .log and create a tar.gz file.

    - by Oscar Reyes
    I'm currently using this script line to find all the log files from a given directory structure and copy them to another directy where I can easily compress them. find . -name "*.log" -exec cp \{\} /tmp/allLogs/ \; The problem I have, is, the directory/subdirectory information gets lost because, I'm copying only the file. For instance I have: ./product/install/install.log ./product/execution/daily.log ./other/conf/blah.log And I end up with: /tmp/allLogs/install.log /tmp/allLogs/daily.log /tmp/allLogs/blah.log And I would like to have: /tmp/allLogs/product/install/install.log /tmp/allLogs/product/execution/daily.log /tmp/allLogs/other/conf/blah.log

    Read the article

  • Using 64 bit wuauclt from 32 bit command prompt

    - by Tim Brigham
    I have a script that for legacy reasons needs to run inside a 32 bit command shell. This script also includes references to certain core windows binaries - most notably wuauclt but others as well - which are not accessible by default within the 32 bit environment. This script is being run in several locations including many windows 7 and server 2008 r2 boxes. I'm aware of the possibility to copy files from the system32 to syswow64 in order to get around this. Is there any better method - something along the lines of adding an entry to the path variable - which will allow me to fall back to these 64 bit binaries from within a 32 bit script?

    Read the article

  • Setting up Raid 1 Array for Home Server

    - by user1048116
    I'm not sure if this is even possible, but it's worth asking on here! Essentially I have a old machine at home (well, not old hardware wise, but I recently built a new gaming rig), which I decided to install a copy of W2008 R2 on and use as a file/backup server and media center'ish machine. As of now, it has a single drive partitioned into C and D, with D being the Data partition. I have happened to find an old 1TB SATA drive lying around at home, and was wondering if it's possible to setup a Raid 1 array in my rig within Windows without needing to lose everything on my first drive (or maybe even just mirror a specific partition, say the Data partition, as this is just what stores my photos etc). Maybe this isn't possible, but you never know :) Regards, T.C

    Read the article

  • Mirroring MySQL server with diffrent configuration

    - by HTF
    I have to migrate MySQL server to a different data centre so I would like to create another MySQL slave server in new DC and then promote it to a master later on. I previously used LVM snapshots and Percona Xtrabackup for this purpose but this time I've optimized MySQL configuration file that prevents me from using these methods. Old server (backup): innodb_log_file_size = 256M innodb_log_files_in_group = 3 New server (restore): innodb_log_file_size = 512M innodb_log_files_in_group = 2 The Xtrabackup script and LVM snapshots copy the whole directory structure so the MySQL server won't start because there is a different size for InnoDB logs. Is there any solution to avoid a downtime in this case? I can't use mysqldumps as there is around 8000 databases so I would have to take the server down for a couple of hours. I was also thinking to use the old settings with Xtrabackup and then change it once the new server is promoted to a master - less downtime but I'm not sure if this will work? Thank you Regards

    Read the article

  • ms access template where to find the vb code

    - by tintincutes
    I'm very new to this ms access 2007. I have a copy of a charitable contribution template charitablecontributions.accdb. I would like to know where I can find the code of it? I opened it by holding down the shift button and double click on it and it will open to a normal ms access where you can modify the table and so. But when I just click the file: charitablecontributions.accdb it will open to a formular view where the ribbon bars are gone. I wanted to know how, how is it possible to have this formular view? Because I have a Test.mdb from 2003 and I also would like it to be open like the formular view of charitablecontributions.accdb. I'll appreciate your help. Thanks

    Read the article

  • How to create one additional hidden partition in a USB drive

    - by backslash17
    For security purposes I need to locate a additional hidden partition on USB flash drive. The USB drive contains a security application that will check (in code) if the hidden partition exists. If not I will assume that the application is a non valid copy. Any idea about it? Thanks in advance. EDIT: There is already a programmed secure method to check if the USB drive is correct using the WMI Win32_drive class. The idea is to locate the drive info into the hidden partition and to check if it correct for anti-tampering procedures.

    Read the article

  • Brand new Seagate HDD has high raw read error rate

    - by kpax
    I've just purchased a brand new Seagate ST31000524AS 1TB HDD. Manufacture date shows as January 2012 (yes that's as new as new can get), so must be one of the new batches from the post-flood Thailand. Anyway, I downloaded a copy of Active Hard Disk Monitor tool to check the S.M.A.R.T. parameters and I find the parameter Raw Read Error Rate is very low. Should I be worried? Will this rectify over time? This hdd is just 7 hours old; what gives? Edit: I meant high raw read error rate - Title updated accordingly

    Read the article

  • What's the utility of the return command in autohotkey?

    - by Shashank Sawant
    In the instances where the return command returns a value, the utility is obvious. I have seen the return command being used even when it is seemingly unnecessary. Let me show the following examples: Example 1: Loop { if a_index > 25 break ; Terminate the loop if a_index < 20 continue ; Skip the below and start a new iteration MsgBox, a_index = %a_index% ; This will display only the numbers 20 through 25 } Example 2: IfWinExist, Untitled - Notepad { WinActivate ; Automatically uses the window found above. return } Why is the return command used in Example 2 but is not used in Example 1? Both examples are copy-pasted/modified-pasted from the autohotkey.com's documentation.

    Read the article

  • Pull row of data from one place in spreadsheet to another

    - by bookcasey
    I have a list of names and their assorted properties. I am trying to find a way to quickly move an entire row's worth of data into a different sheet so I can manipulate it. Currently I can accomplish this by copy and pasting the entire row, but this is not elegant nor dynamic (the number in blue does not update with this method): My ideal solution: A name is typed into the "Name" column. The spreadsheet program pulls the associated row from the "main" sheet into the new sheet. I am currently using Numbers.app, but am curious if this is possible in Excel, OpenOffice Calc, Google Docs, or any other spreadsheet program. Thank you.

    Read the article

  • Trigger ZFS dedup one-off scan/rededup

    - by Jake Wharton
    I have a ZFS filesystems which has been running for some time and I recently had the opportunity to upgrade it (finally!) to the latest ZFS version. Our data doesn't scream dedup but I firmly believe based on small tests that we could gain anywhere from 5-10% of our space back for free by utilizing it. I have enabled dedup on the filesystem and new files are slowly being dedupified but the majority (95%+) of our data already exists on the filesystem. Short of moving the data off-pool and then recopying it back, is there any way to trigger a dedup scan of existing data? It doesn't have to be asynchronous or live. (And FYI there isn't enough room on the pool to copy the entire filesystem to another and then just switch the mounts.)

    Read the article

  • DNS Zone file and virtual host question

    - by Jake
    Hi all, I'm trying to set up a virtual host for redmine.SITENAME.com. I've edited the httpd.conf file and now I'm trying to edit my DNS settings. However, I'm not sure exactly what to do. Here's an snippet of what's already in the named.conf file (the file was made by someone else who is unreachable): zone "SITENAME.com" { type master; file "SITENAME.com"; allow-transfer { ip.address.here.00; common-allow-transfer; }; }; I figure if I want to get redmine.SITENAME.com working, I need to copy that entry and just replace SITENAME.com with redmine.SITENAME.com but will that work? I was under the impression I needed a .db file but I don't see any reference to one in the current named.conf file. Any advice would be great and if you need more info to answer the question, don't hesitate to ask.

    Read the article

  • IRC Services with failover support?

    - by insertjokehere
    I run a single server (call it 'server A') IRC 'network', and thank to the generosity of some friends, I have been given a second server ('server B') that I can run an IRCd on in order to provide redundancy in case server A crashes. This is fine, I can set up a round-robin DNS with the servers linked. The problem I have is what to do about services? Does anyone know of a way to get the services to 'fail over' in case of a server failure? Eg, Server A starts off running the services, but suddenly crashes. Server B detects this and starts its own copy of the services (ideally with the same configuration and data as the services on Server B) One solution that comes it mind is to write a bot that each server runs, that sit in a channel periodically checking if the bot from the other server is in the channel. If it is, then all is well. If not, then failover. I would prefer not to have to code this myself though We are currently using Unreal IRCd and Anope services on Linux

    Read the article

  • How to access USB stick content from VMWare running Ubuntu 10.10?

    - by JVerstry
    Hi, I am running Ubuntu 10.10 via VMWare under Windows 7. I have followed the procedure to install the USB stick. It is now connected to the host. However, I don't know how to access the content of the stick. My Google research indicates that this may be a mounting issue. I read somewhere that I should check /proc/bus/usb, but the usb directory does not exist in /proc/bus. Unfortunately, I am not a Linux expert at this. The ultimate issue I am trying to solve is the one describe here. I am trying to use vi to create ~/.vmware/config, but it is virtually impossible to use vi, since I don't have access to the arrow keys (chicken & egg problem). I have created the config file on my usb stick and want to copy it where it should be. Thanks!

    Read the article

< Previous Page | 299 300 301 302 303 304 305 306 307 308 309 310  | Next Page >