Search Results

Search found 7533 results on 302 pages for 'knowledge transfer'.

Page 84/302 | < Previous Page | 80 81 82 83 84 85 86 87 88 89 90 91  | Next Page >

  • Restoring a SD card after using it to root a device

    - by Raz
    I recently purchased the Nook Touch reading device. I rooted it using the instructions on nookdevs.com Rooting included using win32diskimager to somehow transfer an .img file onto the card. Rooting is now completed and as far as I can tell, I don't need to keep the files used on the SD card, which has been reduced from a 3.5 GB or so unit to a 75 MB unit. Is there a simple way for me to somehow reformat the SD card to its original state or at least 3.5 GB. I cannot simply format it.

    Read the article

  • Could not upload .htaccess

    - by syalam
    I am using Springloops to automatically take my SVN repo and deploy onto my server. I am getting the following error: Could not upload .htaccess Could not upload .htaccess using BINARY transfer ---------------------------------------------------- Connecting to dev.convrrt.com Logging in as convrrt Entering destination directory ~/ Entering passive mode REVISION: 1 -> 30 Getting changes Deleting files Removing directories Creating directories and files Extracting file: .htaccess...OK Uploading file: .htaccess [644] R: interrupted How can I diagnose this?

    Read the article

  • Transferring programs to a different Windows partition

    - by SquareWheel
    I have two Windows partitions installed on a HDD, as well as a third partition where I have hundreds of programs installed. I've installed all the programs with my Windows 7 RC, but now that the Release Candidate is coming to an end I have to switch over to my other Windows installation. Is there an easy way to transfer all of my programs registry entries and other references to my new Windows installation, or do I have to reinstall all of my programs on my new partition? Thanks.

    Read the article

  • Freeware (preferably open-source) tool for creating multi-file spanning archives as a self merging SFX

    - by Lockszmith
    I have a large file I want to transfer using either Internet storage hosting, DVD-Rs or USB storage, which sometimes is limited to FAT file-systems (for example: mobile phones) What I'm basically looking for is a tool that create multiple files/volumes (less than 2GB each - FAT's file size limit) which are packed with a self-extracting executable. Currently the only tool I found doing this is WinRAR, but that's shareware, and not free. Is there any Free, preferably Open-Source tool that does that? Thank in advance

    Read the article

  • Virtual Box between differnt PCs

    - by Jim
    I am using Virtual Box machine (as .vdi) on external HDD drive so I can move it between different pcs. On each PC I have created new VM machine, attached that vdi and it seems working well. However, I've noticed that during my last transfer the latest installed applications were not saved on the .vdi, which made me to loose 1 day of work. I don't remember of doing any snapshots which I could merge. Can someone give me a hint how to handle this efficiently in feature?

    Read the article

  • BIND DNS server (Windows) - Unable to access my local domain from other computers on LAN

    - by Ricardo Saraiva
    I have a BIND DNS server running on my Windows 7 development machine and I'm serving pages with WAMPSERVER. My ideia is to develop some tools (in PHP) for my intranet at work and I want them to be accessible via LAN in this format: http://tools.mycompany.com I've already placed BIND and I can access http://tools.mycompany.com on the machine that holds BIND server, but I cannot access it from other LAN computers. I've done the following on my router: defined static IP's for all LAN computers set Port Forwarding to my server (remember: it serves DNS and Web pages) set DNS server configuration to point to my LAN server On LAN computers, I went to Local Area Network properties and also changed the DNS server IP in order to point to my local DNS server. If it helps, here is my named.conf file: options { directory "c:\windows\SysWOW64\dns\etc"; forwarders {127.0.0.1; 8.8.8.8; 8.8.4.4;}; pid-file "run\named.pid"; allow-transfer { none; }; recursion no; }; logging{ channel my_log{ file "log\named.log" versions 3 size 2m; severity info; print-time yes; print-severity yes; print-category yes; }; category default{ my_log; }; }; zone "mycompany.com" IN { type master; file "zones\db.mycompany.com.txt"; allow-transfer { none; }; }; key "rndc-key" { algorithm hmac-md5; secret "qfApxn0NxXiaacFHpI86Rg=="; }; controls { inet 127.0.0.1 port 953 allow { 127.0.0.1; } keys { "rndc-key"; }; }; ...and a single zone I've defined - file db.mycompany.com.txt: $TTL 6h @ IN SOA tools.mycompany.com. hostmaster.mycompany.com. ( 2014042601 10800 3600 604800 86400 ) @ NS tools.mycompany.com. tools IN A 192.168.1.4 www IN A 192.168.1.4 On the file above 192.168.1.4 is the IP of the local machine inside my LAN. Can someone help me here? I need my web pages to be accessible from other computers inside my LAN using my custom domain name. I've tried on other computers and they can access my server via http://192.168.1.4/, but no able when using http://tools.mycompany.com . Please, consider the following: I'm completely new to BIND I have basic knowledge in Apache configuration Thanks a lot for your help.

    Read the article

  • NTFS External Drive takes too long to takes data in Mac ?

    - by mgpyone
    I've an Seagate 500 GB external HD (NTFS). To read/write it on Mac (OSX 10.6.2) , I've tried MacFUSE and NTFS-3G to write my HD on Mac. Though I could be able to see the hard drive, it takes too long to see the contents like this is this normal ? also the data transfer takes too long time and the hard disk becomes too hot . Any suggestions are most welcome.

    Read the article

  • How to port VirtualBox machines with the recent snapshot?

    - by Liran Orevi
    I've ported a 'hard disk' using the 'clonehd' advice given on How to easily port VirtualBox machines? But it's not at the latest state, it's probably before any 'snapshot' was taken. I have only copied the 'hard disk', I haven't copied any 'snapshot' mainly because there are so many. I'm interested to transfer the 'hard disk' at it's latest state. How can this be done? (easily and not by coping the entire snapshots, machine etc...)

    Read the article

  • Is it better to use N WiFi or Cat5e ethernet for large file transfers?

    - by Seth
    I have a network containing 802.11 N WiFi Devices and gigabit ethernet devices, however using only Cat5e wiring. I know that if I used Cat6, the ethernet would be much faster however if I need to transfer a large file between a device thats connected normally connected via ethernet and a device thats normally connected via WiFi, would it be advantageous to bother plugging the WiFi device in to the network? or should I switch the ethernet device to WiFi? or does it really matter?

    Read the article

  • If I uninstall Bonjour and Apple Mobile Device Support will transferring songs to my ipod via itunes

    - by Kate
    I am one of those people who hate to have even one unnecessary program installed. I noticed when I downloaded and installed itunes two other components were installed and shows up in my control panel Uninstall programs screen: "Bonjour" and "Apple Mobile Device Support". Are these two necessary for itunes to work? I never buy songs via itunes, just use the program to transfer smp3s to my ipod.

    Read the article

  • Read NTFS partition on RHEL 5.8

    - by Alex Farber
    I have RHEL 5.8 64 bit, and NTFS partition on the same disk. How can I get access to this partition? This answer Unable to mount NTFS drive with RHEL 6 doesn't work for me: [root@localhost alex]# rpm -Uvh http://download.fedora.redhat.com/pub/epel/6/i386/epel-release-6-5.noarch.rpm Retrieving http://download.fedora.redhat.com/pub/epel/6/i386/epel-release-6-5.noarch.rpm error: skipping http://download.fedora.redhat.com/pub/epel/6/i386/epel-release-6-5.noarch.rpm - transfer failed - Unknown or unexpected error

    Read the article

  • Upload database backup from mysql to Amazon S3 or Glacier without creating local file

    - by Rubem Azenha
    Is there a tool that makes possible to backup a Mysql database to Amazon S3 or Amazon Glacier without having o create a local file with the database contents? Something like that: mysqldump -u root -ppass -h host --all-databases | magical-s3-tool s3-bucket backup-yyyy-mm-dd.sql This magical tool would use the pipe data and transfer the backup data directly to S3, without creating a local file.

    Read the article

  • Backup data to remote dedicated server

    - by Alex Bagnolini
    My company already has a "local" backup strategy, but is willing to also backup data on our remote dedicated server as an additional "plus". Some info: Both machines are Windows Server (client is 2003, server is 2008) Administrator rights on both machines Valid SSL Certificate available FTP/IIS Server available and in use Required cryptation during transfer & storage Free space is not a problem Which software (both client and server side) you advice us to take?

    Read the article

  • How do private BitTorrent trackers monitor how much users upload/download?

    - by Jon-Eric
    From Wikipedia: Most private trackers monitor how much users upload or download, and in most situations, enforce a minimum upload-to-download ratio. How exactly can a tracker figure out how much data was uploaded and downloaded by each user? My understanding is that a BitTorrent tracker is merely a registry of users that are currently downloading/seeding and that peers, once connected, transfer data directly. So I wouldn't think that the tracker would know anything about the amount of data transferred, much less, where it came from.

    Read the article

  • Looking for network monitoring software for Win7 x64 (similar to the one included with Sunbelt perso

    - by rep_movsd
    I've upgraded from Windows XP to Windows 7 (64bit) and I found that the very convenient Sunbelt Kerio Personal Firewall will not work on this version. I hear the Windows 7 firewall is good enough(even though it never prompts for outbound connections) but Kerio had a nice network monitor feature which showed all applications with inbound and outbound connections and the current speed and bytes transferred for each one. Is there any software out there that can give me similar monitoring as what Kerio does? TCPView is almost as good, but doesnt show transfer statistics.

    Read the article

  • Nginx + Passenger running a RoR app is returning 401 when 302 is expected

    - by DBruns
    I've got a RoR app running on Passenger on top of Nginx. I'm using devise for my authentication method and have a link that gets sent in an email to users that requires authentication to view. If a user clicks the link from Outlook, and IE is the default browser, IE makes an HTTP request using the following headers: GET http://www.company.com/custom_layouts/108 HTTP/1.1 Accept: */* Accept-Language: en-us User-Agent: Mozilla/4.0 (compatible; MSIE 8.0; Windows NT 6.1; Trident/4.0; SLCC2; .NET CLR 2.0.50727; .NET CLR 3.5.30729; .NET CLR 3.0.30729; Media Center PC 6.0; InfoPath.2; .NET4.0C; .NET4.0E) Accept-Encoding: gzip, deflate Connection: Keep-Alive Host: www.company.com Returning: HTTP/1.1 401 Unauthorized Content-Type: /; charset=utf-8 Transfer-Encoding: chunked Connection: keep-alive Status: 401 X-Powered-By: Phusion Passenger (mod_rails/mod_rack) 2.2.15 WWW-Authenticate: Basic realm="Application" Cache-Control: no-cache X-UA-Compatible: IE=Edge,chrome=1 Set-Cookie: _vxwer_session=[sessionstr]; path=/; HttpOnly X-Runtime: 0.011918 Server: nginx/0.7.67 + Phusion Passenger 2.2.15 (mod_rails/mod_rack) 31 You need to sign in or sign up before continuing. 0 When the exact same URL is typed into the address bar, it does this: GET http://www.company.com/custom_layouts/108 HTTP/1.1 Accept: image/jpeg, application/x-ms-application, image/gif, application/xaml+xml, image/pjpeg, application/x-ms-xbap, application/vnd.ms-excel, application/vnd.ms-powerpoint, application/msword, */* Accept-Language: en-US User-Agent: Mozilla/4.0 (compatible; MSIE 8.0; Windows NT 6.1; Trident/4.0; SLCC2; .NET CLR 2.0.50727; .NET CLR 3.5.30729; .NET CLR 3.0.30729; Media Center PC 6.0; InfoPath.2; .NET4.0C; .NET4.0E) Accept-Encoding: gzip, deflate Connection: Keep-Alive Host: www.company.com Returning: HTTP/1.1 302 Found Content-Type: text/html; charset=utf-8 Transfer-Encoding: chunked Connection: keep-alive Status: 302 X-Powered-By: Phusion Passenger (mod_rails/mod_rack) 2.2.15 Location: http://www.company.com/users/sign_in Cache-Control: no-cache X-UA-Compatible: IE=Edge,chrome=1 Set-Cookie: _xswer_session=[session_info_here]; path=/; HttpOnly X-Runtime: 0.010798 Server: nginx/0.7.67 + Phusion Passenger 2.2.15 (mod_rails/mod_rack) 6f <html><body>You are being <a href="http://www.company.com/users/sign_in">redirected</a>.</body></html> 0 I expect them to return the same thing regardless.

    Read the article

  • What are NS records?

    - by Deano
    I am changing host and my new host said: Meanwhile, in order your newly registered hosting account to function correctly until the transfer is ready, you can change the NS records to the following: ns1.itahost.com ns2.itahost.com What does this mean?

    Read the article

  • scp stalls and ssh sessions freeze up (but eventually start again)

    - by coleifer
    I am running ubuntu on various computers on a home network. Some are on 9.04x64, some 10.04x64 and one 9.04x32. Running scp with a large file starts out at 2.1 mbps and drops down to about 200k, stalling and dropping until the transfer is complete. I've noticed this when I have a secure shell open on any of these servers as well. I have tried this with 2 different routers, both brand new, different brands.

    Read the article

  • Spotlight Infinite Indexing issue (external data drive)

    - by Manca Weeks
    This is an external drive, formerly a boot drive which is now in use only to access music files (sibelius, audio, midi, live, logic etc.) without transferring the data into a new boot system, partly because of the issue I am about to describe, but mostly because the majority of the data is mainly there for archival purposes. The user is a composer and prominent musician and needs to be able to rehash the data at will. I have tried several things - here is a list: - make complete filesystem clone with antonio diaz's ddrescue - run Disk Warrior on copy, repair whatever errors occurred - wipe out all ACLs on entire drive - set all permissions to the same value - wide open 777 - remove any system data (applications, system files, including hidden files to the best of my knowledge) by selecting only non-system/app data and using Carbon Copy Cloner to put only the data of interest onto a newly formatted drive - transfer data to newly formatted drive folder by folder, resetting the spotlight index in between adding each to observe for issues (interesting here is that no issues occurred except for in Documents folder - when I transferred only the Documents folder to a newly formatted drive on its own - no trouble. It appears almost as thought it may not be the content but the quantity or specific combination of data that results in problems) - use DataRescue to transfer the data to yet another newly formatted drive to expose any missed hidden files Between each of the above steps I stopped Spotlight (search for anything beginning with md in Activity Monitor - All Processes and quitting it), deleted the .Spotlight-V100 directory from the affected drive. Restart Splotlight indexing by adding drive to Spotlight privacy list and removing it. In each case the same issue occurs - Spotlight begins indexing normally (or so it seems), then the index estimated time increases, usually to 4 hours remaining. This is where it gets stuck and continues to predict 4 hours remaining but never finishes. Sometimes I can't eject the drive and have to quit the md.. processes from Activity Monitor to be able to eject the drive without Force Eject. Once I disconnect the drive after the 4 hours remaining situation - if I reattach it, Spotlight forever estimates remaining time and never gets going again. So there it is. It is apparently not a filesystem issue, not a permissions issue and not tied to any particular piece of hardware or protocol (used USB and FW drives). I have tried this on several machines (3 to be precise) and in 10.5.8 and 10.6.5. Simply disabling Spotlight on this volume is not an option because the owner has no clue where things are as the data on the volume dates back to music projects and compositions from 2003 and before. He needs to be able to query for results. Anyone got any ideas? Thanks, M

    Read the article

  • recommendations for efficient offsite remote backup solution of vm's

    - by senorsmile
    I am looking for recommendations for backing up my current 6 vm's(and soon to grow to up to 20). Currently I am running a two node proxmox cluster(which is a debian base using kvm for virtualization with a custom web front end to administer). I have two nearly identical boxes with amd phenom II x4's and asus motherboards. Each has 4 500 GB sata2 hdd's, 1 for the os and other data for the proxmox install, and 3 using mdadm+drbd+lvm to share the 1.5 TB's of storage between the two machines. I mount lvm images to kvm for all of the virtual machines. I currently have the ability to do live transfer from one machine to the other, typically within seconds(it takes about 2 minutes on the largest vm running win2008 with m$ sql server). I am using proxmox's built-in vzdump utility to take snapshots of the vm's and store those on an external harddrive on the network. I then have jungledisk service (using rackspace) to sync the vzdump folder for remote offsite backup. This is all fine and dandy, but it's not very scalable. For one, the backups themselves can take up to a few hours every night. With jungledisk's block level incremental transfers, the sync only transfers a small portion of the data offsite, but that still takes at least a half an hour. The much better solution would of course be something that allows me to instantly take the difference of two time points (say what was written from 6am to 7am), zip it, then send that difference file to the backup server which would instantly transfer to the remote storage on rackspace. I have looked a little into zfs and it's ability to do send/receive. That coupled with a pipe of the data in bzip or something would seem perfect. However, it seems that implementing a nexenta server with zfs would essentially require at least one or two more dedicated storage servers to serve iSCSI block volumes (via zvol's???) to the proxmox servers. I would prefer to keep the setup as minimal as possible (i.e. NOT having separate storage servers) if at all possible. I have also briefly read about zumastor. It looks like it could also do what I want, but it appears to have halted development in 2008. So, zfs, zumastor or other?

    Read the article

  • Change a Munin server and keep the data

    - by Khelben
    We are migrating some servers, and we need tp change our Munin server. Most of the Munin nodes are not changed, and we would want to keep track of the historical data, if possible. I can set up a new Munin server, but I like to know if it's possible to transfer the old data to the new server, and how to do it.

    Read the article

< Previous Page | 80 81 82 83 84 85 86 87 88 89 90 91  | Next Page >