Search Results

Search found 16386 results on 656 pages for 'flash drive'.

Page 357/656 | < Previous Page | 353 354 355 356 357 358 359 360 361 362 363 364  | Next Page >

  • boot failure trying to install linux distro from a cd

    - by jdamae
    Hoping someone can help me get this installation from a cd going. I'm using an older pc: hp pavillion 2.66Ghz, 512RAM with a BIOS revision of 6/30/2003. I reclaimed some an older drive (Seagate ST340810A) that seems to be working as its recognized in the bios (auto-detected) I downloaded a mini.iso of ubuntu 10.10 that I want to install and burned the image to a CD for install. My boot sequence is: First Boot Device [CDROM] and I disabled devices 2-4 so I can just force it to read first from the CDROM. This old pc also has a separate CD writer which is a Sec.Slave, so the Sec.Master is the Toshiba DVD/ROM DSM-171 drive where I placed the burned cd with the linux distro. So, with these settings I cannot get it to boot. I get the "DISK BOOT FAILURE, INSERT SYSTEM DISK AND PRESS ENTER". Where do I go from here? Thanks.

    Read the article

  • CSC folder data access AND roaming profiles issues (Vista with Server 2003, then 2008)

    - by Alex Jones
    I'm a junior sysadmin for an IT contractor that helps small, local government agencies, like little towns and the like. One of our clients, a public library with ~ 50 staff users, was recently migrated from Server 2003 Standard to Server 2008 R2 Standard in a very short timeframe; our senior employee, the only network engineer, had suddenly put in his two weeks notice, so management pushed him to do this project before quitting. A bit hasty on management's part? Perhaps. Could we do anything about that? Nope. Do I have to fix this all by myself? Pretty much. The network is set up like this: a) 50ish staff workstations, all running Vista Business SP2. All staff use MS Outlook, which uses RPC-over-HTTPS ("Outlook Anywhere") for cached Exchange access to an offsite location. b) One new (virtualized) Server 2008 R2 Standard instance, running atop a Server 2008 R2 host via Hyper-V. The VM is the domain's DC, and also the site's one and only file server. Let's call that VM "NEWBOX". c) One old physical Server 2003 Standard server, running the same roles. Let's call it "OLDBOX". It's still on the network and accessible, but it's been demoted, and its shares have been disabled. No data has been deleted. c) Gigabit Ethernet everywhere. The organization's only has one domain, and it did not change during the migration. d) Most users were set up for a combo of redirected folders + offline files, but some older employees who had been with the organization a long time are still on roaming profiles. To sum up: the servers in question handle user accounts and files, nothing else (eg, no TS, no mail, no IIS, etc.) I have two major problems I'm hoping you can help me with: 1) Even though all domain users have had their redirected folders moved to the new server, and loggin in to their workstations and testing confirms that the Documents/Music/Whatever folders point to the new paths, it appears some users (not laptops or anything either!) had been working offline from OLDBOX for a long time, and nobody realized it. Here's the ugly implication: a bunch of their data now lives only in their CSC folders, because they can't access the share on OLDBOX and sync with it finally. How do I get this data out of those CSC folders, and onto NEWBOX? 2) What's the best way to migrate roaming profile users to non-roaming ones, without losing vital data like documents, any lingering PSTs, etc? Things I've thought about trying: For problem 1: a) Reenable the documents share on OLDBOX, force an Offline Files sync for ALL domain users, then copy OLDBOX's share's data to the equivalent share on NEWBOX. Reinitialize the Offline Files cache for every user. With this: How do I safely force a domain-wide Offline Files sync? Could I lose data by reenabling the share on OLDBOX and forcing the sync? Afterwards, how can I reinitialize the Offline Files cache for every user, without doing it manually, workstation by workstation? b) Determine which users have unsynced changes to OLDBOX (again, how?), search each user's CSC folder domain-wide via workstation admin shares, and grab the unsynched data. Reinitialize the Offline Files cache for every user. With this: How can I detect which users have unsynched changes with a script? How can I search each user's CSC folder, when the ownership and permissions set for CSC folders are so restrictive? Again, afterwards, how can I reinitialize the Offline Files cache for every user, without doing it manually, workstation by workstation? c) Manually visit each workstation, copy the contents of the CSC folder, and manually copy that data onto NEWBOX. Reinitialize the Offline Files cache for every user. With this: Again, how do I 'break into' the CSC folder and get to its data? As an experiment, I took one workstation's HD offsite, imaged it for safety, and then tried the following with one of our shop PCs, after attaching the drive: grant myself full control of the folder (failed), grant myself ownership of the folder (failed), run chkdsk on the whole drive to make sure nothing's messed up (all OK), try to take full control of the entire drive (failed), try to take ownership of the entire drive (failed) MS KB articles and Googling around suggests there's a utility called CSCCMD that's meant for this exact scenario...but it looks like it's available for XP, not Vista, no? Again, afterwards, how can I reinitialize the Offline Files cache for every user, without doing it manually, workstation by workstation? For problem 2: a) Figure out which users are on roaming profiles, and where their profiles 'live' on the server. Create new folders for them in the redirected folders repository, migrate existing data, and disable the roaming. With this: Finding out who's roaming isn't hard. But what's the best way to disable the roaming itself? In AD Users and Computers, or on each user's workstation? Doing it centrally on the server seems more efficient; that said, all of the KB research I've done turns up articles on how to go from local to roaming, not the other way around, so I don't have good documentation on this. In closing: we have good backups of NEWBOX and OLDBOX, but not of the workstations themselves, so anything drastic on the client side would need imaging and testing for safety. Thanks for reading along this far! Hopefully you can help me dig us out of this mess.

    Read the article

  • Domino 8.5.3: Attaching an Object Residing on Server (Lotusscript preferred)

    - by Void
    Not sure if this question is more appropriate for ServerFault or StackOverflow, sorry if it should belong elsewhere! I am working on an application and one of the function is to automatically send an email with an attachment. I can code the application to attach the object when it resides on local or on a mapped drive. Newbie Question: Is there a way to have the object reside on the Domino server, and still be able to point to it and have the application automatically attach and send? Is there any method that allows me to do this? Users have no direct access to the server/filesystem, so mapped drive of the Domino server is out of the question. Hope someone can shed some light on this question. Thanks!

    Read the article

  • USB AutoRun Dialog: Open a webpage option not shown in dialog

    - by Jake M
    I am attempting to make our webpage open when someone plugs in our USB device. My Problem: When we plug in the USB device the autorun dialog appears but there is never an option to open/run the webpage we have specified. See below pic of the dialog we see. Isn't there meant to be an option that says something like "Open in Internet Explorer" or etc.? What we are hoping to achieve is to have another option below 'Use this drive for backup' that says something like 'Open in Web Browser' or something that will ultimately open our webpage when the users decides to. Heres our code that is inside the file autorun.inf which is placed on the USB device: [autorun] shellexecute=http://exds-test.epicservices.com.au/V10InstallationInfo.aspx action=Open Website label=EXDS USB Drive

    Read the article

  • Annoying Search Behavior - Search Companion

    - by DavidStein
    I'm running Windows XP Professional, and ever since the last service pack I've had a searching problem. When I want to search a network drive, I get the following message: This folder is not indexed. To search this directory plase use Search Companion or add this directory to your index via options. Basically, I have two questions. Is there some way that I can use the indexed search where appropriate and then have it switch over to the Search Companion automatically? Second, how does a programmer look at this code and think this is a good idea? I realize that this question is rhetorical. However, I must enter my search string into one search, receive the error, and then click "Search Companion" to bring up the new search window. This window doesn't even take the defaults from the previous one so I have to specify the search string and drive again.

    Read the article

  • My laptop can connect to every wireless network except fios

    - by going crazy
    I have always been able to connect to every wireless router secured or unsecured wep or wpa. I had Fios installed and could not connect. Verizon suggested it was my computer and gave me an outside wirless drive to use, it worked. I got rid of fios and went back to comcast and threw out the drive, but now 2 years later, I am sitting at my friends house haveing the same problem. My tech savy friend told me it is a firewall setting or something in my antivirus software, but I disabled them both and still nothing works........Funny it is only FIOS

    Read the article

  • Can't boot into XP after setting up dual boot with Win 7 vhd.

    - by bebop
    I set up a dual boot for win 7 from vhd on 2 xp machines. One of them everything went fine, and I get the option to choose the os when I turn the machine on. The other was slightly different in that rather than seeing the current OS disk (the one with XP on) as the c:\ when I was setting up windows 7 vhd during install it saw the disk as d:. I didn't think anything of it and went ahead and created a vhd on the d: drive. Now when I turn this machine on, it boots straight to win7 and I never get the option to choose xp. When I look at the boot option in msconfig, I only see Windows 7. How can I go about adding the old XP drive as a boot option at startup again? Edit: Strangely when I rebooted this time (perhaps the first time since I removed the install DVD) it boots to XP. I suppose I'll just have to reinstall windows 7 again in a new vhd...

    Read the article

  • Is it possible at all to install Ubuntu 10.04 on Windows 7 (64-bit) using its Virtual PC?

    - by Jian Lin
    It was said that Win 7's Virtual PC is not suitable for installing Ubuntu 10.04... Is there any method at all that it will work? The following is the scenario I ran into: The first time Ubuntu 10.04 installation CD-R boots up, it asked for the Language, and "Install Ubuntu" and then the screen has vertical green bars and then the VPC just closed. The 2nd or 3rd time it booted up, there is no asking of Language or "Install Ubuntu" and just shut down the VPC, sometimes with vertical green bars. I even created another new hard drive and same thing happened. And created VPC 02, and same thing happened. Created VPC 03 with a fixed hard drive size of 60GB and same thing happened.

    Read the article

  • Windows 2003 Server - Can I map a folder to another folder on the same server?

    - by TheCleaner
    Scenario: I have a server that was running low on space. We have an external iscsi SAN that it now has a LUN on and connected to the server as E:\ We are moving the PHOTOS folder from the old location on D:\ to the new E:. That new drive is being shared out as "ARCHIVE". So: \server\shared\photos now becomes \server\archive\photos I can easily place a shortcut in the original location saying "DOUBLE CLICK HERE FOR THE PHOTOS", but it isn't ideal. What I'd like is to be able to have \server\shared\photos simply point to \server\archive\photos. So that if someone wants to map a drive to \server\shared and then browse to the photos folder once in there they will see what is in the \server\archive\photos location. Is that possible? I was thinking about SUBST or DFS, but I don't think either of those will do it.

    Read the article

  • Intermittent USB 3.0 access - How do I troubleshoot?

    - by Billy ONeal
    I've got a WD Passport hard disk with "superspeed" USB 3.0 support. When I use my USB 3.0 flash drive (this is a Lenovo X220 laptop), USB 3.0 consistently works. But when I use the passport drive, almost without fail the connection drops to USB 2.0. Touching the cable seems to immediately trigger the problem, but it seems to happen on its own though. I've got another cable on order right now... but is it likely the cable's the issue here? Is there anything else I can check?

    Read the article

  • Vantec NexStar NAS Encloser - Writing large files

    - by peter
    I have one of these 'Vantec NexStar LX - NST-475LX-BK' drive enclosures. It is a NAS device. When I write a file to the device using eSata, or a SMB share I cannot write files over 4GB. I think this is because the drive is formatted with FAT32. But when I access the device using FTP it doesn't matter. I can write files of any size. E.g. I wrote one on there last night which was 30GB. Does this make any sense? Why? I guess the most important thing for me is data integrity.

    Read the article

  • Too Many Files In Debian Linux Folder?

    - by Dave Potts
    I've been using an external USB drive on a Debian server for backup. The drive is formatted as NTFS and mounted with ntsfmount. This was working fine, but I was filling up a directory with lots of files. Eventually the backup failed. When I then tried to look at the directory using ls it reported: ls: reading directory .: Numerical result out of range Looking in syslog, I also saw this: Sep 23 07:35:31 tosh ntfsmount[28040]: Failed to read index block: Numerical result out of range. Is this simply that I've reached the upper limit of number of files in a directory? If so, is there any way to extend the number of allowed files?

    Read the article

  • Macbook Pro 2.2ghz 2011 (OSX 10.6.7) problem with NTFS 3G

    - by James
    I installed NTFS 3G but now get the following error message when I try to plug in my external drive. I also get it on startup about my Windows partition. Uninstall/ reinstall does not work. NTFS-3G could not mount /dev/disk1s1at /volumes/freeagent GoFlex Drive because the following error occured: /library/filesystems/fuse.fs/support/fusefs.kext failed to load- (libkern/kext) link error; check the system/ kernel logs for errors or try kextutil(8). The MacFUSE file system is not available (71) Any help would be great. I'd hope to avoid reinstalling OS X if possible!

    Read the article

  • Hardware RAID 0 without OS re-installation

    - by sterz
    I have Ubuntu & Windows 7 installed on my hdd. Can I mirror the image of the hdd to the second identical drive? Is this not recommended (i.e have to re-install every OS)? If it is okay to mirror, is there anything else to do to make hw RAID 0 work? Does RAID 0 have the same risk as a single drive? What sector size would you recommend for read/write/extract video files (mostly each around 2 GB)?

    Read the article

  • Alternatives to Crashplan for VPS?

    - by Chloe
    I use SFTP Net Drive to mount a remote VPS so I can back it up. However, it's taken over 3+ days to scan! I ran 'ls -lR' from my desktop over the mounted network drive and it only took about 5m to list all the files! There are only about 5000 files and 2 GB. I know Crashplan can run headless on the VPS itself, but that sounds like a pain to set up, and it takes so much memory on the server. The VPS doesn't have a lot of memory to spare - it's less than my desktop. Is there another program that can communicate with a Crashplan backup protocol and has a command line interface? backup /home

    Read the article

  • Clarification On Write-Caching Policy, Its Underlying Options And How It Applies To Hard Drives And Solid-State Drives

    - by Boris_yo
    In last week after doing more research on subject matter, I have been wondering about what I have been neglecting all those years to understand write-caching policy, always leaving it on default setting. Write-caching policy improves writing performance and consists of write-back caching and write-cache buffer flushing. This is how I understand all the above, but correct me if I erred somewhere: Write-through cache / Write-through caching itself is not a part of write caching policy per se and it's when data is written to both cache and storage device so if Windows will need that data later again, it is retrieved from cache and not from storage device which means only improved read performance as there is no need for waiting for storage device to read required data again. Since data is still written to storage device, write performance isn't improved and represents no risk of data loss or corruption in case of power failure or system crash while only data in cache gets lost. This option seems to be enabled by default and is recommended for removable devices with no need to use function of "Safely Remove Hardware" on user's part. Write-back caching is similar to above but without writing data to storage device, periodically releasing data from cache and writing to storage device when it is idle. In my opinion this option improves both read and write performance but represents risk if power failure or system crash occurs with the outcome of not only losing data eventually to be written to storage device, but causing file inconsistencies or corrupted file system. Write-back caching cannot be enabled together with write-through caching and it is not recommended to be enabled if no backup power supply is availabe. Write-cache buffer flushing I reckon is similar to write-back caching but enables immediate release and writing of data from cache to storage device right before power outage occurs but I don't know if it applies also to occasional system crash. This option seem to be complementary to write-back cache reducing or potentially eliminating risk of data loss or corruption of file system. I have questions about relevance of last 2 options to today's modern SSDs in order to get best performance and with less wear on SSDs: I know that traditional hard drives come with onboard cache (I wonder what type of cache that is), but do SSDs also come with cache? Assuming they do, is this cache faster than their NAND flash and system RAM and worth taking the risk of utilizing it by enabling write-back cache? I read somewhere that generally storage device's cache is faster than RAM, but I want to be sure. Additionally I read that write-caching should be enabled since current data that is to be written later to NAND flash is kept for a while in cache and provided there is data that gets modified a lot before finally being written, holding of this data and its periodic release reduces its write times to SSD thereby reducing its wearing. Now regarding to write-cache buffer flushing, I heard that SSD controllers are so fast by themselves that enabling this option is not required, because they manage flushing. However, once again, I don't know if SSDs have their own onboard cache and whether or not it is faster than their NAND flash and system RAM because if it is, keeping this option enabled would make sense. Recently I have posted question about issue with my Intel 330 SSD 120GB which was main reason to do deeper research having suspicion of write-caching policy being the culprit of SSD's freezing issue assuming data being released is what causes freezes. Currently I have write-cache enabled and write-cache buffer flushing disabled because I believe SSD controller's management of write-cache flushing and Windows write-cache buffer flushing are conflicting with each other: Since I want to troubleshoot in small steps to finally determine the source of issue, I have decided to start with write-caching policy and the move to drivers, switching to AHCI later on and finally disabling DIPM (device initiated power management) through registry modification thanks to @TomWijsman

    Read the article

  • Unable to map to web folder using WebDAV client on Windows Server 2008 R2

    - by user74989
    Hello, I have a client running Windows Server 2008 R2 on several servers. One of the servers is also running SharePoint 3.0 and my client has created a web folder to map to. I can map to the web folder from all Server 2008 R2 boxes that have the WebDAV client (part of Desktop Experience feature) installed, except for the server the folder resides on. When I attempt to map to the web folder on the server which the folder resides, I am repeatedly prompted to enter my credentials. I am using the same account that I used to map the web folder on the other servers. I have also tried mapping from the command line and receive 'Access Denied' What may be causing the problem? I would think that if I can map to the drive from one server, I should be able to map the drive from the rest as long as the WebDAV client is installed, especially on the server where the folder is located. Jesse

    Read the article

  • Intel RAID0 on Windows 8 not Displaying Correct Media Type

    - by kobaltz
    I have my primary C Drive which consists of 2 Intel 120GB SSD Drives in a RAID0. I have a clean install of Windows 8 Pro, latest MEI software, latest RST software, latest Intel Toolbox. Prior to this I had installed Windows 8 Pro as an upgrade. When I went into the Optimize Drives while in the Upgrade installation, it showed the Media Types as Solid State Drives. However, now since I am in a brand new install, it is showing the Media Type as Hard Disk Drive. I am worried about this because of the trim not working properly. Before when in the upgrade, it showed SSD as the media type and the Optimize option would perform a manual trim. Unfortunately, my search credentials on Google are so common to many other things (ie Raid0, SSD, Windows 8, Media Type) that all I am finding are useless topics. Before, (found on random site) it showed the Media Type as below

    Read the article

  • Xen P2V for large physical hosts with much free space

    - by Sirex
    I need to P2V a rhel5 machine to xen under rhel5. I know I can use dd if=/dev/sda then using virt-install --import on the host, but the downside of this is the original machine has 80% free space on its drive. Does anyone know of (or can document) a quick and easy method which works reliably, to produce a bootable xen image which can run under a hvm in such cases ? I tried clonezilla to make the image, to avoid the free space problem, but it failed to do the clone with "something went wrong" (useless info, i know). At the moment im looking at doing a dd of each partition, and a file level copy of the partition which is mostly empty, then creating a new virtual disk, copying the partitions over to it by mounting both the new image and the virtual drive on a second vm, then copying the boot sectors over, then copying the file level backup..... there must be an easier way ? Oh, and budget is $0. :)

    Read the article

  • Environment variable for volume names in Windows?

    - by Shinrai
    I'm trying to write a batch file that will look at the volume label of the current drive and report if it's not equivalent to a certain string. Is there a default variable in the shell for this? Can I define one? Am I SOL and I'll actually have to do some (shudder) programming? EDIT: If this is possible in PowerShell that would work fine. (For the curious, we ship our machines with software cloning as a rapid bootable backup solution since most of our customers are daytraders and aren't interested in RAID due to urgency of getting-the-hell-back-to-work-right-away if there's a software corruption problem, and we want to make it immediately obvious if they're booting to the backup drive unintentionally, like say the primary failed entirely. The hope was just to write a simple batch file that would autostart on boot and throw a warning in the event of a problem.)

    Read the article

  • Unused Index Entries: What causes them?

    - by Bart Silverstrim
    I was running chkdsk on one of our servers and it fixed 800 unused index entries (minor inconsistencies) on the drive. I also see these on lab machines running Deep Freeze, where most drive changes are supposed to be prevented although I'm not sure of the mechanism through which it does this. The thing that got me more curious than usual is that this server does very little. Very little activity, it remotely serves out a web app with little use so it's rarely ever touched at the console...so what is it on NTFS drives that causes this? It seems to be treated as if it's nothing, but should this be happening on a barely used filesystem? Is it a sign of corruption? I would think it's something if a disk repair utility sees fit to report it and repair it. Maybe this is more curiosity while doing some routine maintenance on servers, but I'd like to know what this is doing and why it happens, if anyone can give insight on this phenomena.

    Read the article

  • decrypting a dvd and testing that it is decrypted and playing it locally with its menu?

    - by barlop
    I have a DVD , it may be decrypted. How do I get it to hard drive decrypt it and see if it is decrypted? I have got it to hard drive... I can get VOB files to play(does that mean it is not encrypted?) but I can't get it to play with the menu. I can get the DVD itself to play with a menu. I'd like to get the local copy to. Also, I can play the first VOB file in VLC and MPC but not in mplayer. But that's a secondary thing. I will use whatever player you suggest. I have tried VLC, and mplayer, and media player classic.

    Read the article

  • Understanding RewriteCond in .htacces files

    - by Paulo Bu
    I'm having problems understanding how RewriteCond directive works. So far, it's pretty clear that it compares to strings to apply a RewriteRule. I have this file: <IfModule rewrite_module> RewriteEngine on RewriteCond %{REQUEST_FILENAME} !-f RewriteRule ^(.*)$ app_dev.php </IfModule> This works for me but I don't know why it works. So far in the RewriteCond directive I understand: if the value of REQUEST_FILENAME is NOT a file in the hard drive then allow the rule This doesn't have sense becouse app_dev.php after substituting is a file in the hard drive. Anyways, could someone enlighten me with this issue? I am having a very harsh time figuring out how this works.

    Read the article

  • Newbie: VirtualBox "no bootable medium found"

    - by Newton Falls
    My first foray into VM land so please excuse the stupidity expressed below. Running VirtualBox on Kubuntu 9.10 and trying to set up a VM running Windows 7. Windows 7 came installed and is functional as a dual-boot but I want to stay in kubuntu hence the VM. Received "Fatal: no bootable medium found" when trying to start a VM. Not surprising since VBox never gave me an option (that I recognized anyway) to tell it to look on the hard drive. I did not receive a CD of Windows 7 with the machine. How can I tell VirtualBox to grab what it needs from the hard drive? Or am I out of luck because I don't have a CD? Update: This might not be possible. The installed Windows 7 is useful only to pull the installation key. It is easier to just to download or get an ISO from a friend and enter in your own key.

    Read the article

  • Slow write speeds on new Gigabit home file server

    - by Ryan Holder
    So I finally got all my parts delivered to setup a home file/backup server this week. It's currently running Ubuntu Server and I'm using Samba to share files on my network. The server currently has a 2TB WD Green drive in it connected to a Asus M5A78L-M This is then connected via CAT6a to my new Gigabit switch (TP-Link TL-SG1005D). My home desktop is then also connected to this switch and again also through CAT6a cable. Currently when transfering files I will get a perfect 100MB/s read from the server to my Windows machine. When copying from my Windows machine to the server I get around 30/38MB/s. I know this drive is capable is faster speeds so would anybody have an idea of where the bottleneck is? Any help would be greatly appreciated :) EDIT: I have found ftp's write speed is much closer to what my Samba read speed is so I'm going to give it a guess that is a software problem rather than hardware

    Read the article

< Previous Page | 353 354 355 356 357 358 359 360 361 362 363 364  | Next Page >