Search Results

Search found 2630 results on 106 pages for 'mount'.

Page 28/106 | < Previous Page | 24 25 26 27 28 29 30 31 32 33 34 35  | Next Page >

  • How to mount a LOFS in Solaris that doesn’t cross mountpoints

    - by jcea
    I need to access my "root" ZFS dataset to delete a file under "/var". But "/var" is overlayed by another ZFS dataset. Since these are system datasets I can't "umount" them while the machine is running. And I want to avoid to reboot the system in "failsafe" mode, since this is a production machine. Teorically ZFS would refuse to mount "/var" dataset over the underlying "/var", because it is not empty. But it works, possibly because they are system datasets mounted early in the boot process. But having the underlying "/var" not empty is preventing me to create an ABE (Alternate Boot Environment), so patching is risky, and I can't upgrade my system using Live Upgrade. The machine is remote. I have an IP KVM, but I rather prefer to avoid booting this machine in "failsafe" mode, if I can. I know there is a file in "/var/" because I can snapshot the "root" dataset and check it. But snapshots are read-only, so I can't get rid of the file. I tried "mkdir /tmp/zzz; mount -F lofs / /tmp/zzz", but when I go to "/tmp/zzz/var", I see the "/var" dataset, not the underlying "root" dataset. That is, the LOFS is crossing mountpoints. I would usually like it, but not this time!. Any suggestion, beside rebooting the machine in "failsafe" and mess with it thru the IP KVM?

    Read the article

  • Protocol to mount fat32 network filesystem on Linux with ability to lock files ( not advisory locks

    - by nagul
    I have a fat32 filesystem sitting on a NAS storage device (nslu2) that I need to mount on my Ubuntu system. I've tried Samba and NFS mounts, but both don't seem to support proper locking. More specifically, I am unable to save files to the mounted drive through GNUcash, KeepassX etc, which makes the share fairly useless. Is there a protocol that allows me to achieve this ? Note that the NAS storage device is running a linux OS so I can run pretty much any protocol that has a linux implementation. The only option I'm not looking for is to reformat the partition to ext3, which I'm not able to do due to other constraints. Alternatively, has anyone managed proper locking of a fat32 system over the network using Samba ? Or, is advisory locking the best you get with a network-mounted fat32 file system ? I've thought of trying sshfs but I've not found any indication that this will solve my problem. Edit: Okay, maybe I can reformat the drive, but to any file system except ext3. The "unslung" nslu2 doesn't like more than one ext3 drive, and I already have one attached. So any solution that involves reformatting the drive to ntfs, hfs etc is fine, as long as I can mount it on linux and lock files.

    Read the article

  • mounting SD card on android emulator

    - by Lo'oris
    On the emulator, I can unmount the SD card from the Settings. On OSX, I can then mount it using hdiutil, then unmount it normally. I haven't been able to figure out how to re-mount it then on the emulator (without rebooting it). hints: the adb command "remount" is unrelated: it's about /system the emulator command is unrelated: it's only about starting the emulator mounting the SD card on OSX without unmounting it from the emulator, causes horrible pain to the emulator

    Read the article

  • (re)mounting the SD card on android emulator

    - by Lo'oris
    On the emulator, I can unmount the SD card from the Settings. On OSX, I can then mount it using hdiutil, then unmount it normally. I haven't been able to figure out how to re-mount it then on the emulator (without rebooting it). hints: the adb command remount is unrelated: it's about /system the emulator command is unrelated: it's only about starting the emulator mounting the SD card on OSX without unmounting it from the emulator causes horrible pain to the emulator

    Read the article

  • read and write permission for FAT32 partition in Ubuntu

    - by Dean
    This is a strange problem. I have the following partition table Device Boot Start End Blocks Id System /dev/sda1 * 1 13 102400 7 HPFS/NTFS Partition 1 does not end on cylinder boundary. /dev/sda2 13 5737 45978624 7 HPFS/NTFS /dev/sda3 5738 10600 39062047+ 83 Linux /dev/sda4 10601 19457 71143852+ 5 Extended /dev/sda5 10601 11208 4883728+ 82 Linux swap / Solaris /dev/sda6 11209 15033 30720000 b W95 FAT32 /dev/sda7 15033 19457 35537920 7 HPFS/NTFS I dual boot Win7 (sda2) and Ubuntu (sda3) and wanted to use the FAT23 partition to share files across two OS's. I followed some online tutorial and have done these: sudo mkdir /media/FAT32 sudo chmod 777 /media/FAT32 sudo mount /dev/sda6/ /media/FAT32 after I mounted the file, I can only read but not be able to write to it. I checked the file permission, it becomes: drwxr-xr-x but after I unmounted the it then becomes drwxrwxrwx and I can read and write to it. very strange. I don't know where I've down wrong. Cheers.

    Read the article

  • Second partition with Windows 7 unmountable

    - by Florian Pilz
    I'm using GRUB2 to dualboot Ubuntu and Windows 7. I installed Windows 7 some days ago and rewrote GRUB2 to cope with the dualboot. Everything went fine till yesterday - I could boot into both OS. Since this morning my laptop restarts every time I choose Windows 7 from the menu. Ubuntu is still working. As I tried to mount the partition with Windows 7 on Ubuntu it gave me return code 2, so this isn't working as well. I tried to reinstall GRUB2 to the MBR but it didn't help. I also tried to repair Windows 7 boot with it's install DVD, while trying so it showed me the following error: "Status: 0xc000000f Info: Boot selection failed because a required device is inaccessible." Tanks in advance

    Read the article

  • fdisk -l shows a partition is not in /dev directory

    - by zitronic
    I am having troubles on mounting my ntfs hard drive on linux. I am running ophcrack live CD. fdisk shows me the hard drive I have installed on my computer fdisk -l Disk /dev/hdc: 30.0 GB, 30005821440 bytes 255 heads, 63 sectors/track, 3648 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/hdc1 * 1 3647 29294496 7 HPFS/NTFS but I dont have that device on my /dev directory so I can not mount it ls /dev/hd* /dev/hda /dev/hda4 /dev/hda8 /dev/hdb2 /dev/hdb6 /dev/hdc /dev/hda1 /dev/hda5 /dev/hda9 /dev/hdb3 /dev/hdb7 /dev/hdd /dev/hda2 /dev/hda6 /dev/hdb /dev/hdb4 /dev/hdb8 /dev/hda3 /dev/hda7 /dev/hdb1 /dev/hdb5 /dev/hdb9 What could be going on_

    Read the article

  • Dealing with different usernames when mounting removable media in Linux

    - by dimatura
    I have a laptop in which my username is, say, "foo". I have an external drive, formatted with Ext4, for which all files are owned by "foo" (at a filesystem level). Now, I have a desktop in which my username is, say, "bar". If I mount this external drive in this computer the files are considered to not be owned by "bar". This makes sense, but it is annoying because their bits mode are set so that only the owner can modify/delete them. What's the cleanest way to deal with this? Create a group with "foo" and "bar" and add group modification permissions?

    Read the article

  • OSX 10.6 Give Apache2 read&write access to mounted windows share

    - by JohEngstrom
    On Mac OS X Snow Leopard I'm trying to give the apache2 user _www full rights to a mounted hidden windows server share. I've used Connect to Server with smb://servername/share$ and saved the username/password in the keychain. The domain username used for the mount got full rights to the share on the windows server. It all works this far. I can browse and edit the files in the share from the Mac. However I can't find a way to give the apache2 user _www rights to write to the mounted share. I have a perl script that is supposed to create a file in the mounted folder but only get permission denied. I've tried all kinds of chmod and chown but it doesn't change the permissions of the share. Does anyone know how this can be done please?

    Read the article

  • NTFS-3G is only mounting external drives as read-only

    - by Phanto
    I'm currently running RHEL 5.5, and I installed the ntfs-3g utility from here: http://www.tuxera.com/community/ntfs-3g-download/. I have also followed their instructions for auto-mounting NTFS USB drives here: http://www.tuxera.com/community/ntfs-3g-faq/#plugandplay. The problem I'm experiencing is that ntfs-3g is automatically mounting as root. In order for me to obtain write support, I need to navigate to the mounted device as root, and perform write actions with elevated privileges. Is there a way to mount USB NTFS volumes automatically without needing to sudo every write command? Thanks!

    Read the article

  • How to determine the used size of device associated's buffer

    - by dubbaluga
    Hi, when mounting a device without the "sync" option, e. g. by invoking the following: mount -o async /dev/sdc1 /mnt a buffer is associated with a device to optimize (speed) read/write operations. Is there a way to determine the size of this buffer? Another question that comes into my mind is, if it's possible to find out how much of it is used currently. This can be interesting to determine the time it would take to "sync" or "umount" slow devices, such as flash-based media. Thanks in advance for your answers, Rainer

    Read the article

  • Virtual box - How to add disks and move var, opt and home to them?

    - by Jarrod Roberson
    I created a CentOS 5.6 Guest OS Virtual Machine. I made the first disk 10GB, I am rapidly outgrowing it. It was suggested that I make disks for my /var, /opt and /home directories and move them so I can better manage the disks for backing up and what not. This sounds like a good idea. I know how to create the disks in Virtual Box. I have dug around Google and the internet in general and all my attempts at doing this have failed. Snapshots are awesome! I can get the drives fdisked, and I have had limited success mounting them to /mnt/var, /mnt/home and /mnt/opt, but even in single user mode ( init 1 ) I can't get the entire contents of the directories to move over, and then the machine won't reboot correctly. cd /var cp * -ax /mnt/var The /var directory in particular is not wanting to move everything to the new location. How do I format, mount and move the /var, /opt and /home to my new disks?

    Read the article

  • How to find the /dev name of my USB device

    - by mustafa
    I am running Ubuntu 11 on VmWare on Windows XP. I want to format an SD card in Ubuntu. But, I can't figure out which /dev/xxx device the SD card is. I plug the card into the built-in socket of my laptop. I "safely remove" the device in Windows. Then, I "connect" the PCMCIA reader in VmWare. Now, I was supposing to see a new device like /dev/sdx. But, it doesn't appear :( How can I find what the name of my USB device's name and mount it? /var/log/message is empty. Here is the output of dmesg: [ 5268.927308] usb 2-1: new full speed USB device number 12 using uhci_hcd And, here is the last lines of /var/log/syslog: Oct 31 18:51:21 ubuntu kernel: [ 5268.927308] usb 2-1: new full speed USB device number 12 using uhci_hcd Oct 31 18:51:21 ubuntu mtp-probe: checking bus 2, device 12: "/sys/devices/pci0000:00/0000:00:11.0/0000:02:00.0/usb2/2-1" Oct 31 18:51:21 ubuntu mtp-probe: bus: 2, device: 12 was not an MTP device

    Read the article

  • Any good method for mounting Hadoop HDFS from another system?

    - by Beel
    I want to mount the Cloudera Hadoop as a Linux file system over the LAN. As a setup, I already have the hadoop cluster running on a set of Ubuntu machines. But now I need to be able to use it as a normal file system from a Fedora system over the LAN. I tried FUSe but two things: 1. Cloudera says FUSE loses data (click here for that comment by a Cloudera employee on the official Cloudera support site) 2. I've had no success making it work the way we want As a point of clarification, I am using Hadoop ONLY for the file system, not for its other capabilities.

    Read the article

  • fdisk -l only displays boot partition

    - by Franklin
    I have a SAN, and it's able to read and write to the 50TB RAID just fine, but when I run fdisk -l it only lists the boot partition of the SAN server, and doesn't display anything about the other partitions on the RAID. I've also tried using parted -l with the same result. Now when I type mount it shows that the partitions are mounted just fine. I've never seen this happen. The box is running Openfiler 2.3 (I know it's old, we're in the process of upgrading all our old equipment). We have another SAN that's configured almost identically, and it's able to display the partition info with either of the two commands I mentioned above.

    Read the article

  • fstab and cifs mounting, possible to store authentication information outside of fstab?

    - by tj111
    I am currently using cifs to mount some network shares (that require authentication) in /etc/fstab. It works excellently, but I would like to move the authentication details (username/pass) outside of fstab and be able to chmod it 600 (as fstab can have issues if I were to change its permissions). I was wondering if it is possible to do this (many-user system, don't want these permissions to be viewable by all users). from: //server/foo/bar /mnt/bar cifs username=user,password=pass,r 0 0 to: //server/foo/bar /mnt/bar cifs <link to permissions>,r 0 0 (or something analogous to this). Thanks.

    Read the article

  • HFS partition mounting read-only

    - by Sid
    Hey, I have an external Western Digital Hard drive with two HFS partitions with journaling disabled. When I connect it to a computer running Linux (Debian or Ubuntu), frequently both partitions are mounted read-only. In the past, mounting them on my Macbook and executing the command to disable the journaling often worked (even though it would tell me that journaling was already disabled) but I would love to have a solution which works every time. Thanks! Edit: In light of Chris Johnsen's comment below - my question is how to mount the filesystem read+write on Linux since it is not automatically doing so itself

    Read the article

  • What are my options for a disk with what seems to be a corrupted filesystem?

    - by CT
    I have a friend with an old Dell that will not boot into Windows. It has an IDE drive. It spins up. I have an IDE to USB device. I've attached the drive via that device to a working laptop. The drive does not mount. If I go into Disk Management I can see the drive but it will not initalize, says "Drive not ready." I've also booted into a linux live cd to see if the drive mounts, it does not. I am just trying to recover some pictures from the drive. The data is not important enough to send to a professional. The issue is more of a curosity on how to recover data if and when these situations would occur in the future.

    Read the article

  • Slow speed for UFS mounted drive in Linux

    - by Incredible
    Hi, I have a disk that has Sun OS disk, (ufs filesystem). And I want to mount it in my debain machine with read/write mode. Since by deafult linux doesn't support write to ufs filesystem. I had to recompile the kernel by setting to the flag CONFIG_UFS_FS_WRITE=y. Now I am able to write to the filesystem, but the read/write speed is very slow. It is around 120 KB/s. Any idea what is wrong and how to resolve this issue? Thank you in advance.

    Read the article

  • Drobo Pros won't mount via iSCSI

    - by niklassaers
    Hi guys, I've got an Xserve where I've configured a Drobo Pro. Connected via firewire and USB, it works fine, but when connected to a separate network port, it doesn't mount. I've got the Drobo Pro configured to 2.0.1.2/255.255.0.0 and the network port to 2.0.0.1/255.255.0.0. I can ping the IP of the Drobo Pro just fine. I've installed Drobo Dashboard 1.6.8. Cheers Nik

    Read the article

  • how to mount a partition inside a partition

    - by facha
    Hello, everyone I have a block device (/dev/sda5) that has been partitioned inside by a virtual machine. So, when I look inside with fdisk /dev/sda5, I see: sda5p1 sda5p2 and so on. Is it possible to mount them on my host system? Thanks in advance.

    Read the article

  • Mount an additional Xserve volume with autofs on Linux

    - by daustin777
    A few years ago I setup autofs on a RH Linux box to mount volumes from four XServes. I need to add a couple new volumes from these same Xserves so that I can access files from the Linux box. I've completely forgotten how to do this and haven't been able to find a solution online. How do I add the new volumes? Do I need to add paths to the new volumes?

    Read the article

  • (Amazon AWS) EBS mount error: Stale NFS file handle

    - by May
    I have an EC2 instance that just went offline (cannot even be pinged) but is still reflected as operational. In an effort to retrieve data stored on an attached EBS, I did a forced detach of the mounted volume, launched a new instance, and tried attaching the EBS volume. However, I keep getting an error - mount: Stale NFS file handle whenever I do so. Did I just lose all my files?

    Read the article

< Previous Page | 24 25 26 27 28 29 30 31 32 33 34 35  | Next Page >