Search Results

Search found 700 results on 28 pages for 'nas'.

Page 22/28 | < Previous Page | 18 19 20 21 22 23 24 25 26 27 28  | Next Page >

  • Can I run a web site from my home network without jeapordizing other computers on my LAN?

    - by Alchemical
    I have a home LAN with 5 computers and a NAS, all connected to a Linksys router which is connected to my Cox cable modem. I'm interested in having one of my computers run an IIS-based web site and to have it be accessible to the internet with a static IP. However, I do not want o jeapordize the safety of the other computers on my home network! Is there anyway to do this safely, or as safe as possible? I may also like to run an FTP server from this computer. Finally, optionally I would like to allow remote access to this computer from the internet--but it seems to me that that may increase the security risk to the other computers substantially.

    Read the article

  • Copying files to zfs mountpoint doesn't work - the files aren't actually copied to the other filesystem,

    - by user113904
    I have 3 x 4 TB disks in a NAS that I want to group together and access as if they were one whole 'unit' of some kind. I also have a 250GB disk containing the OS - this is full of films and tv shows currently. I thought zfs sounded good so I created a raidz zpool, after installing the ppa sudo zpool create store raidz /dev/sdb /dev/sdc /dev/sdd and set the mountpoint to /mnt/store sudo zfs set mountpoint=/mnt/store /store checked it was successful - I think it was sudo zfs list NAME USED AVAIL REFER MOUNTPOINT store 266K 7.16T 170K /mnt/store Then I wanted to move over a whole load of files from my home directory. I went to where the to-be-copied folder was (called media) and entered sudo cp -R * /mnt/store cp: cannot create directory `/mnt/store/media': No space left on device It seems like it's not copying over to the new filesystem I made (or thought I did). I've never really done this type of thing until a few days ago so may be running before I can walk... is this not the right way to copy files across? I've only used windows before so the idea of mountpoints is a bit mind boggling. I'm using XBMCbuntu 12 beta 2.0 which is based on 12.04. Will retry with normal Ubuntu 12.04 desktop to see if that's the problem. thanks for the help!

    Read the article

  • Drawbacks of installing linux on usb stick?

    - by Znarkus
    I am setting up a router/nas/http/whatever server based on an ION mini-ITX board. I've installed Ubuntu Server on an old 160 GB drive, but it generates a lot more heat and vibrates more than my other new drive (storage). It just doesn't fit the concept, and worse: it takes up a SATA port. As SSD's are crazy expensive I'm thinking of buying an extra 4 GB USB stick, and raid0 it. From my point of view, these are the pros/cons: Pros Low power consumption No vibrations No heat Smaller Get to buy new, larger USB stick (:D) Cons Shorter life time Slower Raid 0 More work maintaing/installing? I think the pros overweighs the cons. Shorter life time and raid 0 is countered by regular backups of the configs/settings. Slower is partially countered by raid 0, and I don't know about the last one. What do You think? Experience? Another solution?

    Read the article

  • Best practice? Using DPM to backup VMs within each VM or through the host?

    - by andrew
    We've got two Hyper-V hosts running multiple VMs (all flavors of Windows Servers). One of the VMs is running MS Data Protection Manager 2010, which runs beautifully (most of the time) and is connected to a separate NAS via iSCSI for the DPM storage. I noticed when I installed the DPM agent on the Hyper-V hosts, it enumerates the VMs in the DPM Protection listing. I don't want to burn through my storage space too fast with duplicate protection, so I was wondering: Is it recommended to back up VMs through the host, or is it better to install the DPM agent on each VM and backup as I would any other machine? It would seem as though most people (currently including me) do it the second way, but is there any advantage to including the entries under HyperV (Backup using Child Partition Snapshop)?

    Read the article

  • FreeBSD: Samba performance over GBit-Ethernet

    - by Axel Gneiting
    I'm using a FreeBSD NAS with RAID-Z. I can read ~300MB/s from the ZFS disks to /dev/null on the box, but only get about 50MB/s over GBit-Ethernet with SMB to Windows 7 (Samba 3.5.6). Both systems have Intel-PCIe-NICs and are connected directly. Samba is configured to use AIO and I already tried to tune TCP/IP: kern.ipc.maxsockbuf=16777216 net.inet.tcp.sendspace=1048576 net.inet.tcp.recvspace=1048576 net.inet.tcp.sendbuf_max=8388608 net.inet.tcp.recvbuf_max=8388608 net.inet.tcp.delayed_ack=0 Any ideas what's causing the bottleneck? I think the link should handle 100 MB/s easily.

    Read the article

  • Online Backup strategy for family individuals

    - by marlon brando
    The majority of my family, including myself, uses Dropbox and Syncplicity free accounts for our syncing needs. However our backup strategy is pretty non-existent, we all have access (via webDav) to our qnap nas located at home, however we copy files accross when we can remember to do so, terrible I know. Is there tools like CrashPlan or Spideroak that allows each the accomodation of each family members syncing and back-up needs under one main account? I'm not sure how this would work, as each family member would need a seperate sub-account or would I manage each computer's syncing and backup lists from a single account? Any thoughts?

    Read the article

  • Clonezilla is not able to clone a RAID 1 disk

    - by Adrian
    I have a HP Server DL320 G5. There are two SATA hard disks configured as RAID 1 through HP embedded RAID controller. Server OS is running GNU/Linux (Fedora) Server booted up with clonezilla live CD. The image will be stored on a NAS connected through NFS. Clonezilla could mount the NFS share and could see the two hard disks /dev/sda and /dev/sdb. I selected /dev/sda for disk cloning. However I could not see the cloning progress and got straight into a prompt for reboot, poweroff, command line I tried to select /dev/sdb but the same issue.

    Read the article

  • Beginners advice on Small business network disk(s)

    - by Rob
    We are having 10 PCs used by various user and presently use one network disk (a LaCie NAS) for all our data. Everything is Windows Vista and our collective IT hardware knowledge is minimal. This worked well generally. However, recently the disk freqently loses connection from the network (2-3 times per week) and the only way back seems to be the "turn it off and back on" trick. This obviously cant be any good for the disk. I understand that there are various more sophisticated ways of storing data and was wondering what people would recommend. One of the worries is obviously disk failure (either in part or as a whole) and the lack of continued availability due to network issues. I would guess that a disk which replicates data wouldnt work as a sole solution due to the network connection, but dont know what hardware (and/or software) would/could work in our case. In terms of size, we are looking at very small amounts, ie. less than 500 GB in total.

    Read the article

  • Using rsync to take backup of folder

    - by Ali
    Hi, I have a server (Linux) with NAS which is mounted as folder "mount" I have website in "public_html" folder. I want to take backup of website in mount folder automatically at certain intervals for e.g. every hour. I read that there is something called "rsync" which is used to make two folders sync. And it doesn't copy all files every time and instead matches if the file has been changed and then only update changed files. How do I use it to make automatic backups? I have root access to server. Thanks

    Read the article

  • Does RAID 1 protect against corruption?

    - by Shaun
    Does Raid 1 protect against data corruption? For example, let's say that I am keeping all of my important files on a NAS that uses 2 disks in a RAID 1. If one hard drive has some kind of internal problem and the data becomes corrupted, does the RAID recognize this automatically and correct it using data from the other good disk? Could it even know which copy is the good one? Does RAID 5 protect against corruption? I know that RAID is not a backup solution. I am trying to figure out how to make sure that I am not backing up corrupt data!

    Read the article

  • Debian Harddrive Fdisk - Same ID's and changing letters

    - by James Willson
    I am trying to create and install a debain NAS and ive been having a hard time because I am new to all of this. I used ntfs-3g in order to automount my 4 NTFS drives. I also have a partitioned harddrive which is for the OS. When I was working on it and I ran this command I got this: fdisk -l /dev/sdae1 fdisk -l | grep NTFS /dev/sda1 /dev/sdb1 /dev/sdc1 /dev/sdd1 The weird thing is, all of the NTFS harddrives listed had an ID of 7. The next time I boot up my machine, I get an error about mounting /dev/sda1 and I run this command, and get the following results: fdisk -l | grep NTFS /dev/sdb1 /dev/sdc1 /dev/sdd1 /dev/sde1 I havent plugged in any drives, so whats going on? How to I make sure that my drives are mounted with the same sdXX name every time, and is the reason for this because they dont have unique ID numbers, if so, how do I solve this?

    Read the article

  • Connect to wired and wireless networks at same time, Ubuntu

    - by Gary Chambers
    Currently, I have a media PC running Ubuntu 10.04 that I am trying to connect via a wired network cable directly to a NAS box, and wirelessly to the router. This works no problem after I run sudo /etc/init.d/networking restart but I can't get both interfaces to come up on system startup. My /etc/network/interfaces file reads as follows: auto eth0 iface eth0 inet static address 10.0.1.2 netmask 255.255.254.0 broadcast 10.0.1.255 network 10.0.1.0 auto wlan2 iface wlan2 inet dhcp As I say, I know this works, because I can get it to work by restarting the network interfaces, but I can't bring them both up on system startup. Does anyone know why this might be?

    Read the article

  • SAN alternative for VMWare

    - by CogitoErgoSum
    Has anyone utilized something aside from a SAN to run their VMWare images off? We are looking to drop in two HP Servers and VMWare on them and run them off a SAN. Due to the cost of SAN though our CFO and VP are wondering if there are any viable alternatives (I.e. NAS) that can effectively run a VMWare. I can't think of any off of the top of my head. IF anyone can provide one or a good article outlining why to stick to SAN that'd be great.

    Read the article

  • Linksys EA3500 : use WAN port as WAN + LAN possible?

    - by Dough
    I'm waiting for deliver of a Linksys EA3500 router. I was wondering if it was possible to get the WAN port act as normal WAN port (Internet IP) + LAN switch ? My problem is that my Internet box is in a room with NAS and PC, and the router will be in another room with TV, PS3 and PC. The Internet box have a 4 port Gigabit switch so one cable will plug from there to the router's WAN in the other room. The router will get the Internet IP from the Internet box, but will it be possible to have private network through this link too ? I could have a first network between the Internet box and the router, and another network form router and devices but that will do double NAT for nothing... Thanks for reading me and your help !

    Read the article

  • Can I mark a folder as mountpoint-only?

    - by Collin
    I have a folder ~/nas which I usually use sshfs to mount a network drive on. Today, I didn't realize the share hadn't been mounted yet, and copied some data into it. It took me a bit to realize that I'd just copied data into my own local drive rather than the network share. Is there some way to mark in the system that this folder is supposed to be a mount point, and to not let anyone copy data into it? I tried the permissions solution here: How to only allow a program to write to a directory if it is mounted?, but if I don't have write access I also can't mount anything to it.

    Read the article

  • How to maximise performance in computers connected into LAN via Gigabit ethernet router?

    - by penyuan
    Our group is setting up a server (which might just be a NAS, but we're not sure yet), which shares files, so that it connects to all other computers in the room (about 10 of them). I am thinking just hooking all of them up via a gigabit router/switch. Is there anything I should watch out for, in terms of cables, connections, or the connection capabilities of each computer in the network? For instance, I don't want a slow computer in the LAN to slow down everyone else's connection, etc., etc. Thanks for the education.

    Read the article

  • Should I cache the data or hit the database?

    - by JD01
    I have not worked with any caching mechanisms and was wondering what my options are in the .net world for the following scenario. We basically have a a REST Service where the user passes an ID of a Category (think folder) and this category may have lots of sub categories and each of the sub categories could have 1000 of media containers (think file reference objects) which contain information about a file that may be on a NAS or SAN server (files are videos in this case). The relationship between these categories is stored in a database together with some permission rules and meta data about the sub categories. So from a UI perspective we have a lazy loaded tree control which is driven by the user by clicking on each sub folder (think of Windows explorer). Once they come to a URL of the video file, they then can watch the video. The number of users could grow into the 1000s and the sub categories and videos could be in the 10000s as the system grows. The question is should we carry on the way it is currently working where each request hits the database or should we think about caching the data? We are on using IIS 6/7 and Asp.net.

    Read the article

  • Connect to wired and wireless networks at same time, Ubuntu

    - by Gary Chambers
    Currently, I have a media PC running Ubuntu 10.04 that I am trying to connect via a wired network cable directly to a NAS box, and wirelessly to the router. This works no problem after I run sudo /etc/init.d/networking restart but I can't get both interfaces to come up on system startup. My /etc/network/interfaces file reads as follows: auto eth0 iface eth0 inet static address 10.0.1.2 netmask 255.255.254.0 broadcast 10.0.1.255 network 10.0.1.0 auto wlan2 iface wlan2 inet dhcp As I say, I know this works, because I can get it to work by restarting the network interfaces, but I can't bring them both up on system startup. Does anyone know why this might be?

    Read the article

  • What is the safest and least expensive way to store 10 terabytes of data?

    - by Josh T
    I'm a member of a production company and we're preparing for our first feature film. We've been discussing methods of data storage to keep all of our original content safe (for as long as possible). While we understand data is never 100% safe, we'd like to find the safest solution for us. We've considered: 16TB NAS for on-site storage 4-5 2TB hard drives (cheap, but not redundant), copy original footage to drives then seal in static free bag Burn data to Blu-Ray disks (time consuming and expensive: 200 disks == $5000) Tape drive(s)? I know the least about tape drives, except the fact that they're more reliable than disks. Any experience/knowledge with this amount of data is hugely appreciated.

    Read the article

  • redundant/multi-site terminal server

    - by Adam
    Hi We have a Hyper-V cluster running 5 virtual terminal servers using HA. We need to be able make this system redundant and so if this site was to fail our users could log into the backup system at another location and access their data via the terminal servers. Any ideas? We were thinking of maybe using a NAS which replicated the data to the other location in real-time(pass-through disks)? and having a similar Hyper-V cluster setup in the backup location. However we would need to create the users in both location and create a virtual mirror without the data ie applications, directories, settings etc. Is this the best way to achieve this? We have read that using Hyper-v pass through disks is a big performance de-grade.

    Read the article

  • CIFS shares do not mount after upgrade to 12.10 from 12.04

    - by Mothball
    I have seen issues close to my problem but no one seems to have a definitive answer as to what is going on and why the failure occurs. I have a number of NAS devices on my home network and on a previous install of 12.04 and version prior mounting at login worked using this entry for each in fstab: //servername/sharename /media/windowsshare cifs guest,uid=1000,iocharset=utf8,codepage=cp850,cp850 0 0 Now when I use this, 12.10 reports the standard - cannot mount bad option ... blah blah... The kern log reports that the CIFS option "codepage" unknown... changed entry to "unicode" and received the same error message. There are no other error messages or log entries that would indicate another issue, but this is the statement I used for quite awhile with version 12.04 and before. Is the codepage option obsolete in 12.10/CIFS now? Is there a codepage support program that I must load? Is there some kind of helper program that is required to supports the codepage option? A current review of the man pages at samba.org does not make mention of the option "codepage". Extremely confused - any help/insight would be greatly appreciated.

    Read the article

  • HP DL370 G6 expansion

    - by user72185
    Hi, we are running a HP DL370 server with 6 x 300 gb disks on RAID 5. Due to a Windows update causing our server to fail recently, we couldn't access the data. I now want to separate the data from the OS (Windows server 2008 r2) so that if anything like that happens again, we can route everyone through a separate server. I have seen these HP storageworks enclosures (msa70) and have a couple of questions: Can I just take out our 2.5 inch 10k SAS drives, install them in the new Storageworks NAS and hey presto we would be up and running? If I wanted to then add another drive (I think there are 25 bays), can I just insert a blank but identical drive and the RAID 5 would dynamically expand to incorporate the new drive. Many thanks Adrian

    Read the article

  • How can I improve performance over SMB/CIFS for an application that has poor write speeds?

    - by Jeremy
    I have a third party application that reads several large files and generates a third large file. Its performance is quite good when the generated file is stored on "local storage", i.e. either a direct attached or iSCSI-based disk. The source files that are read can be stored remotely on our NAS and accessed via SMB with little effect on performance. However, if we attempt to write the target file to any kind of SMB/CIFS share (Samba or Windows Server) the performance drops almost ten-fold. This is unacceptably slow in our case. Writing files to network shares is not otherwise slow. I can copy large files to SMB shares and get great performance - near what I would expect is possible given the disks and network in question. I have a theory that this application's problem with SMB shares has something to do with a lack of write caching over the share and perhaps lots of network roundtrips. Is this possible and is there anything that can be done about it?

    Read the article

  • Oracle Database 12c Technical Trainin

    - by mseika
    Audience Database Administrators, Solutions Architects, System Engineers, Technical Consultants, Implementation and support personnel, Technical Analysts, and Developers. What We Are Announcing During his opening keynote at Oracle OpenWorld 2012, Larry Ellison previewed Oracle Database 12c - the latest generation of the database market leader and Oracle flagship product. Oracle Database 12c introduces many groundbreaking features making it the database foundation of choice for the cloud. Many years of development effort have been focused on introducing innumerable new technological innovations centered on the cloud computing platform. This training session will focus on the specific needs of our Oracle partner community and developers, and provide insight into the many features and capabilities your customers will be looking to leverage in their own environments. Topics includes: Consolidation and Cloud Strategies Deep dive into the key Database 12c Options Migrating to Oracle Database 12c Webcast Details Speaker: Sean Stacey, Director of Platform Technology Solutions.Please note that you will need to join both the Audio and Web Conferences to attend. Please plan on joining 10 minutes before the scheduled time. Region: Date & Time Audio Conference Web Conference Calendar NAS, LAD, EMEA July 28am PT (US) Duration: 1 hour US/Canada: (866) 900-7470Click for local numberIf your country is not listed, dial +1 (706) 634-7953. Local charges may applyCONFERENCE ID: 98498078 Click here to joinPassword: Oracle123 Add this session to your calendar If you have any questions, please contact: Yvonne OungSenior Manager, Channel [email protected]

    Read the article

  • Significant OS X Finder lag when listing directories/files

    - by Jack Sleight
    I'm experiencing some significant OS X finder lag, that seems to be purely an issue with Finder itself, and not the HD or any other part of OS X (I'll explain below). The lag only appears to be when listing directories/files, where I'm seeing up to twelve or so seconds of lag (the folder opens with a blank list and the spinner going in the bottom right). This happens with both the local SSD and network drives (connected via ethernet or wifi) Browsing both local and network drives in terminal and listing directories is instant I can actually browse files on my NAS from my phone over a 3G connection from the other side of the country faster than Finder can when connected to the local network (madness!) Can anyone help? Thanks.

    Read the article

< Previous Page | 18 19 20 21 22 23 24 25 26 27 28  | Next Page >