Search Results

Search found 56780 results on 2272 pages for 'system design'.

Page 458/2272 | < Previous Page | 454 455 456 457 458 459 460 461 462 463 464 465  | Next Page >

  • Is there anyway to build a raid system without all drives?

    - by xenoterracide
    I'm building a raid1 (ok it will probably be a raid10,f2 but the difference with 2 drives... isn't much) system with 2 1TB drives. However, 1 of the drives I've ordered is bad so I'm RMA-ing it. I'm wondering if I could partition and install to the 1 drive and then rebuild the array when I get the second drive (after I test it of course) My initial investigation doesn't show me a way of creating the array without specifying all devices... and the device the second drive will be is one that has data that I will need to migrate (plus it's not big enough). Is it possible that I could create an array without specifying all devices? or specify false ones and reconfigure to the right ones later? Or some other method I'm not thinking of.

    Read the article

  • Play system sounds from one computer through another computer's speakers?

    - by kevinconner
    I have a Mac and a Windows PC at work. I use both at the same time with Synergy and SynergyKM, which is a great setup for me. The Mac has built-in speakers and the PC has a sound card, but no speakers. I don't want to get speakers for use at work; instead I'd rather just have all sounds from both computers play through the Mac's speakers. Is there some sort of audio service I can run on each computer to accomplish that? I would want: System alert sounds, like you configure in the Control Panel Sounds from videos in a web browser, including from Flash

    Read the article

  • What could be causing Windows to randomly reset the system time to a random time?

    - by Jonathan Dumaine
    My Windows 7 machine infuriates me. It cannot hold a date. At one point it all worked fine, but now it will decide that it needs to change the system time to a random time and date, either in the future or past. There seems to be no correlation or set interval of when it happens. In attempt to remedy this, I have: Correctly set the time in BIOS. Replaced the motherboard battery with a new CR2032 (even checked it with a multimeter). Tried disabling automatic internet synchronizing via "Date and Time" dialog. Stopped, restarted, left disabled the Windows Time service. Yet with all of these actions, the time will continue to change. Any ideas?

    Read the article

  • How to calculate time spent for Linux system administration?

    - by Jaakko
    I would love figuring out how much time I actually have spent in shell performing administrative duties. For billing of course, this information is not sufficient, but it would be helpful to figure out how much time I have actually spent in shell. My question is: How anyone got tips on how to calculate time spent in shell (in hours + minutes), let's say per user account? Could anyone recommend a software that could be used to analyze this time? Optionally I would like to know about ways how to categorise time spent (e.g. 1 hour for Apache, 2 hours for Samba, etc.) Also, your takes on how you monitor your admin time, are also much appreciated! To narrow down the problem, I am specifically interested in ways of tracking time spent in shell for Linux system administration

    Read the article

  • split virtualization design based on environment or server role?

    - by Dan
    I'm setting up the server environment for a new software development group, which will include 4 test environments. These are web applications, so each environment will have an application server and a database server. I'm planning on buying two physical servers (e.g. 6-core CPU each with 12GB or so of RAM), and I'm thinking virtualization is appropriate here. With that in mind, I've thought of a couple ways that I could organize the virtualization strategy: - Separated by server role: Server 1 has all the application servers, each in their own guest VM. Server 2 has all the databases. OR - Separated by environment: Server 1 has a VM for two of the environments, with the VM containing both the app server and the database server. Server 2 would also contain two test environments, with the same style (app server and database in same VM). The advantages I see with all the app servers on one server and all the databases on another server is that I could probably be more efficient with the database server (one instance running multiple databases). But the other option seems easier to manage (archives/restorations would be contained in a single VM). Any recommendations? TIA.

    Read the article

  • Deploying website content via Subversion

    - by Johann
    we have recently set up a new development infrastructure and process for one of our clients. This involves the strict use of subversion as a central source code repository. The svn repositories contains a seperate branch for code on the live system (/branches/live/). The repositories are use for PHP content (mainly Wordpress Blogs), but in future they may hold other asp code as well. Bonus points for a solutions which more or less in the same way with ASP code on Windows Server 2008 R2. We have two servers: one staging system and one live system. The staging system is updated regularly with the code of the trunk. The live system is update manually. Each webroot on the servers are working copy of either the trunk (staging system) or the live branch (live system). The current workflow is: Developing on the dev's box - commit into the trunk - auto-deploy on staging system - testing on the staging system - merging into /branches/live/ - manual deployment on live system. This works for one-way changes very well, however we have some troubles on every wordpress (or plugin) update: The WP update process removes the directories and unpack the archive of the new version. This removes the svn admin area as well, which produces a lot of errors. We could switch to SVN 1.7 with a single, global admin area, but this would only solve on part of the problem. Finally, we have done the update via the WP Gui, restored the svn admin area, added/removed the files and committed the changes to the trunk. After testing, we had to do basically the same thing on the live server (except the commit, we just reverted the changes and merged the new files from the staging system to the live system). I'm currently thinking of the following: The htdocs of each website is a svn export Each website has a svn working copy beside the htdocs directory a script which "replays" the changes in the wc from htdocs after an update in WP (rsync'ing the changed files to the working copy, rsync'ing new files and svn add them and finally svn delete the deleted files). The script would have to exclude some files (like wp-config.php, uploads/temp directories, etc.). Are there better ways to do this? Unfortunaly, a complete CI server is out of scope due to time and budget limitations.

    Read the article

  • NAS is intermittently inaccessible

    - by Natalie
    Model: QNAP TS-410 Turbo NAS Firmware version: 3.2.5 Build 0409T Issue: Each day, users connect to share folders on the NAS system and have read/write permissions for the share folders to which they need access. However, it often asks them for their log-in details and - when provided with right (or wrong) credentials for a user with read/write permissions - it denies them access. I've checked the logs and I keep seeing the following warnings: 2011-11-23 16:26:29 System 127.0.0.1 localhost Re-launch process [rpc.mountd]. 2011-11-23 16:26:16 System 127.0.0.1 localhost Re-launch process [proftpd]. 2011-11-23 16:25:30 System 127.0.0.1 localhost Re-launch process [rpc.mountd]. 2011-11-23 16:25:15 System 127.0.0.1 localhost Re-launch process [proftpd]. 2011-11-23 16:24:33 System 127.0.0.1 localhost Re-launch process [rpc.mountd]. 2011-11-23 16:24:21 System 127.0.0.1 localhost Re-launch process [proftpd]. 2011-11-23 16:23:37 System 127.0.0.1 localhost Re-launch process [rpc.mountd]. 2011-11-23 16:23:25 System 127.0.0.1 localhost Re-launch process [proftpd]. They seem to occur per minute but I am uncertain about whether or not they are relevant to this issue. The "Login failed" warning has also displayed in the system connection logs which tells me when and which user was unable to log in, as shown below: 2011-11-22 16:11:07 Administrator 192.168.0.xx computer-01 SAMBA --- Login Fail 2011-11-22 16:11:07 Administrator 192.168.0.xx computer-01 SAMBA --- Login Fail 2011-11-22 16:11:06 Administrator 192.168.0.xx computer-01 SAMBA --- Login Fail 2011-11-22 13:46:14 administrator 192.168.0.yy --- HTTP Administration Login Fail 2011-11-22 13:46:09 administrator 192.168.0.yy --- HTTP Administration Login Fail 2011-11-21 15:17:22 user 192.168.0.zz computer-02 SAMBA --- Login Fail 2011-11-21 15:17:18 user 192.168.0.zz computer-02 SAMBA --- Login Fail 2011-11-21 15:17:17 user 192.168.0.zz computer-02 SAMBA --- Login Fail I've researched this on Google and the QNAP forums and have not come up with a resolution as yet.

    Read the article

  • Windows 7 ReadyBoost - What File System To Use With Flash Card/Drive?

    - by Boris_yo
    NTFS, exFAT or FAT32? I know that FAT32 has a limit of 4GB transfer per file, but is it faster and better than NTFS or exFAT? Since Windows 7 by itself uses NTFS, it remains logical to format flash card/drive with NTFS file system, however will NTFS or even exFAT be fine for flash card/drive? P.S. In case i decide to use SD flash card, what should i do if it is already plugged in and i decide to use another SD flash card in order to transfer photos? What will happen if i take out suddenly ReadyBoost SD flash card?

    Read the article

  • Trying to find a duplicate version of PHP on my system. Where is it?

    - by macek
    I have having a helluva time trying to track down which php binary my apache is using. locate bin/php returns this list /usr/bin/php /usr/bin/php-cgi /usr/bin/php-config /usr/bin/phpize /usr/local/bin/php /usr/local/bin/php-cgi /usr/local/bin/php-config /usr/local/bin/php-shell.sh /usr/local/bin/phpize Let's see the versions: /usr/bin/php -v shows 5.3.2 /usr/bin/local/php -v shows 5.3.2 What about which? [macek ~]$ which php /usr/bin/php The problem phpinfo(); when executed by apache shows 5.2.11 Where is this phantom 5.2.11 on my system?

    Read the article

  • Unable to load type error in ASP.NET 4 application on Windows Server 2003 / IIS6 -- only happens after first worker process recycle

    - by Daniel Coffman
    I'm running an ASP.NET 4.0 web application on IIS6 (Windows Server 2003 x64). This app is one of many running on this server under the Default Web Site -- but is alone on it's own application pool because the other sites are all running ASP.NET 2.0 still. When I deploy my application, it works just fine until the application pool recycles or kills its worker process (by default 2 hours or 20 minutes with no activity). After this, I get the error: "Unable to load one or more of the requested types. Retrieve the LoaderExceptions property for more information. System.Reflection.ReflectionTypeLoadException: Unable to load one or more of the requested types. Retrieve the LoaderExceptions property for more information." Refreshing the page, recycling the application pool, and iisreset do nothing. But, I can bring the site back online again for a little while by simply redeploying it. The stack trace seems to start at an EntityDataSource -- see below: [ReflectionTypeLoadException: Unable to load one or more of the requested types. Retrieve the LoaderExceptions property for more information.] System.Reflection.RuntimeModule.GetTypes(RuntimeModule module) +0 System.Reflection.Assembly.GetTypes() +144 System.Data.Metadata.Edm.ObjectItemConventionAssemblyLoader.LoadTypesFromAssembly() +45 System.Data.Metadata.Edm.ObjectItemAssemblyLoader.Load() +34 System.Data.Metadata.Edm.AssemblyCache.LoadAssembly(Assembly assembly, Boolean loadReferencedAssemblies, ObjectItemLoadingSessionData loadingData) +130 System.Data.Metadata.Edm.AssemblyCache.LoadAssembly(Assembly assembly, Boolean loadReferencedAssemblies, KnownAssembliesSet knownAssemblies, EdmItemCollection edmItemCollection, Action`1 logLoadMessage, Object& loaderCookie, Dictionary`2& typesInLoading, List`1& errors) +248 System.Data.Metadata.Edm.ObjectItemCollection.LoadAssemblyFromCache(ObjectItemCollection objectItemCollection, Assembly assembly, Boolean loadReferencedAssemblies, EdmItemCollection edmItemCollection, Action`1 logLoadMessage) +580 System.Data.Metadata.Edm.ObjectItemCollection.ExplicitLoadFromAssembly(Assembly assembly, EdmItemCollection edmItemCollection, Action`1 logLoadMessage) +193 System.Data.Metadata.Edm.MetadataWorkspace.ExplicitLoadFromAssembly(Assembly assembly, ObjectItemCollection collection, Action`1 logLoadMessage) +140 System.Web.UI.WebControls.EntityDataSourceView.ConstructContext() +756 System.Web.UI.WebControls.EntityDataSourceView.ExecuteSelect(DataSourceSelectArguments arguments) +147 This is a bug filed for the same (or similar) problem: http://connect.microsoft.com/VisualStudio/feedback/details/541962/unable-to-load-one-or-more-of-the-requested-types-connected-with-entitydatasource Question: Has anyone seen this and have advice? I've tried copy-local on all the references... Works just fine on my dev machine. Works on the server until the application pool worker process recycles. I'm building in release mode, but experience the same result when I build for debug. I'm stumped.

    Read the article

  • launchctl - use rvm instead of system Ruby in executed scripts?

    - by Stefan Kendall
    I have a launchctl job I define as such: <key>ProgramArguments</key> <array> <string>/bin/sh</string> <string>-c</string> <string>~/projects/script.sh</string> </array> When I run script.sh manually, the script works fine, as it uses the currently configured rvm version of ruby. When I run this through launchctl, the system version of Ruby is used, which breaks the script. How can I get this script to run with the right version of ruby available?

    Read the article

  • How do I log file system read/writes by filename in Linux?

    - by Casey
    I'm looking for a simple method that will log file system operations. It should display the name of the file being accessed or modified. I'm familiar with powertop, and it appears this works to an extent, in so much that it show the user files that were written to. Is there any other utilities that support this feature. Some of my findings: powertop: best for write access logging, but more focused on CPU activity iotop: shows real time disk access by process, but not file name lsof: shows the open files per process, but not real time file access iostat: shows the real time I/O performance of disk/arrays but does not indicate file or process

    Read the article

  • how to design network for connectivity between private and corporate LANs?

    - by maruti
    there is a bunch of servers connected to shared storage in a private LAN (10.x.x.x). this privateLAN is managed by a windows server (DHCP, DNS and directory services) these hosts need to be from outside of the datacenter Eg. Remote desktop. can the NIC2 on each of the hosts be connected to the other public LAN (compromising speed or security? what are improtant considerations: additional hardware? like switches? routing&DNS software? currently available hardware : Dell Powerconnect 6224 switch .... planning this for storage network. software: windows 2003 server for DHCP, DNS, A/D ? would it be more flexible to use Linux distributions like IPCOP, Untangle etc? all that I am looking for is good isolation between private and other networks, avoid DHCP, DNS, AD clashes.

    Read the article

  • A design for watching IPTV anywhere in the house?

    - by Carlos
    I'm currently getting TV and internet via IP to my house. The service comes into an ISP-supplied Router (ST585) which is in turn connected to an IPTV box. I need to replace the router, as it's port forwarding seems to be broken, so I was thinking of getting a box with an IGMP proxy. I once mistakenly got a non-IGMP box, with the result that the internet worked, but the IPTV didn't. The trouble is, I have no idea how to set up the IPTV part of the installation. I do have a copy of the ST585 configuration, but it doesn't look anything like the Cisco stuff that I learned about at CCNA. What are the steps necessary to reproduce the IGMP setup? I was thinking of WireSharking the communications on the network, but I suppose I'd need a hub for that? As a bonus, since the packets are coming in with the TV signal, is it possible to mess with the IGMP setup so I can watch TV anywhere in the house?

    Read the article

  • A design for watching IPTV anywhere in the house?

    - by Carlos
    I'm currently getting TV and internet via IP to my house. The service comes into an ISP-supplied Router (ST585) which is in turn connected to an IPTV box. I need to replace the router, as it's port forwarding seems to be broken, so I was thinking of getting a box with an IGMP proxy. I once mistakenly got a non-IGMP box, with the result that the internet worked, but the IPTV didn't. The trouble is, I have no idea how to set up the IPTV part of the installation. I do have a copy of the ST585 configuration, but it doesn't look anything like the Cisco stuff that I learned about at CCNA. What are the steps necessary to reproduce the IGMP setup? I was thinking of WireSharking the communications on the network, but I suppose I'd need a hub for that? As a bonus, since the packets are coming in with the TV signal, is it possible to mess with the IGMP setup so I can watch TV anywhere in the house?

    Read the article

  • Is there an encrypted write-only file system for Linux?

    - by Grumbel
    I am searching for an encrypted filesystem for Linux that can be mounted in a write-only mode, by that I mean you should be able to mount it without supplying a password, yet still be able to write/append files, but neither should you be able to read the files you have written nor read the files already on the filesystem. Access to the files should only be given when the filesystem is mounted via the password. The purpose of this is to write log files or similar data that is only written, but never modified, without having the files themselves be exposed. File permissions don't help here as I want the data to be inaccessible even when the system is fully compromised. Does such a thing exist on Linux? Or if not, what would be the best alternative to create encrypted log files? My current workaround consists of simply piping the data through gpg --encrypt, which works, but is very cumbersome, as you can't easily get access to the filesystem as a whole, you have to pipe each file through gpg --decrypt manually.

    Read the article

  • Create a bootable .wim from Win 7 system image?

    - by Christian Nikkanen
    I was given an assigment to create a clean install of Windows 7, take a system image of it and make it a bootable image, so it can be booted from our local server. I've readed a few tutorials, but only one worked so far that it created a .wim . But when I transferred it to the server (Windows 2008 Server), navigated to Boot Images -- Add a new boot image But it says it's not bootable. So how would I create a .wim that is good for the server from a 64bit Windows? I have Windows AIK installed.

    Read the article

  • System will only boot into a Live CD with core disabled in BIOS and ACPI off.

    - by CookieOfFortune
    I have a system that was running a Q6600 and Windows 7 RC. It crashed last night into a BSOD with a MACHINE_CHECK_EXCEPTION. Now it does that every time it boots into Windows. I tried using my Ubuntu Live CD, but the kernel would error out, the stack trace showed something along the linse of "Can not synchronized to one of the CPUs". Working from this, I enabled the "Limit CPU to 3 cores" option in my BIOS and tried again. This time, it seemed to have died after an ACPI call, so I disabled that during the boot and now it is running from the Ubuntu Live CD, showing 2 cores. Does anyone here think I have any hope or is it simply a CPU waiting to die? EDIT: 1 core now.

    Read the article

  • On Ubuntu, how to design a monitor to run a process when a file is written to a directory?

    - by J G
    I want to run a process when a file is written to a directory on Ubuntu. I understand I can write a monitor to do this. (ie not an event based trigger but a 'monitor'). To implement this - other than writing a for loop of infinite duration in C that polls the directory - what steps would you take to implement this? What steps would you take to implement a monitor on Ubuntu that runs a process (another application) when a file is written to a directory? (The reason that you'd do this is the that process writing the file is under a limited set of permissions (eg tomcat6) but the executed process (an existing software package) needs root (ie to run a process in a chroot jail))

    Read the article

  • How do I properly dual boot Ubuntu and Windows 7 on an UEFI system?

    - by S. Robert James
    I have an HP Z210 with Windows 7, which is UEFI based. I'd like to dual boot it to Ubuntu 11.10 with GRUB2, but am having problems due to the UEFI. The install CD goes to the end, but then the machine always boots right into Windows. These problems are apparently documented (here and here), but there's no consensus as to what the simplest solution is. Any recommendations? I want Ubuntu and Windows to both be bootable. (Perhaps if I knew more about UEFI and its system partition, and how it differs from BIOS and MBR, I'd be in a better position. So background answers explaining how UEFI loads up are very appreciated.)

    Read the article

  • Asterisk doesn't start properly at system startup. DNS lookup fails.

    - by leiflundgren
    When I start my Ubuntu system it attempts two DNS lookups. One to find out what my internet-routers external ip is. And one to find the IP of my PSTN-SIP-provider. Both fails. [Apr 7 22:14:54] WARNING[1675] chan_sip.c: Invalid address for externhost keyword: sip.mydomain.com ... [Apr 7 22:14:54] WARNING[1675] acl.c: Unable to lookup 'sip.myprovider.com' And since the DNS fails it cannot register properly a cannot make outgoing or incoming calls. If I later, after bootup, restart asterisk everything works excelent. Any idea how I should setup things so that either: Delay Asterisk startup so that DNS is up and healthy first. Somehow get Asterisk to re-try the DNS thing later. Regards Leif

    Read the article

  • What are steps to upgrade an cisco UCS B series system vmware vsphere from 4.1 to 5.0

    - by Gk.
    I have a Cisco UCS B-series system with 1.4 firmware running vsphere 4.1 (ESX) + Nexus 1000V. I want upgrade all that stuff to vsphere 5.0 without downtime. I tried to find any documentation describe all steps needed to do it, but cannot found anything clear. Here is my plan: Upgrade firmware of UCS from 1.4 to 2.0. Doc: http://www.cisco.com/en/US/docs/unified_computing/ucs/sw/upgrading/from1.4/to2.0/b_UpgradingCiscoUCSFrom1.4To2.0.html Upgrade all vcenter, hosts+VEM, virtual machines, datastores using VMWare best practice. Is it OK? Am I missing something? Thank you, giobuon.

    Read the article

  • Asterisk doesn't start properly at system startup. DNS lookup fails.

    - by leiflundgren
    When I start my Ubuntu system it attempts two DNS lookups. One to find out what my internet-routers external ip is. And one to find the IP of my PSTN-SIP-provider. Both fails. [Apr 7 22:14:54] WARNING[1675] chan_sip.c: Invalid address for externhost keyword: sip.mydomain.com ... [Apr 7 22:14:54] WARNING[1675] acl.c: Unable to lookup 'sip.myprovider.com' And since the DNS fails it cannot register properly a cannot make outgoing or incoming calls. If I later, after bootup, restart asterisk everything works excelent. Any idea how I should setup things so that either: Delay Asterisk startup so that DNS is up and healthy first. Somehow get Asterisk to re-try the DNS thing later. Regards Leif

    Read the article

  • Why can't I install Ubuntu 10.04 on a system that already has Ubuntu 8.04?

    - by Android Eve
    Ubuntu 10.04 is beautiful. I love it. I am dying to install it on my PC, alongside the existing Ubuntu 8.04 (from which I write this message right now). But... it won't let me! When I reach the partitioning stage (manual!) Ubuntu 10.04 sees my two HDDs as one RAID volume. It doesn't see all the partitions I already have in place in /dev/sda and /dev/sdb. Even Windows 7 doesn't behave like this... (yes, I actually managed to install Windows 7 64-bit in dual-boot configuration with Ubuntu 8.04 on this same system). Note: GParted on Ubuntu 10.04 (live CD) sees the partition intended for Ubuntu 10.04 (/dev/sda4) perfectly, but is unable to format it. Any idea how to solve this problem?

    Read the article

  • Material to use for computer system cover against UV and salty air?

    - by hippietrail
    I live right next to the sea and have a large window quite close to my computer setup which allows a lot of indirect sunlight to enter. I'd like to buy or make a cover for my computer system. From visiting my usual mom & pop computer shop yesterday I got the impression these might not really exist any more. If I make my own I need a material with these qualities: Block or reduce ultraviolet light which can depolymerize plastics (the sun here in Australia is much stronger than in the northern hemisphere). Block salt-laden sea air which can oxidize USB and other connectors. Not cause static electricity when covering or uncovering. Keep dust off of course (-: My setup is a laptop plugged into a wide-screen LCD with a few external drives. So I think I'd want a largish sheet to flop over the whole desk. Are such covers commonly sold these days? What material(s) should I look for which provides the listed attributes?

    Read the article

< Previous Page | 454 455 456 457 458 459 460 461 462 463 464 465  | Next Page >