Search Results

Search found 20475 results on 819 pages for 'multiple repositories'.

Page 406/819 | < Previous Page | 402 403 404 405 406 407 408 409 410 411 412 413  | Next Page >

  • Virtualize Mac OS X on a Mac

    - by Joe Mako
    I use Virtual Box on a Mac OS X 10.6.2 MacBook Pro. Can I create/run a Mac OS X 10.5 guest OS on this setup? If not, what setup would allow me to run multiple versions of Mac OS X on the same machine?

    Read the article

  • VirtualBox 4.1.12 Ubuntu Host Bridged Networking

    - by hidden_premise
    Setup Ubuntu Server 11.10 x64 VirtualBox 4.1.12 One network adapter eth0 with a static IP address My /etc/network/interfaces looks like #Local loopback auto lo iface lo inet loopback #Primary Interface auto eth0 iface eth0 inet static address 10.0.0.6 netmask 255.255.255.0 gateway 10.0.0.1 Need Static IP for Ubuntu Host Static IP for multiple Ubuntu guests My attempt Downloaded the .deb for 64bit virtual box dpkg -i virtualbox.deb Create a VM and install ubuntu server 11.10 x64 from ISO Set the networking mode to bridged and choose 'eth0' as the interface Fail. None of the guides I can find for bridged networking work because everything changed since vbox 2

    Read the article

  • Windows web server

    - by klork
    I need a windows 2003 hosted server for various websites that I develop and host. Any suggestions on a good deal out there? I need full root access to IIS 6/7 and to be able to host multiple websites using host headers. Thanks

    Read the article

  • Why create many partitions?

    - by ryeguy
    I have noticed that when installing Ubuntu some people create multiple partitions for directories. Like one for root, one for home, one for boot. What is the advantage to doing this over installing them all on one partition, assuming there is only one hard drive?

    Read the article

  • Good Wireless Range Extender

    - by Joseph Sturtevant
    Can anyone recommend a good wireless (802.11g) range extender? I would like to find something that will provide reliable wireless to areas of the building that aren't covered or get poor reception. I would also like a product that won't require big changes to my current wireless setup (multiple APs with a wireless controller are out). Latency and bandwidth aren't terribly important. Does anyone have experience with a product like this?

    Read the article

  • splitting off an emacs window into a new frame

    - by user38983
    From a frame displaying multiple windows, I'd like to be able to 'break' one of the windows off into another frame, perhaps with a modifier-click on the modeline, where the window would now have its own frame but hold the same buffer and position. I was wondering if this behavior was available somewhere in the standard library, if there was an extension for it, or perhaps if a small elisp function would work.

    Read the article

  • DPM 2010 iSCSI Mirror

    - by Thermionix
    We're using DPM 2010 for exchange backups, The backup Disk(s) are iSCSI attached drives from multiple NAS boxes. We'd like to mirror iqn.2009-07.com.example.example:RAID.iscsi4.vg0.iscsi05 onto iqn.2012-3.com.example.example:RAID.iscsi4.vg0.iscsi05 DPM 2010 requires the disk for itself and handles volume creation, Therefore we can't just create a mirrored volume in Disk Management. DPM itself doesn't seem to have any ability to mirror the Disks in its storage pool. Any tips on how to mirror the volumes from one drive to the other?

    Read the article

  • Instant Messenger proxy with web-based, searchable chat logs and mobile support

    - by zpinter
    Does anybody here know of a decent setup for having multiple computers and devices (iPhone/Android) logged into the same IM accounts (yahoo, gtalk, AIM) with consolidated web-based chat logs? I've tried/thought of a few approaches: IRC w/ IRSSI and bitlbee (was nice, but not a great solution for phones and chat logs were painful) Google Talk (would be great if I could just use this, but I need to support Yahoo - perhaps a Jabber relay?) Meebo (can this be used as a proxy?)

    Read the article

  • Web server hits from Akamai IP addresses

    - by Joel
    I am getting a lot of hits to my web server from from multiple IP addresses that are owned by Akamai. This stood out to me as the requests are for resources that are no longer available. These resources should not be being requested - that pages that refer to them are marked as being non-cachable. Any clues?

    Read the article

  • What email server should I choose?

    - by DCC
    I need a secure email server installed in debian-lenny with users in a mysql table. Also users are from multiple domains. Quota should be in mysql or a global variable for all users. What are my options ? THanks in advance for your help.

    Read the article

  • Tomcat with virtual hosts - 404

    - by Thardas
    I have a CentOS 5.2 server set up with Apache 2.2.3 and Tomcat 5.5.27. The server hosts multiple virtual hosts connected to multiple Tomcats. For instance we have one tomcat for development and testing and one tomcat for production. project.demo.us.com points to dev tomcat and project.us.com points to production tomcat. Here's the virtual host's configuration: <VirtualHost *:80> ServerName project.demo.us.com CustomLog logs/project.demo.us.com/access_log combined env=!VLOG ErrorLog logs/project.demo.us.com/error_log DocumentRoot /var/www/vhosts/project.demo.us.com <Directory /var/www/vhosts/project.demo.us.com> Allow from all AllowOverride All Options -Indexes FollowSymLinks </Directory> ########## ########## ########## JkMount /project/* online </VirtualHost> JkMount line defines that we use online worker and our workers.properties contains this: worker.list=..., online, ... worker.online.port=7703 worker.online.host=localhost worker.online.type=ajp13 worker.online.lbfactor=1 And tomcat's conf/server.xml contains: <Connector port="7703" enableLookups="false" redirectPort="8443" protocol="AJP/1.3" URIEncoding="UTF-8" maxThreads="80" minSpareThreads="10" maxSpareThreads="15"/> I'm not sure what redirectPort is but I tried to telnet to that port and there's no one answering, so it shouldn't matter? Tomcat's webapps directory contains project.war and the server automatically deployed it under project directory which contains index.jsp and hello.html. The latter is for static debugging purposes. Now when I try to access http://project.demo.us.com/project/index.jsp, I get Tomcat's HTTP Status 404 - The requested resource () is not available. The same thing happens to hello.html so it's not working with static content either. Apache's access_log contains: 88.112.152.31 - - [10/Aug/2009:12:15:14 +0300] "GET /demo/index.jsp HTTP/1.1" 404 952 "-" "Mozilla/5.0 (Macintosh; U; Intel Mac OS X 10.5; en-US; rv:1.9.1.2) Gecko/20090729 Firefox/3.5.2" I couldn't find any mention of the request in Tomcat's logs. If I shutdown this specific tomcat, I no longer get Tomcat's 404 but Apache's 503 Service Temporarily Unavailable, so I should be configuring the correct Tomcat. Is there something obvious that I'm missing? Is there any place where I could find out what path the Tomcat is using to look for requested files?

    Read the article

  • google docs document in Hebrew and English?

    - by JoelFan
    Is there any way to have Hebrew and English in the same document in Google Docs? So far, everything I have found about multiple languages seems to assume that you want the user interface (menus, etc) and document text in the same language, and that you only want that one language. I would like the interface to stay in English but let me type Hebrew and English into the document text.

    Read the article

  • Windows 7 file-based backup service

    - by Ben Voigt
    I'm looking for a good replacement for Lazy Mirror, since it doesn't support Windows 7 well. Pros: One of the things I really loved about Lazy Mirror is that it always maintains a "full" backup, but does so by only copying modified files. As each file was copied, the old version got archived (moved to an out-of-the-way location). So after mirroring ran, there'd be a complete copy of the file system, which could even be booted if necessary. At the same time, extra space on the backup media was used to store as many older versions of files as possible, without wasting space storing multiple copies of the same version. It seems that with Windows 7 backup, there'd be wasted space storing the same data in both the system image and file backup. It was completely file-based, but also aware of the registry (it had a feature to dump the live registry to hive files in the correct format). The backups were normal NTFS filesystems, no special tool was needed to read them. It automatically cleaned out the oldest previous versions when space ran out (unlike Windows 7 backup which apparently simply starts failing the the backup media fills.) It copied all file attributes including security. Cons: It doesn't deal well with junction points, symbolic links, and hard links. It didn't run as a service without lots of help from firesrv or srvany, and then you couldn't interact with the GUI. Running as a service was necessary to be able to mirror protected OS files. It didn't have open file handling, except for registry hives. I guess that the file-by-file archive and replacement could leave mismatched sets of files, if the mirror was interrupted. This would be the advantage of incremental backup techniques that require old full backup + all intermediate incremental backups to restore. But I don't see this as presenting much of a problem, you'd really only have a boot failure if you had a mixture of pre- and post-service pack files, and I can run a full image backup using another tool before applying a service pack. Does anyone know of a tool that does both full-system backup and storage of old versions of files like Lazy Mirror did (without storing the same data multiple times), and also can run as a service in Windows 7? Free is best of course, but a reasonably priced paid program (e.g. It would be absolutely awesome if it also triggered a backup/mirror pass when a particular external drive was plugged in and generated popup warnings if backups hadn't been run recently)

    Read the article

  • Looking for an application to act as a file bucket / consolidate - with auto tag and duplicate elimi

    - by Notitze
    Is there a windows application that can help with consolidating file libraries (documents, projects, pictures, music) into an archive? I'm looking for: eliminate redundant files by replacing duplicates with links add tags from keywords and date (year) allow an easy dump from multiple external media files (USB sticks, USB drives) Nice to have would be: auto indexing for easy search anything else that makes it easy to find and retrieve a file

    Read the article

  • Solution to route/proxy SNMP Traps (or Netflow, generic UDP, etc) for network monitoring?

    - by Christopher Cashell
    I'm implementing a network monitoring solution for a very large network (approximately 5000 network devices). We'd like to have all devices on our network send SNMP traps to a single box (technically this will probably be an HA pair of boxes) and then have that box pass the SNMP traps on to the real processing boxes. This will allow us to have multiple back-end boxes handling traps, and to distribute load among those back end boxes. One key feature that we need is the ability to forward the traps to a specific box depending on the source address of the trap. Any suggestions for the best way to handle this? Among the things we've considered are: Using snmptrapd to accept the traps, and have it pass them off to a custom written perl handler script to rewrite the trap and send it to the proper processing box Using some sort of load balancing software running on a Linux box to handle this (having some difficulty finding many load balancing programs that will handle UDP) Using a Load Balancing Appliance (F5, etc) Using IPTables on a Linux box to route the SNMP traps with NATing We've currently implemented and are testing the last solution, with a Linux box with IPTables configured to receive the traps, and then depending on the source address of the trap, rewrite it with a destination nat (DNAT) so the packet gets sent to the proper server. For example: # Range: 10.0.0.0/19 Site: abc01 Destination: foo01 iptables -t nat -A PREROUTING -p udp --dport 162 -s 10.0.0.0/19 -j DNAT --to-destination 10.1.2.3 # Range: 10.0.33.0/21 Site: abc01 Destination: foo01 iptables -t nat -A PREROUTING -p udp --dport 162 -s 10.0.33.0/21 -j DNAT --to-destination 10.1.2.3 # Range: 10.1.0.0/16 Site: xyz01 Destination: bar01 iptables -t nat -A PREROUTING -p udp --dport 162 -s 10.1.0.0/16 -j DNAT --to-destination 10.3.2.1 This should work with excellent efficiency for basic trap routing, but it leaves us completely limited to what we can mach and filter on with IPTables, so we're concerned about flexibility for the future. Another feature that we'd really like, but isn't quite a "must have" is the ability to duplicate or mirror the UDP packets. Being able to take one incoming trap and route it to multiple destinations would be very useful. Has anyone tried any of the possible solutions above for SNMP traps (or Netflow, general UDP, etc) load balancing? Or can anyone think of any other alternatives to solve this?

    Read the article

  • Extracting one file from archive: 7-zip requires decompressing entire archive?

    - by siikamiika
    I've noticed that when browsing an archive containing multiple files with 7-zip 9.20 Windows GUI, extracting one file for previewing takes significantly longer with .7z than .rar archives. With .7zips it also cycles through the filenames in the archive. To me it looks like decompressing the entire archive and keeping just one file. Is there a setting in 7-zip (current or beta/alpha versions) that allows RAR-like behavior?

    Read the article

  • Securing against DDoS

    - by gAMBOOKa
    My client would like to have his eCommerce (custom-made) site secured from DDoS attacks. What strategies can I implement? There are multiple forms in the purchasing flow --- searching, drilling-down to the product, user information and payment and I want to avoid captchas.

    Read the article

  • Windows DFS File System Clustering

    - by tearman
    We're attempted to set up a high availability network for our file servers, and we're wanting to do a DFS file system cluster using the same back-end storage (our back-end storage has its own clustering mechanisms that it manages itself). The question being, A. how would one go about setting up DFS clustering, and B. how can we get Windows to cooperate with multiple servers accessing the same SAN volumes?

    Read the article

  • Backing Up vs. Redundancy

    - by TK Kocheran
    I'm currently in stage 2 of 3 of building my home workstation. What this means is that my RAID-0 array of solid state disks will be backed up nightly to a RAID-5 or RAID-6 array of traditional spinning hard disks. However, it recently dawned on me that redundancy is not backup. The main reason for setting up a RAID array with redundancy was to protect myself in the event of a drive failure to serve as an effective backup solution. Wait. What if a bolt of lightning finds a way to travel into my house, through my surge-protector, into my power supply and physically destroys all of my hard disks and SSDs? Well, in that case, I guess I'd be fine because I generally keep most important files (music, pictures, videos) stored in multiple places like on my laptop, my wife's laptop, and an encrypted USB hard drive. Wait. What if a giant hedgehog meteor attacks my house from space traveling at mach 3 and all machines and hard disks are blown to smithereens. Well, I guess I could find a way to do ridiculously slow and cumbersome rsyncs or backups to Amazon's Glacier. Wait. What if there's a nuclear apocalypse... and at this point I start laughing hysterically. At what point does backing up become irrelevant? I completely understand situation one (mechanical drive failure), situation two (workstation compromised or destroyed somehow), possibly even situation three (all machines and disks destroyed), but situation four? There's no questioning the need for backups. None. However, there are three questions I'd really like addressed: To what level should one backup? I definitely understand the merits of physical disk redundancy. I also believe in keeping important files on multiple machines and thinning out the possibility of losing all of my files. Online backups make sense, but they beg the following question. What should I be backing up remotely and how often? It's no problem storage-wise to back up important files (music, pictures, videos) and even configuration and temporal data for all of the machines in my network (all Linux based)... albeit locally. Transferring to the cloud is another story. Worst-case scenario, if I lost all of my configuration for my individual computers, the reality is that I probably lost the machines too. The cloud is a long way away from here; I can run backups over CAT-6 here and see 100MB/s easily, but I'm afraid that I'm only going to see 2MB/s at best when transferring up to the cloud.

    Read the article

  • Debian tuning for increasing read/write buffer.

    - by Claudiu
    Is there a way to modify Debian settings so the memory could be used more for disk read/write caching ? I am already using RAID 0 but thats not enough for multiple users, and the disk is almost struggled. Torrents use the disk very much and rTorrent doesn't have cache settings.

    Read the article

  • Mac OS X VirtualBox Update to 10.6.6

    - by DickSavagewood
    Hi. I was able to get Mac OS X 10.6.2 to work on my 64-bit Acer on VirtualBox multiple times, but I would like to update to 10.6.6. Every time I do, however, I continue to get this annoying error, saying that the kernal and OS version mismatch, and the OS Version is no longer set. What is the proper way to update to Mac OS 10.6.6 because I have heard of it being done before. Thanks in advance for your help, Thommaye

    Read the article

  • What are the best browser based administration tools for Linux?

    - by Kevin Kuphal
    Looking for recommendations the best comprehensive browser based administration tool for administering a Linux machine. Ideally this would be distribution agnostic. This could also be an "appliance" type product that would administer multiple machines on network. This is marked community wiki. Give your answer preferably one product per answer and vote up packages that you recommend also.

    Read the article

< Previous Page | 402 403 404 405 406 407 408 409 410 411 412 413  | Next Page >