Search Results

Search found 12497 results on 500 pages for 'linked servers'.

Page 335/500 | < Previous Page | 331 332 333 334 335 336 337 338 339 340 341 342  | Next Page >

  • Operating Systems supported by the Intel SR1435VP2 Server Platform?

    - by Xspence
    I recently had two Intel SR1435VP2 Servers (with SE7320VP2 server boards) donated to me from a colleague. Google hasn't yielded much more than user manuals when searching for OS-compatibility answers. I have worked with flavors of Linux such as Ubuntu and Debian, but Intel only documents that they have tested proprietary operating systems such as SuSE, Solaris and Red Hat as documented on their driver downloads page. Has anyone worked with these machines before, and if so, do you know if the SR1435VP2/E7320 chipset supports OS's such as CentOS, Debian or Ubuntu? If you need more information or clarification, let me know. This is all new for me. Thanks in advance.

    Read the article

  • Setting up a DNS server for a 3rd level domain

    - by user45339
    If I would like a client to use HIS DNS for a 3rd level domain name (ie test.domain.com), how would I be able to do that? So I have domain servers ns1.domain.com + ns2.domain.com for domain.com but now I want ns1.rabbit.com and ns2.rabbit.com for test.domain.com. How can this be done? I know it's possible because I saw it at some providers. Then, as a second part (and related) part; how to set up a WHOIS for that test.domain.com? So that, if you query my server for the information about test.domain.com, it'll be different than the info on domain.com? Thanks

    Read the article

  • Load-balanced Linux server across internet?

    - by LinuxGnut
    I'm investigating setting up a load balanced server solution consisting of three CentOS 5.4 boxes. Two of these boxes will reside in one facility, while a third will reside in a different facility. I'm currently working to set up heartbeat, ldirectord, ipvsadm to load-balance the machines, but I'm not sure its going to work with I'm not overly familiar with the details behind how all of these work, but is the load balancing going to work correctly when these servers are not all on the same LAN? I'm not sure if heartbeat is using SNMP to send signals or not, which would only work over a LAN. Has anyone tried this or found a different solution?

    Read the article

  • WSUS Updates - Best Practice

    - by What'sTheStoryWishBone
    We have an isolated enviornment of a few hundred servers in which we use WSUS to push updates too. We have thousands of updates which to manage and push to devices testing along the way to ensure the update will not break anything. What are the best practices that you all follow in your enteprise networks to ensure an update does not go out to all the machines that will break something? We currently have ours broken into customized groups for each type of machine. There is one "Test Group" which has one PC of each type which we apply updates to for error checking. Is this a similar procedure others follow or is their an easier safer way to manage the thousands of WSUS updates?

    Read the article

  • Server 2003 Filter mobile devices via MAC

    - by msindle
    At one of my client's sites I need to keep my users unauthorized devices off the wireless. They all know the SSID and Password because many of them have laptops that need the wireless. I'm running out of IP address's and we have sent out numerous emails asking them to stay off, but like most users they ignore IT's email. I'm currently running Server 2003 as the GC/DC (but have 2008 servers in place) and 2 Netgear WNAP320. I've seen several posts similar to what I'm looking for but they seem to deal with Linux. My question is how do I go about doing this without migrating (scheduled for the end of the year) to a new server and is it possible to do this within Server 2003? Thanks msindle

    Read the article

  • PHP and load balancing

    - by StCee
    I have one major domain but the server spec behind it is not good enough. Hence I want to relay the traffic, in particular php-mysql queries to multiple smaller servers. How is that normally be done? (BTW I wonder how much traffic or number of php/mysql request a normal setup on ec2 micro instance can handle? ) I did have a look of EC2 load balancer. But is it only possible to load balance on machines of your own account?

    Read the article

  • GVim asks for passphrase on every action when using scp

    - by Ashnur
    I want to use my vim config when editing files, but there are at least 5 different servers right now where I have to edit them. Of course I could use console (where I set up ssh-keys and and have a script so it wont asks for passhphrase), but then I have to maintain the vim config on every machine. so I decided to use gvim and browse/edit the remote machines via scp://, but on every action a popup appears asking for the passphrase. this is a ubuntu 10.10 install, with xfce installed later on. i checked in the xfce settings so gnome services should start, but it still won't remember the passphrase.

    Read the article

  • How to find dom0 name from hosted domU

    - by svigan
    I'm actually testing RHEL 5.3 with Xen between two servers in order to have a disaster recovery solution. So I'm playing with moving my domU from one dom0 server to the other server. Unfortunatly when somebody else move the domU I don't have any clue where my domain is hosted. I'm wondering where I can't find my dom0 name with inside domU. I'm looking for something like the gzonename command on a solaris sparse zone. I check inside /proc/xen but I don't see anything special except the dom0 kernel release. Does anybody know a wait to find this ?

    Read the article

  • Exchange 2010 allows outside access to network files

    - by user2891127
    One of our users discovered by accident he could access our network files from his smartphone while at home. No VPN needed. He was sent an email with an internal link to a network share on his android. When he opened the email and clicked on the link, he could browse our files while at home. Looking at the access logs, the connection to the share and files he accessed came from our mail server (Exchange 2010). We have no sharepoint servers running at all, and certainly not on the Exchange server. What is this function/feature called, and is it possible to turn this function/feature off? Should I turn this off?

    Read the article

  • Do all domains on the same shared hosting server have the same IP or ID

    - by silow
    Here's what I've got: siteA.com and siteB.com are hosted on hostgator. They're hosted on the same account of a shared hosting server (not VPS or dedicated). script.php is an external site that each of these 2 sites are accessing. I noticed that when siteA.com or siteB.com access the outside script.php, the script identifies them both as 1a.12.12ab.static.theplanet.com (apparently because hostgator uses theplanet.com servers). The fact that they're identified as the same value isn't surprising because after all they're hosted on the same account /home/user123/public_html. What I'm wondering about is how about other websites that are hosted on the same shared hosting server, but under other accounts. Basically other websites that are under another developer's control, but just happen to share the same hardware (hosting server). Do they also have the exact same identifier 1a.12.12ab.static.theplanet.com or that changes by account?

    Read the article

  • Share laptop wifi with router

    - by obie
    I hi guys i'm trying to accomplish something: i have a modem (ADSL) which servers as a wireless router as well. There's a laptop connected to it through cable. On the other room i have a laptop which is connected through the router wirelessly. Now the problem is, since i have other appliances that need to get connected but are not wifi cable i have another wifi router. What i want to know is how can i share the second laptop's wifi with the second router so that the router can then serve as an access point to give internet to my other applinaces through ethernet cables. so in breaf i want: INTERNETMODEM (WIFI)LAPTOP 1 (WIFI)LAPTOP 2 (WIFI)Router 2 (ETHERNET)DREAMBOX & WDTV

    Read the article

  • An international mobile app - Should I set up EC2 instances in multiple regions?

    - by ashiina
    I am currently trying to launch an mobile app for users around the world. It is not a spectacular launch which will get millions of users in weeks - just another individual developer releasing an app. I know enough about the techniques of managing timezones, internationalizing string, and what not ( the application layer ). But I cannot find any information on how I should manage my EC2 instances... Should I be setting up EC2 instances in different regions around the world? Is that a must-do, or is it an overkill? I'm aware that it's the ideal solution in terms of performance, but it becomes very tough managing servers in multiple regions. DB issues, AMI management, etc... I'd much rather NOT do so. So I would like to know the general best practice when launching an international app/website. Note: For static contents, I know it's better to use a CDN, so I'm planning on doing so.

    Read the article

  • Ubuntu+Mono+Postgres+ASP.NET 4.0. No problem?

    - by wreck_of_u
    Would this be ok? I'm an ASP.NET developer and I'm planning to build "portable" web app servers based on Atom D510 mini-ITX. I have ran Ubuntu 10 with MySQL along with a separate IIS machines (win 2k3, 2k8) before with no problems. But now I'm thinking of "packaging" a web/db server into one small, cheap machine. I thought of Ubuntu/Mono/Postgres/ASP.NET, that it would be a good idea but I'm not sure? I have not actually tried it yet. Your thoughts?

    Read the article

  • Vim: How can I :cd to an sftp directory?

    - by ChrisInCambo
    Hi, I'm currently an emacs user, but thought I would come back and take another look at Vim for fun. My client machine is Ubuntu, and I frequently work on other ubuntu servers within my network, emacs and tramp is really nice in this regard, I can just connect to a file via sftp and then easily navigate to it's parents/children/sibling, just as if I was dealing with a file on the local file system. With Vim I can sftp to a file, but none of the navigation stuff works, when I use :e or fuzzy file search, vim still thinks I'm back on the client machine. If I try to :cd to the sftp directory I get an error. Is there anyway I can get the same file/folder navigation regardless of whether I'm looking at an sftp folder or a folder on my client machine? Cheers, Chrsi

    Read the article

  • Removing integrated video card drivers broke other hardware?

    - by jsnlxndrlv
    I just purchased and installed an XFX ATI Radeon HD 4650 video card. The installation instructions said to remove the old video card drivers before installing the new driver. I did both things, but now every time I start my computer, the Found New Hardware wizard pops up for the following three devices: PCI Device SM Bus Controller Other PCI Bridge Device Furthermore, my LAN port no longer functions. The cable's connected, and the light on the port comes on, but the computer can't find any servers when I attempt to connect with it. The modem and router are working, and obviously I'm able to use the other computer here to post this. What do I do to fix this?

    Read the article

  • node.js server not running

    - by CMDadabo
    I am trying to learn node.js, but I'm having trouble getting the simple server to run on localhost:8888. Here is the code for server.js: var http = require("http"); http.createServer(function(request, response) { response.writeHead(200, {"Content-Type": "text/plain"}); response.write("Hello World"); response.end(); }).listen(8888); server.js runs without errors, and trying netstat -an | grep 8888 from terminal returns tcp4 0 0 *.8888 *.* LISTEN However, when I go to localhost:8888 in a browser, it says that it cannot be found. I've looked at all the related questions, and nothing has worked so far. I've tried different ports, etc. I know that my router blocks incoming traffic on port 8888, but shouldn't that not matter if I'm trying to access it locally? I've run tomcat servers on this port before, for example. Thanks so much for your help! node.js version: v0.6.15 OS: Mac OS 10.6.8

    Read the article

  • ADFS http login failure not re-requesting credentials

    - by Devnull
    We have ADFS working with HTTP (401) login. If a user types their password incorrectly, ADFS barfs and requires that the browser be closed, rather than asking the user for to attempt to log in again. Reprompting for user credentials is the typical behavior with other web servers (even IIS). This appears to be an artifact of setting the HTTP session, but other HTTP-login applications dont behave this way. We are having additional issues now because some users are saving that password, and its causing them account lockouts because the browsers do not realize they need to update saved credentials. Anyone know of a workaround? Wed rather not enable forms login if possible.

    Read the article

  • Server drives: 2.5" SCSI less reliable than 3.5" ?

    - by Bill
    Just had an HP 2.5" SAS 10k drive fail on a RAID5 array after about 2.5 years. It made me wonder if this was a fluke or an indication that 2.5" drives are less reliable than 3.5" SAS drives. I've had many 3.5" SAS drives running for many years without any issues (knock on wood). I would think that smaller drives would generate less heat and therefore be more reliable, but couldn't find any evidence of this. I realize all drives will eventually fail and that it's a crap shoot with any particular model, but was hoping someone could point out some related studies or comment on the SCSI drive sizes they've found to be most reliable in servers. Thanks.

    Read the article

  • How to keep programs from source up to date?

    - by wizard
    I'm designing a new server setup for hosting multiple websites. (Shared hosting for my clients over at SliceHost.) I've recently moved away from the traditional LAMP setup and chosen Ubuntu, Nginx, php-fpm and mysql. I like it a lot better then my old Apache, suphp, mysql setup. It works great, provided encapsulation between sites and uses substantiallly less memory. However I have one major maintenance problem. In order to have a recent version of Nginx and in order to use php-fpm I've had to compile these programs from source. The reason I see this as a problem is that keeping track of updates, and build configurations will end up being a lot of work. For two programs (and a patch) I can handle it, but it seems like this setup would not scale with many packages and servers. Are there good ways to manage this situation? I'm sure people do this all the time.

    Read the article

  • Can i change the subnet on the vpn side of a server without having to change its whole lan (to avoid collission)

    - by Gusty
    Ohai, ive got a xp server with a client connecting via VPN. Problem (as we all know) is that sometimes the subnets clash. Instead of changing the whole server network every time this happens, cant i just have the server appear to have a different subnet to vpn clients? I have no interest in accessing other computers than the server on its LAN. I tried just turning off automatic dhcp in the servers vpn settings and changed it to 172.31.255.x. The client gets the address assigned alright and it can ping 172.31.255.1 but i cannot ping the server name (because the dns on the clients vpn connection points at 192.168.0.1?) wisdom? PS everything worked until the client hit a network with the same subnet as the server lan. thanks

    Read the article

  • Yum through http proxy

    - by eodchop
    I have several Fedora 13 servers that have to connect through an http proxy for yum updates. All port 80 traffic has to be routed through this proxy. I have setup the proxy server in the network settings GUI. I can browse the internet just fine. I have also setup my proxy information in /etc/yum.conf as follows: proxy=http:proxy.largecorp.corp/accelerated_pac_base.pac proxy_user=user proxy_password=password I then added the export HTTP_PROXY="http:proxy.largecorp.corp/accelerated_pac_base.pac" to /etc/bashrc and sourced the file. When i run yum update: Loaded plugins:presto, refresh-packagekit Error: Cannot retrieve repository metadata (repomd.xml) fro repository: fedora. Please verify its path and try again. All of the repo urls are the defaults, as this is a fresh install.

    Read the article

  • Failover of mirrored webservice

    - by eflat
    We are using a webservice (over http) which has 2 mirrored servers, accessible as, say www.blah.com and www2.blah.com. Is there a software solution that would help us handle failover? Currently if one server becomes unavailable, I need to manually edit our config to point at the other server. Failover on their side is "in the works", so I don't want to do checking for server availability in code. On our side, we use a mix of linux and windows boxes, so we have both os's in the cage.

    Read the article

  • Time Capsule + Ubee Router?

    - by Charlie
    I can't for the life of me figure this out. I recently had TWC installed in my house, and wanted to disable the NAT and router functions of it. I have a Time Capsule hooked up to it from LAN1 (on the Ubee) to the WAN port on the TC. The problems started occurring here. I figured the settings would be these: Ubee Configuration mode: Bridge DHCP: Off TC IPv4: 192.168.100.2 Subnet Mask: 255.255.255.0 Router Address: 192.168.100.1 DNS Servers: 8.8.8.8, 8.8.4.4 Router Mode: DHCP and NAT But using those settings, my TC says "Double NAT", so I have to change it all around to the default settings of the Ubee using NAT. This leads me to believe bridge mode doesn't actually turn off NAT...

    Read the article

  • Automate setup of constrained kerberos delegation in AD

    - by Grhm
    I have a web app that uses some backend servers (UNC, HTTP and SQL). To get this working I need to configure ServicePrincipalNames for the account running the IIS AppPool and then allow kerberos delegation to the backend services. I know how to configure this through the "Delegation" tab of the AD Users and Computers tool. However, the application is going to be deployed to a number of Active Directory environments. Configuring delegation manually has proved to be error prone and debugging the issues misconfiguration causes is time consuming. I'd like to create an installation script or program that can do this for me. Does anyone know how to script or programmatically set constrained delegation within AD? Failing that how can I script reading the allowed services for a user to validate that it has been setup correctly?

    Read the article

  • What is the world wide web? [closed]

    - by think123
    I don't know where to post this question, so please move it if necessary. Ok, so I've heard of how the professional hosting companies can create 'links' to the world wide web to register an unregistered domain. So that's where my question comes from. Is the world wide web a server to which servers link? Is it created by abstract linkage? I'm not sure. Also, what does it mean for the DNS to be updated throughout the whole world?

    Read the article

< Previous Page | 331 332 333 334 335 336 337 338 339 340 341 342  | Next Page >