Search Results

Search found 13411 results on 537 pages for 'proxy servers'.

Page 369/537 | < Previous Page | 365 366 367 368 369 370 371 372 373 374 375 376  | Next Page >

  • How much power supply do I need for my server, and could a shortage be causing my odd crashing?

    - by dolan
    I have 5 servers, all with similar hardware (i7, four 2tb 7200rpm drives, two 4tb 5400rpm drives, 430 watt power supply), and lately the machines have been freezing up. This has gotten worse in the last day or so, and I can't pinpoint any explanation. One recent change was adding the two 4tb hard drives. The crashes happen most often while running a large Hadoop job, so I was originally thinking the load was causing some issues, but last night one server just froze without any heavy load on the box (or so I think), other than HDFS (Hadoop's distributed file system) was probably rebalancing itself since two of the five nodes were offline. If I plugin a monitor and keyboard to one of these frozen machines, I can't get any response or feedback on the screen. Any ideas on possible points of failure and/or different logs I can look at to investigate? Thanks Edit: The systems are running Ubuntu 10.04 Edit 2: More on hardware: intel core i7-930 bloomfield 2.8ghz processor (quad core) 12gb (6 x 2gb) kingston ddr3 1333 ram antec earthwatts green 430 power supply msi x58m lga 1366 motherboard

    Read the article

  • Slowness on Exchange 2007 Server (network related we think)

    - by Az
    We are seeing some odd behavior out of our exchange 2007 server. I believe it is network related as during the slowness in response from the exchange server. During these times downloading files to server is fraught with connection resets and extreme slow download speeds that are not affecting other servers right on the same subnet with the same gateway and ISP. We have about 50 users at a time using Outlook 2003 from a terminal server not displaying any other signs of slowness, all network slowness is confirmed trying downloads on the exchange server itself. Checked for virus issues, there are none. Checked for hardware issues, none found I am noticing between 200-400 TCPv4 connections established I am seeing around 100k tcpv4(active) connections according to perfmon are these numbers odd? Has anyone run into anything similar? Our server has a gigbabit NIC linked to a gigabit switch using cat6 so I do not believe it is "physical". Thanks!

    Read the article

  • BIG IP - HTTPS Health Monitor setup

    - by djo
    I have a Web site that we have setup a health monitoring pages so we can take our servers in and out of the Big-IP as we see fit. Now we have just moved onto Big-IP and the issue I have hit is that you setup Health Monitors for port 80 and 443, now the 80 check works fine but when I to get the 443 check to look at our file it fails. Now I am aware as I am hitting the this page on the IP address over HTTPS is going to cause a cert error but I would have guessed that BIG-Ip would have been setup just to accept the cert and carry on with the check. Is what I am wanting to do possible? Also is there a way of just using a HTTP monitor for HTTPS? Because if port 80 has stopped sending traffic then if i use the same monitor for 443 it will stop traffic to that. Any help would be great! Thanks

    Read the article

  • Domain name in server other than the website

    - by med
    Hello! I'm not very popular with hosting, but I have a special situation: I live in Tunisia, I can buy a domain name with .tn extention, but the problem is that: 1- The domain could not be pointed to a server outside Tunisia 2- All servers in Tunisia are bad, no one provides really reliable hosting so, I want to use the .tn domain with a basic hosting in Tunisia, and Make other db queries and rich media on another remote server outside Tunisia. How to do it? Is there better alternatives? All suggestions are welcome :) Thank you.

    Read the article

  • MySQL Cluster Failover doesn't work

    - by Lukasz
    I have two servers, where First server 10.100.15.150: 1. one mgm server 2. one ndbd 3. one mysql api Second server 10.100.15.160: 1. one ndbd 2. one mysql api When i start all 'parts' of cluster it looks : Cluster Configuration [ndbd(NDB)] 2 node(s) id=21 @10.100.15.150 (mysql-5.1.56 ndb-7.1.17, Nodegroup: 0) id=22 @10.100.15.160 (mysql-5.1.56 ndb-7.1.17, Nodegroup: 0, Master) [ndb_mgmd(MGM)] 1 node(s) id=3 @10.100.15.150 (mysql-5.1.56 ndb-7.1.17) [mysqld(API)] 2 node(s) id=11 @10.100.15.150 (mysql-5.1.56 ndb-7.1.17) id=12 @10.100.15.160 (mysql-5.1.56 ndb-7.1.17) When i shutdown first machine - 10.100.15.150, on second the nbdb process also has been shutdown so i cannot use this data node and cluster fail ... How i must configure this cluster to get FailOver working ? Thx

    Read the article

  • How Exchange integrates with DC

    - by TheD
    This probably is a very open ended question, but really I'm only looking for a specific aspect I suppose. It relates to a reboot question I posted earlier, relating to restart orders of servers, Server Restart's and Respective Orders. Please take a look if you get a second for input! Basically - in relation to how Exchange integrates with AD, what would happen in two scenarios: 1). The Exchange server is booted up before the DC is online 2). You replace a DC but keep your current Exchange server. I'm fairly new to all of this, so hopefully this isn't a silly question (I'm an apprentice in fact!). Many thanks

    Read the article

  • nginx conditional Accept header

    - by manu_v
    Some mobile devices send the following incorrect requests to our servers : GET / HTTP/1.0 Accept: User-Agent : xxx The empty Accept header causes our Ruby on Rails server to throw back a 500 error. In Apache, the following directive allows us to rewrite the header before sending it to the application RoR server in order to cope with the broken devices : RequestHeader edit Accept ^$ "*/*" early We're currently setting up nginx, but achieving the same work-around is proving difficult. We are able to set : proxy_set_header Accept */*; However, this seems to have to be done inconditionally. Whenever trying to do : if ($http_accept !~ ".") { proxy_set_header Accept */*; } It complains with the message : "proxy_set_header" directive is not allowed here So, using nginx, how can we set the HTTP Accept header to */* when it is empty before sending the request to the application server ?

    Read the article

  • Windows 7 blocks network access to network-installed apps

    - by VokinLoksar
    Windows 2008 R2 domain. Users, running Windows 7 Enterprise, are trying to run some software from a network share. Specifically, I've tested this with MATLAB and PuTTY. When starting, MATLAB has to contact a licensing server to get its license. This action fails for regular users when they start MATLAB from the network share. However, if they copy the installation directory to a local disk everything works fine. Running MATLAB as an admin user from the network share also works. Same story with PuTTY. If the executable is launched from the share, regular users cannot connect to any servers. Something is blocking network communications for programs that are launched from a network drive. Here's the only other mention I could find of the same problem: https://social.technet.microsoft.com/Forums/en-US/w7itpronetworking/thread/4504b192-0bc0-4402-8e00-a936ea7e6dff It's not the Windows firewall or the IE security settings. Does anyone have any clue as to what this is?

    Read the article

  • Do all domains on the same shared hosting server have the same IP or ID

    - by silow
    Here's what I've got: siteA.com and siteB.com are hosted on hostgator. They're hosted on the same account of a shared hosting server (not VPS or dedicated). script.php is an external site that each of these 2 sites are accessing. I noticed that when siteA.com or siteB.com access the outside script.php, the script identifies them both as 1a.12.12ab.static.theplanet.com (apparently because hostgator uses theplanet.com servers). The fact that they're identified as the same value isn't surprising because after all they're hosted on the same account /home/user123/public_html. What I'm wondering about is how about other websites that are hosted on the same shared hosting server, but under other accounts. Basically other websites that are under another developer's control, but just happen to share the same hardware (hosting server). Do they also have the exact same identifier 1a.12.12ab.static.theplanet.com or that changes by account?

    Read the article

  • how to forward IP request to a specific port

    - by Jeremy Talus
    I have 2 servers the first (SRV01) is running Bind and other web app the second (SRV02) is running 2 server Minecraft (^^) in Bind I have 2 A recording for the 2 server MC s1.domain.tld A SRV02IP s2.domain.tld A SRV02IP the 2 MC serv are running on 2 different port 25565 and 25566 so I want that the request from s1.domain.tld:25565 are going to SRV02IP:25565 and the request from s2.domain.tld:25565 are going to SRV02IP:25566 I think I need do this in the SRV02 iptables. I have looking some topic about iptables but nothing pertinent to me. could you help me ? rgds.

    Read the article

  • Can't ssh from Ubuntu to RHEL or CentOS

    - by Alex N
    I am trying to setup publickey based authenitcation for 2 different boxes. One RHEL another on e is CentOS. I am having same issue with both where ssh fails and falls back to password based authentication. Error that seems to be causing this is quite obscure: debug1: Unspecified GSS failure. Minor code may provide more information Credentials cache file '/tmp/krb5cc_1000' not found Both boxes are completely unrelated. I have my public key in .ssh/authorized_keys file on both boxes, all permissions are checked and good(700 for .ssh and 600 for internals) I have bunch of other servers that are running on various flavors(Gentoo, Fedora, FreeBSD etc.) and publickey ssh works just fine, but CentOS and RHEL giving me this for some reason :( Anyone experienced this before? I am not even sure how to further analyze this issue :(

    Read the article

  • using gmail to send mails from your php?

    - by ajsie
    i've read some threads here in SO that someone managed to use gmail's smtp server to send out messages from php script. so if php sends a mail to [email protected] it looks like this: php tells sendmail (smtp server) to send the message sendmail sends the message to gmail's smtp server gmail's smtp server sends the message to hotmail's smtp server so now i wonder, why should one use gmail's smtp server to send a mail? isn't it better just to send the message from sendmail to hotmail's smtp server? correct me if im wrong, but isnt all that requires for a mail to be sent a communication between TWO smtp servers, the senders and the receivers (in this case, hotmail)? and another question, why can't php just send the message directly to hotmail's smtp server, why does it have to go through a local smtp server? and why do you call it a smtp server when sendmail is just acting like a client to send a message and does not receive anything? would be great if someone could shed a light on this topic!

    Read the article

  • puppet agent doesn't retrieve files from master

    - by nicmon
    I have a very basic question regarding to Puppet 3.0.1 configuration. I setup a puppet master server (CentOS) with 2 agents (CentOS and Windows 7), all 3 can ping and access each other. There is no error at all. I have copied a file under /etc/puppet/files/test2.txt my site.pp (/etc/puppet/manifests) contains these lines: node default { include test file { "/tmp/testmaster.txt": owner => root, group => root, mode => 644, source => "puppet:///files/test2.txt" } } but there will no file be created on agent servers under /tmp/ once I run "puppet agent --test" here is the output: [root@agent1 ~]# puppet agent --test Info: Retrieving plugin Info: Caching catalog for agent1.mydomain.com Info: Applying configuration version '1354267916' Finished catalog run in 0.02 seconds "puppet apply /etc/puppet/manifests/site.pp" creates the testmaster.txt under /tmp/ on master.

    Read the article

  • Error configuring virtual hosts

    - by user148351
    i Have a problem using my virtual hosts: When i try to connect to my server on direct ip adress, for example http://111.11.11.111/ in apache error log i see following error: script '/var/www/html/mmm/public/index.php' not found or unable to stat File index.php exists!!! and has correct access rights. I have virtual hosts configured <VirtualHost *:80> DocumentRoot /var/www/html/mmm/public ServerName example.com ServerAlias example.com www.example..com <Directory var/www/html/mmm/public> AllowOverride All </Directory> </VirtualHost> Why when I try to connect to ip address - it try to search index.php not in servers root directory, but in root directory of virtual host.

    Read the article

  • Bind9 forwarding zone not working

    - by JMW
    i've setup a forwarding zone on a RHEL6 Bind server like this: zone "office.local" IN { type forward; forward only; forwarders { 192.168.0.2; 192.168.0.3; }; }; when i try to query using dig @127.0.0.1 monitorsms.office.local i see the following message in the syslog: client 127.0.0.1#39376: query: monitorsms.office.local IN A + (127.0.0.1) validating @0x7ff7640357d0: monitorsms.office.local A: bad cache hit (monitorsms.office.local/DS) google tells me, that there is an issue with DNSSEC, but both servers do not have DNSSEC configured and thus do not send any DNSSEC records. What's wrong with my configuration?

    Read the article

  • Restrict VPN Clients to VPN Server

    - by Sprotty
    I've setup a VPN server using Debians pptpd. This all works, but I want to restrict incoming connections so they can only access ports on the VPN server and not get access to other machines on the VPN servers local network. I'm guessing this is a firewall rule? but i'm new to linux so am struggling a bit... ---- Additional info ----- The VPN Server is also hosting an SVN server, which is its real purpose, and the only service I want it to expose to incoming connections. I guess I could expose the SVN server directly, but I'm not confident that I could secure this correctly, so a VPN seemed to give an additional layer of security.

    Read the article

  • Setting up Mac VPN on simple network with separate IPs

    - by Dave
    Hi there, We have a pretty simple network configuration. A typical ASDL line with two extra IP addresses. We have a basic home network router Dlink DIR-655. We have two Mac OS X Snow Leopard Servers that we want to setup VPN on. The problem is that we can't change the ports which VPN works under ie. we can only setup port forwarding for UDP 500,1701,4500 to a single server. I'm wondering how we can cost effectively re-work our network so that we can take advantage of the separate IP addresses and have both VPNs working with the same router. Is that possible with our current hardware setup or do we need something more powerful? We're also toying with Mobile Access and wondering if that might be easier to setup. Expert help is greatly appreciated.

    Read the article

  • How can I setup a Firewall without NAT?

    - by SRobertJames
    We have 16 IP addresses from our ISP, and are setting up a SonicWall Firewall. I'd like to have the SonicWall do NAT for the LAN, but act as a firewall only (no NAT) for the servers which are using some of the 16 addresses. How do I set this up? If I set the WAN's subnet to include the 16 IPs, the SonicWall won't route the traffic to the LAN interface. Should I set the WAN subnet to only include the ones we are dedicating for NAT, and then keep the others on the LAN? Related point: How can I set multiple IP addresses for a SonicWall LAN interface?

    Read the article

  • Virtual MS Sql Server not consuming enough CPU

    - by rocketman
    We have a Win2008 server 32 bit running as a virtual machine under ESX server. It has 6 CPU cores of 2Gz each and 4GB ram. It's running MS Sql Server 2008 R2 only. Problem: The server is heavily loaded and responds slowly. From windows taskmanagers point of view, it really looks overloaded, CPU wise. However, our external "cloud manager" says it's only using 2.5GHz worth of CPU-cycles in the cluster. I/O times looks "good". We have already tried to set the SQL servers number of working threads from 0(auto) to 256, to no effect. How to tune the VM host, guest or SQL to use all of it's alotted resources? Does it sound possible att all?

    Read the article

  • Cannot redirect ip traffic with iptables to new ip on linux centOS

    - by Kiwi
    today I able to migrate some of the game servers to another server and needed some help to redirect the traffic from old ip to the new one. SERVER1 1.1.1.1 ----- (internet ) ----- SERVER 2.2.2.2 I asume to use iptables to perform this, for that used this rule on my centOS box in the server1. /etc/sysctl.conf: net.ipv4.ip_forward = 1 iptables -t nat -A PREROUTING -p udp --dest 1.1.1.1 --dport 27015 -j DNAT --to-destination 2.2.2.2:27015 iptables -t nat -A POSTROUTING -j MASQUERADE iptables -t nat -A POSTROUTING -d 2.2.2.2 -p udp --dport 27015 -j SNAT --to 1.1.1.1 But the client cannot connect to the server from the old ip, the redirection don't started.

    Read the article

  • What would cause SQL 2008 Log Reader Agent to fail with "This process could not execute 'sp_replcmds

    - by Rick
    I've seen this error message in other posts. They didn't seem to help resolving our issue. We are trying this with two SQL Server 2008 servers. I backed up my database from the source server and then restored it on our destination server. We setup basic Transaction Replication. The Snapshot Agent is working fine. The Log Reader Agent fails with the error above. Is it most likely a login issue for this job or QueryTimeout?

    Read the article

  • How can we implement network search for Windows AND OS X clients?

    - by michielvoo
    We have a network with Windows 7 and OS X (10.5 and 10.6) computers. Our servers run on Windows Server 2003 (1 Small Business Server, 2 Standard). We need to be able to search through about 15.000 - 30.000 documents in our archives. The best solution would be if users can search directly from the Windows menu (on Windows 7) or the Spotlight menu (on OS X 10.5 and 10.6). Also good would be if users can search directly from the search bar in their browsers, or by first visiting a site with the search form. In case the users search through the browser, it's important they they are able to open a file in the search results just by clicking on it. I have tested Microsoft Search Server Express, but it doesn't meet the requirements (no OS X support, results in the browser can't be opened by clicking in anything but Internet Explorer). I have looked at Spotlight server, but that only supports OS X. Thanks!

    Read the article

  • Is Zabbix the right tool for me?

    - by hortitude
    I just want to monitor a small handful of servers (less than 10). From reading various places it sounds like the top leading contenders (for open source at least) are: nagios munin zabbix From what I have read a lot of people tend to use munin and nagios together -- munin for history and graphs, and nagios for alerting. On the other hand it sounds like Zabbix is a more complete solution and easier to configure than either of the other two. So I was thinking of going that route. My thoughts right now are: What are the general disadvantages of Zabbix? Does Zabbix have a small footprint on boxes it is monitoring? Do I really need to setup an entire other server for it? I currently have a server that is under very light load -- can I dual purpose it?

    Read the article

  • Windows 2012 Cluster on P6300 SCSI-3 Persistent Reservation issues

    - by Bruno J. Melo
    Scenario: 1 HP 6300 with latest XCS version 1 Command View 10.1 + with hosts defined as Windows 2008 2 BL460c Gen8 Servers with SPP 2012.10 and Windows Server 2012 Datacenter Edition with all the updates + MPIO feature enabled DSM v4.03.00 Cluster Analyser Tool triggers this error: Test Disk 0 does not support SCSI-3 Persistent Reservations commands needed to support clustered Storage Pools. Some storage devices require specific firmware versions or settings to function properly with failover clusters. Please contact your storage administrator or storage vendor to check the configuration of the storage to allow it to function properly with failover clusters. Any ideas? Thanks for your help!

    Read the article

  • Need to move a debian server from i686 to x86_64 architecture

    - by user64204
    I have a debian server that I need to move from one hosting provider to another. I don't really know how the old server was setup, all I know is that it's running a Ruby on Rails application with a lot of custom libraries installed and that I should prepare myself for a painful migration. Old server: -os: debian 5.0.9 -used disk space: 3.2GB -architecture: i686 New server: -os: debian 5.0.9 -free disk space: 10GB -architecture: x86_64 As you can see the problem is that the servers are running different architectures. Q: Is there anyway I could somehow migrate the old to the new server in a few steps (or am I just dreaming I could) ? I was thinking maybe I could: -get list of packages and gems installed on old server and use for loop to install them all on the new -copy the disk content from old to new server while excluding what is architecture-specific (the problem is that I don't really know what to exclude).

    Read the article

< Previous Page | 365 366 367 368 369 370 371 372 373 374 375 376  | Next Page >