Search Results

Search found 12484 results on 500 pages for 'seraphims host'.

Page 142/500 | < Previous Page | 138 139 140 141 142 143 144 145 146 147 148 149  | Next Page >

  • Why does httpd handle requests for wrong hostnames in SSL mode?

    - by Manuel
    I have an SSL-enabled virtual host for my sites at example.com:10443 Listen 10443 <VirtualHost _default_:10443> ServerName example.com:10443 ServerAdmin [email protected] ErrorLog "/var/log/httpd/error_log" TransferLog "/var/log/httpd/access_log" SSLEngine on SSLProtocol all -SSLv2 SSLCipherSuite HIGH:MEDIUM:!aNULL:!MD5 SSLCertificateFile "/etc/ssl/private/example.com.crt" SSLCertificateKeyFile "/etc/ssl/private/example.com.key" SSLCertificateChainFile "/etc/ssl/private/sub.class1.server.ca.pem" SSLCACertificateFile "/etc/ssl/private/StartCom.pem" </VirtualHost> Browsing to https://example.com:10443/ works as expected. However, also browsing to https://subdomain.example.com:10443/ (with DNS set) shows me the same pages (after SSL certificate warning). I would have expected the directive ServerName example.com:10443 to reject all connection attempts to other server names. How can I tell the virtual host not to serve requests for URLs other than the top-level one?

    Read the article

  • Running bridged-networking vmware player on a Linux machine with 2 interfaces

    - by Roman D
    I have got a laptop running Arch Linux with 2 interfaces: wireless (wlan0) and ethernet (eth0). I use wlan0 to access internet (static IP, networking is configured using netcfg), and I connect a second PC to the eth0. Now, whenever I start vmware player (v. 4.0.4), it chooses wlan0 to connect its bridged virtual NIC to, but I need it to connect to eth0 (I want my guest machine to be able to talk to the second physical PC on eth0). So, I disable the wlan0 interface (netcfg -d wireless) and restart vmware. Now, it connects to eth0, and everything works fine; I can ping the host PC from the virtual one, and I can ping the virtual PC from the second physical PC connected to eth0. Then, if I try to reenable the wlan0 interface (netcfg -u wireless), all of the connectivity between the host and the guest (and between the second physical PC and the guest) gets lost, until I disable wlan0 again. Can someone please give me a hint on what's going on?

    Read the article

  • DNS resolution over DHCP

    - by Eric
    I have a m0n0wall router a VMWare workstation running ubuntu a windows 7 workstation running the VM The ubuntu hostname is "renraku" From the windows machine I can't resolve dns automatically for this host. For example, when I ping renraku Ping request could not find host renraku. Please check the name and try again. However nslookup seems to work nslookup renraku Server: m0n0wall.local Address: 192.168.123.254 Name: renraku Address: 192.168.123.248 I don't get how to have ping to work with hostnames. The main goal behind this is to have my web server to work with hostnames instead of ip addresses EDIT : ping 192.168.123.248 works

    Read the article

  • Unable to call through asterisk

    - by sk
    I want to create a voip service. I have installed asterisk-1.4 on a dedicated remotely hosted debian lenny distro. I made a sip.conf and extensions.conf so as to place a call between two sip phones(i am using xlite 3.0) installed in some other Windows PC. Whenever i switch this phones the asterisk console shows that Registration from '"1000"<sip:[email protected]>' failed for '122.168.10.254' - Peer is not supposed to register Where xx.xx.xx.xx is the server's IP. i.e my sip phones are unable to register with the asterisk server. Please help me to place call between two sip phones #sip show peers Name/username Host Dyn Nat ACL Port Status 2000 (Unspecified) D 0 Unmonitored 1000 (Unspecified) D 0 Unmonitored 2 sip peers [Monitored: 0 online, 0 offline Unmonitored: 0 online, 2 offline] # sip show registry Host Username Refresh State Reg.Time # sip show channels Peer User/ANR Call ID Seq (Tx/Rx) Format Hold Last Message 0 active SIP channels

    Read the article

  • How to copy directories using debugfs?

    - by tjbp
    The debugfs manpage gives the impression that the command 'rdump . .' will recursively copy all files found on the specified filesystem from the debugfs cwd to the native filesystem's cwd. Instead I seem to receive a syntax error, and no copy is initiated? These are the commands I run: cd /path/to/transfer/destination debugfs /dev/sda1 -R rdump . . My task is to copy the entire contents of a clean yet unmountable USB storage device to its host machine's HD. The host machine does not support the inode size used by the USB device's filesystem (256) and its software is not upgradeable, so my intention was to use debugfs to transfer the files. If anyone has any other suggestions for this task I'd be grateful.

    Read the article

  • Guestfish not found on Debian 7 when execute virt-copy-in [on hold]

    - by John Wang
    When I execute virt-copy-in command on a kvm host(Debian7.1), I got error saying "guestfish not found". However according to the dpkg comamnd, guestfish do have been installed: john@sver:~$ dpkg-query -l | grep guestfs* ii libguestfs-perl ... ii libguestfs-tools ... ii libguestfs0 ... What's the problem? Is the libguestfs-tools not the guestfish in Debian? Or it's just a broken dependency in libguestfs-tools in Debian7.1(my KVM Host)?

    Read the article

  • WCF Service in Windows Services

    - by sivakumar
    I create WCF service library and i test that working fine on WCF Test client(default). when i host the WCF service in winodws service that time i got the error. I am using windows XP sp3, .Net 3.5 and Visual Studio 2008. i got error. Error opening host : HTTP could not register URL "http://+:8731/WCFServerDLL/Service1/." Your process does not have access rights to this namespace (see "http://go.microsoft.com/fwlink/?LinkId=70353" for details). the above link for microsoft i implement the httpcfg. Here i run the "httpcfg.exe set urlacl /u http://localhost:8731/WCFServerDLL/Service1/ /a" i get the result HttpSetServiceConfiguration completed with 0. what is the problem i got same error. can you give me a suggation.

    Read the article

  • Does Windows Virtual PC support virtual applications with an XP Home Edition guest?

    - by endolith
    I've installed Windows Virtual PC in Windows 7 and have the XP Mode virtual PC working. I can run virtual applications with it and the integration features all work. I used Disk2VHD to convert my existing XP Home drive into a VHD, so I can use it as a virtual PC, too. It works in general, but it sometimes pops up the "Could not enable integration features" error. I don't see the host computer's drives in the guest, and I don't see the guest's applications in the host's Start menu. Is this just because the guest is XP Home instead of XP Pro? Do I have to reinstall all these apps in the XP Mode VHD in order to get them as virtual apps? Could something else be preventing it from working?

    Read the article

  • WinXP - Having trouble sharing internet with 3G USB modem via ICS

    - by Carlos Nunez
    all! I've been banging my head against a wall with this issue for a few days now and am hoping someone can help out. I recently signed up for T-Mobile's webConnect 3G/4G service to replace the faltering (and slow) DSL connection in my apartment. The goal was to put the SIM in one of my old phones and use its built-in WLAN tethering feature to share Internet out to rest of my computers. I quickly found out that webConnect-provisioned SIMs do not work with regular smartphones, so I was forced to either buy a 4G-compatible router or tether one of my old laptops to my wireless router and share out that way. I chose the latter, and it's sharpening my inner masochistic self by the day. Here's the setup: GSM USB modem (via hub), ICS host - 10/100 Mbps Ethernet NIC, ICS "guest" - WAN port of my SMC WGBR14N wireless router in bridged mode (i.e. wireless access point). Ideally, this would make my laptop the DHCP server and internet gateway with the WAP giving everyone wireless coverage. I can browse internet on the host laptop fine. However, when clients try to connect, they get a DHCP-assigned IP from the laptop and are able to use the Internet for a few minutes before completely dying. After that happens, they are able to re-associate with the WAP and get IP addresses, but are unable to use Internet or resolve IP addresses until the laptop and router are restarted. If they do get access, it's very, very slow. After running Wireshark on the host machine, it turns out that this is because every TCP connection keeps getting RST. DNS seems to work. I would normally think the firewall is the culprit here, but when it drops packets, it drops them completely. The fact that TCP connections are being ACK'ed by the destination rules that out. Of course, none of the event Log isn't saying anything about what's going on. I also tried disabling power management on the NIC, since that's caused problems in the past; that didn't help either. I finally disabled receive-side scaling as per a Microsoft KB (that applied to Windows Server 2003, SP2) to no avail. I'm thinking of trying it with a different NIC (will be tough; don't have a spare Ethernet NIC around for the laptop), but I'm getting the impression that this simply doesn't work. Can anyone please advise? I apologise for the length of this post; all contributions are much appreciated! -Carlos.

    Read the article

  • Plesk 10 Postfix with multiple IP adresses and SSL certificates

    - by JulianB
    We are currently running a root server with Debian 6 and Plesk 10.4.4. We have some virtual hosts using one IP adress (shared) - e.g. example1.com - and another virtual host using a dedicated IP address (example2.com). Is there a way to configure postfix to do the following Always use the IP address of the virtual host to which the e-mail account belongs (so that an e-mail from [email protected] will originate from the shared IP-Address and an e-mail from [email protected] will originate from the dedicated IP? Use different certificates for TLS for example1.com and example2.com? If the latter is not possible: Could any problems arrive when using example1.com as certificate for example2.com users? Of course, example2.com users would have to configure their clients to use example1.com as the SMTP server name to avoid annoying security warnings. But if we still would be able to get the effect of the first point that would still be acceptable.

    Read the article

  • Conect to remote server using

    - by Vlad
    Hi, I’m using Zen Cart (php) for my web site. I have images that is on my supplier’s remote server and my site is on Host Gator the only way to get the images show up on my site is to connect to their server using links to the images (per agreement). They tell me that ZenCart PHP does not allow you to serve images from a remote host. And I need to have a PHP programmer take a look at the code to see if he can manipulate the source code to allow this. Does anyone have solution for this? Thank you!! Vladi

    Read the article

  • Is it possible to run two separate instances (copies, versions) of Audacity?

    - by cipricus
    What I want is two separate instances of Audacity running in parallel as to separate programs with different settings. I want one to use WASAPI as audio host and speakers as input device (the 2.0.5 portable has the WASAPI option), and the other to use the MME as audio host, and microphone as input device. When opening a new (second) window of the same installation/copy of Audacity, by executing twice the .exe file, it is the same as giving the command 'new' under 'File', what happens is not a second instance, but just a separate project/file and window. When settings are changed in one, they change exactly in the other. When trying to start a second copy of Audacity, or even a second different version, when one is already running, it will also open just a second project/file of the application that is already running, as before.

    Read the article

  • Self-hosted image gallery

    - by Robert Munteanu
    I'd like to host an image gallery for family and friends. Is there any solution out here which offers: albums ( I expect most do ); ability to specificy tags/persons/places etc, similar to what f-spot does; an API to upload multiple photos or just a bulk uploader; block out unregistered users; allow users to leave comments. Update: please do assume that I'm able to host the service myself if I've asked for a self-hosted solution.

    Read the article

  • Name Servers changed over 48 hrs ago and site is still down

    - by Solomon Closson
    Ok, I just recently changed Name Servers for my site (Oct. 29th): http://dream-portal.net but there is an issue, the name servers were changed from godaddy, however, the DNS is being hosted elsewhere. Godaddy states that the DNS is resolving the site to www.dream-portal.net and that is making it inaccessible from the DNS on the other host and that they can't do anything because I'm not hosting DNS with godaddy. I've contacted the other host and they say that everything is fine, however, people are not able to access the site and I can't access it from my work computer. I can only access it from my home computer. From my work computer it says that the page can not be found. What should I do? What can I do? Other people reported problems with the site here: http://www.simplemachines.org/community/index.php?topic=487592.msg3629960#msg3629960

    Read the article

  • Virtualized CPU cores vs. threads

    - by nedm
    We've got a KVM host system on Ubuntu 9.10 with a newer Quad-core Xeon CPU with hyperthreading. As detailed on Intel's product page, the processor has 4 cores but 8 threads. /proc/cpuinfo and htop both list 8 processors, though each one states 4 cores in cpuinfo. KVM/QEMU also reports 8 VCPUs available to assign to guests. My question is when I'm allocating VCPUs to VM guests, should I allocate per-core or per-thread? Since KVM/QEMU reports the server has 8 VCPUs to allocate, should I go ahead and set a guest to use 4 CPUs where I previously would have set it to use 2 (assuming 4 total VCPUs available)? I'd like to get the most possible out of the host hardware without over-allocating.

    Read the article

  • Audio doesn't work on Windows XP guest (WS 7.0)

    - by Mads
    Hi, I can't get audio to work with on a Windows XP guest running on VMware Workstation 7.0 and Ubuntu 9.10 host. Windows fails to produce any audio output and the Windows device manager says the Multimedia Audio Controller is not working properly. Audio is working fine in the host OS. When I open Multimedia Audio Controller properties it says: Device status: The drivers for this device are not installed (Code 28) If I try to reinstall the driver I get the following error message: "Cannot Install this Hardware There was a problem installing this hardware: Multimedia Audio Controller An Error occurred during the installation of the device Driver is not intended for this platform" Has anyone else experienced this problem?

    Read the article

  • DNS resolution Windows 7 & browsing to locally hosted web site

    - by Aidan Whitehall
    We host two Intranet sites, http://intranet/ and http://sales.intranet/, both on the same server on the LAN. Local DNS (a Windows 2003 Server) was updated and both hostnames are configured to be CNAMEs that point to the FQDN name of the server on which they're hosted. On the LAN, Windows XP Professional clients can browse to both sites. However, Windows 7 Professional clients can browse to the main Intranet site, but not the Sales Intranet (neither using Firefox 3 nor Internet Explorer 8). Using nslookup on the command line on the Windows 7 boxes, intranet and sales.intranet both correctly resolve as CNAMEs of the server hosting them, and that in turn correctly resolves to the host's IP address. So the Q is... can anyone think why this might be, or what test to try next? Thank you for any suggestions!

    Read the article

  • Snapshot/rollback for libvirt+KVM?

    - by jtimberman
    I've recently begun using KVM for my development/test environment on a Linux host system with 8G memory. Prior, I was using VMware Fusion for my virtual environment, but my Macbook only has 2G memory. I tried VMware Server and ESX on the host instead of KVM, but the webUI doesn't run on Mac OSX's Firefox, and we're going to be doing more with KVM anyway. The main feature of VMware I miss is robust snapshot/rollback, but I'm missing this in KVM. I understand the snapshot command, but it shuts down the guest OS when complete, and then copying the disk image to preserve its state is cumbersome. Is this really the best way to manage snapshots on KVM?

    Read the article

  • 403 forbidden while submitting a POST request with image data via iPhone application

    - by binnyb
    I am creating an iOS application which allows users to send image/text data to my webserver via a POST request. I am successfully sending POSTS to the server when image data is not included in the request. Any time i POST with image data the server spits back a 403 forbidden. I have tried adding the following to the .htaccess file in the directory of the script with no luck: Options +Indexes FollowSymLinks +ExecCGI Order allow,deny Allow from all web browsers and Android devices can successfully POST with image data to the script, the only device which cannot is the iPhone. POSTING with data to other hosting providers works as expected - it is just this host(ipowerweb.com). i noticed that when i try to POST to ANY script on the server with data returns a 403 forbidden. another note: i can successfully post to another server that is hosted by ipowerweb, but mine cant seem to handle it. My host has tried to resolve the issue but cannot, and they have marked it on their end as "resolved", so no more help from them. I wish to keep this host as moving would be a pain - i will change hosts as a last resort, so please help me! Why am i getting this 403 forbidden error only when i submit data via my iPhone application? How can i resolve the issue so i can successfully POST data? any advice on what i can do would be greatly appreciated. edit: as request, here are the response headers: { Connection = close; "Content-Length" = 217; "Content-Type" = "text/html; charset=iso-8859-1"; Date = "Wed, 12 Jan 2011 19:11:19 GMT"; Server = "Apache/2"; } edit: as request here are the request headers(oops): { "Accept-Encoding" = gzip; "Content-Length" = 5781; "Content-Type" = "multipart/form-data; charset=utf-8; boundary=0xKhTmLbOuNdArY"; "User-Agent" = "YeahIAteThat 1.0 (iPhone; iPhone OS 4.2.1; en_US)"; }

    Read the article

  • 403 Forbidden

    - by demas
    Here is my Nginx config: user pass users; worker_processes 1; events { worker_connections 1024; } http { passenger_root /usr/lib64/ruby/gems/1.8/gems/passenger-3.0.7; passenger_ruby /usr/bin/ruby; include mime.types; default_type application/octet-stream; sendfile on; keepalive_timeout 65; server { listen 80; server_name some.another.ru; root /www/public/redmine; passenger_enabled on; rails_env development; } } Here is Nginx log: 2011/06/02 12:53:57 [error] 45986#0: *1 directory index of "/www/public/redmine/" is forbidden, client: **.*.**.***, server: some.another.ru, request: "GET / HTTP/1.1", host: "some.another.ru" 2011/06/02 12:53:59 [error] 45986#0: *1 open() "/www/public/redmine/favicon.ico" failed (2: No such file or directory), client: **.*.**.***, server: some.another.ru, request: "GET /favicon.ico HTTP/1.1", host: "some.another.ru" What is the reason of this error and how can I fix it?

    Read the article

  • Help about NAT with virtual server

    - by Thanh Tran
    I have a dedicated server running Linux CentOS 5.3 with 2 IP addresses. I've installed a virtual machine using VMware Server. The host and the guest have a host-only network. Now I want to map the 2nd IP address to the virtual machine so that it can run as a second dedicated server for me. Here is what I do: modprobe iptable_nat echo "1" > /proc/sys/net/ipv4/ip_forward iptables -t filter -A FORWARD -s 192.168.78.128 -d 64.85.164.184 -j ACCEPT iptables -t nat -A PREROUTING -d 64.85.164.184 -i eth0 -j DNAT --to-destination 192.168.78.128 iptables -t nat -A POSTROUTING -s 192.168.78.128 -o eth0 -j SNAT --to-source 64.85.164.184</p> But it not working as intended. What is the matter?

    Read the article

  • Testing Tomcat with Virtual Hosts

    - by Marty Pitt
    I'm trying to test Tomcat virtual hosts on my dev machine (windows 7/Tomcat 6). I'd like to have requests for localhost, test1.localhost and test2.localhost all route through to the same tomcat instance. I've edited my hosts file to look as follows: 127.0.0.1 localhost ::1 localhost 127.0.0.1 test1.localhost 127.0.0.1 test2.localhost And added modified the Engine in server.xml as follows: <Engine defaultHost="localhost" name="Catalina"> <Realm className="org.apache.catalina.realm.UserDatabaseRealm" resourceName="UserDatabase" /> <Host appBase="webapps" autoDeploy="true" name="localhost" unpackWARs="true" xmlNamespaceAware="false" xmlValidation="false"> <Alias>test1.localhost</Alias> <Alias>test2.localhost</Alias> </Host> </Engine> However, I'm getting a 404 when hitting test1.localhost:8080/myWebApp, although localhost:8080/myWebApp works fine. I can ping test1.localhost fine. What have I missed?

    Read the article

  • Spam issues while using Postfix as a two-way relay

    - by BenGC
    I want to use a Postfix box to do two things: Relay mail from any host on the internet addressed to one of my domains to my Zimbra server Relay mail from my Zimbra server to any address on the internet. To try and accomplish this I have configured Postfix thusly: mynetworks = 127.0.0.0/8, zimbra_ip/32 myorigin = zimbra_server mydestination = localhost, zimbra_server relay_domains = example.com example.org transport_maps = hash:/etc/postfix/transport_map local_transport = error:no mailboxes on this host transport_map looks like this: example.com smtp:[zimbra_server] example.org smtp:[zimbra_server] Now, this works and passes the Open Relay tests. However, I am seeing in the maillog that the server is relaying spam that has a From: address of <> to domains that are not mine. How do I stop this behavior?

    Read the article

  • Can't create a valid symlink under VMWare HGFS

    - by Alexander Gladysh
    Host: OS X 10.6.5 VMWare Fusion: 3.1.2 Guest: Ubuntu x86 10.10 $ uname -a Linux ubuntu 2.6.35-24-generic #42-Ubuntu SMP Thu Dec 2 01:41:57 UTC 2010 i686 GNU/Linux I can not create a symlink, readable from the Guest OS anywhere in the directory, mounted with hgfs: /mnt/hgfs/projects/tmp$ touch aaa /mnt/hgfs/projects/tmp$ ln -s aaa bbb /mnt/hgfs/projects/tmp$ less bbb bbb: No such file or directory /mnt/hgfs/projects/tmp$ ls -la total 6 drwxr-xr-x 1 501 users 136 2010-12-28 18:12 . drwxr-xr-x 1 501 users 8602 2010-12-28 18:12 .. -rw-r--r-- 1 501 users 0 2010-12-28 18:12 aaa lrwxr-xr-x 1 501 users 3 2010-12-28 18:12 bbb - aaa /mnt/hgfs/projects/tmp$ readlink bbb aaa The same symlink is perfectly accessible in OS X host. Is there a workaround for this?

    Read the article

  • Trying to install wordpress inside rails app with nginx and fastcgi

    - by pinouchon
    I have a rails app (let's call it myapp) running at www.myapp.com. I want to add a wordpress blog at www.myapp.com/blog. The webserver for the rails app is thin (see the upstream block). The wordpress runs with php-fastcgi. The rails app works fine. My problem is the following: in /home/myapp/myapp/log/error.log error I get: 2013/06/24 10:19:40 [error] 26066#0: *4 connect() failed (111: Connection refused) while connecti\ ng to upstream, client: xx.xx.138.20, server: www.myapp.com, request: "GET /blog/ HTTP/1.1", \ upstream: "fastcgi://127.0.0.1:9000", host: "www.myapp.com" Here is the nginx conf file: upstream myapp { server unix:/tmp/thin_myapp.0.sock; server unix:/tmp/thin_myapp.1.sock; server unix:/tmp/thin_myapp2.sock; } server { listen 80; server_name www.myapp.com; client_max_body_size 20M; access_log /home/myapp/myapp/log/access.log; error_log /home/myapp/myapp/log/error.log error; root /home/myapp/myapp/public; index index.html; location / { proxy_set_header X-Real-IP $remote_addr; proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for; proxy_set_header Host $http_host; proxy_redirect off; # Index HTML Files if (-f $document_root/cache/$uri/index.html) { rewrite (.*) /cache/$1/index.html break; } if (!-f $request_filename) { proxy_pass http://myapp; break; } # try_files /system/maintenance.html $uri $uri/index.html $uri.html @ruby; } location /blog/ { root /var/www/wordpress; fastcgi_index index.php; if (!-e $request_filename) { rewrite ^(.*)$ /blog/index.php?q=$1 last; } include /etc/nginx/fastcgi_params; fastcgi_param SCRIPT_FILENAME /var/www/wordpress$fastcgi_script_name; fastcgi_pass localhost:9000; # port to FastCGI } } Any ideas why that doesn't work ? How do I make sure that php-factcgi is configured properly ? Edit: I cant test if fastcgi is running with telnet: $> telnet 127.0.0.1 9000 Trying 127.0.0.1... telnet: Unable to connect to remote host: Connection refused And it's not.

    Read the article

< Previous Page | 138 139 140 141 142 143 144 145 146 147 148 149  | Next Page >