Search Results

Search found 59254 results on 2371 pages for 'http host'.

Page 37/2371 | < Previous Page | 33 34 35 36 37 38 39 40 41 42 43 44  | Next Page >

  • After changing web host, I get a 'file does not exist' error

    - by Jordan
    I run a WordPress blog, and have recently changed web hosts. When changing web hosts, I copied all files and exported/imported the database etc as explained by lots of tutorials found easily on Google. The blog home page works fine. What goes wrong: When I click on any link from the home page, the browser gets stuck in a redirect loop. Looking at the error log, I see: File does not exist: /usr/local/apache/htdocs/index.php The directory /usr doesn't even exist for my website - so perhaps this is looking for a file that was present using my old Web Host and is no longer present with my new web host? What is going on, and how might I resolve it?

    Read the article

  • Loading GUI on host machine through SSH with X11 forwarding

    - by cdated
    I've got applications on my host machine running fine by runnning: $ssh -X user@host Welcome to host $firefox Firefox proceeds to run in client display from the host machine. How do I get firefox to load on the host machine's display so I could remotely manage the host machine's running GUI applications? I need my application to load a GUI on a remote machine without walking to that machine.

    Read the article

  • Why does a Non-existant page returns 302 status when using a custom 404 page in asp.net

    - by webdevbytes
    I have setup custom 404 page custom404.aspx that returns a 404 Not Found error correctly, however the non-existant page that was initially requested returns a 302 Found status. So when I test thispagedoesnotexist.aspx, it returns a 302 Found then the custom404.aspx loads and returns a 404 Not Found status. I want to make sure that search spiders/bots understand that the requested page does not exist and should not show up in any search results. Is this the case? Cheers

    Read the article

  • Why would certain browsers request all pages on my ASP.Net Web site twice?

    - by Deane
    Firefox is issuing duplicate requests to my ASP.Net web site. It will request a page, get the response, then immediately issue the same request again (well, almost the same -- see below). This happens on every page of this particular Web site (but not any others). IE does not do this, but Chrome also does this. I have confirmed that there is no Location header in the response, and no Javascript or meta tag in the page which would cause the page to be re-requested (if any of these were true, IE would be re-requesting pages as well). I have confirmed this behavior on multiple Firefox installs on multiple machines. Versions vary, but all are 3.x. The only difference between the two requests is the Accepts header. For the first request, it looks like this: Accept: text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8 For the second request, it looks like this: Accept: */* The Content-Type response header in all cases is: Content-Type: text/html; charset=utf-8 Something else odd -- even though Firefox requests the page twice, it uses the first response and discards the second. I put a counter on a page that increments with every request. I can watch the responses come back (via the Charles proxy). Firefox will get a "1" the first time, and a "2" the second time. Yet it will display the "1," for some reason. Chrome exhibits this exact same behavior. I suspect it's a protocol-level issue, given the difference in Accepts header, but I've never seen this before.

    Read the article

  • Proper status codes for JSON responses to Ajax calls?

    - by anonymous coward
    My project is returning JSON to Ajax calls from the browser. I'm wondering what the proper status code is for sending back with responses to invalid (but successfully handled) data submissions. For example, jQuery has the following two particular callbacks when making Ajax requests: success: Fired when a 200/2xx status code is delivered along with the response. error: Fired when 4xx, 5xx, etc, status codes come back with the response. If a user attempts to create a new "Person" object, I send back a JSON representation of the newly created object upon success, thus giving javascript access to the necessary unique ID's for the new object, etc. This, of course, is sent with a 200 status code. If a user submits malformed or invalid data (say, an invalid/incomplete "name" field), I would like to send back the validation error messages via JSON. (I don't see why this would be a bad thing). My question is: in doing so, should I send a 200 status code, because I successfully handled their invalid data? Therefore, I'd be using the jQuery success callback, but simply check for errors... Or, should I use a 4xx status code, perhaps 'Bad Request', because the data they sent me is invalid? (and thus, use the error callback to do the necessary client-side notifications).

    Read the article

  • Can CFHEADER values be read by other code?

    - by Aidan Whitehall
    The code <cfheader name="Test" value="1"> <cfheader name="Test" value="2"> results in the header "Test: 2" being sent to the browser (as seen using HttpFox). Is there a way for the second line of code to determine if a header with the same name has already been written using CFHEADER? Thanks!

    Read the article

  • cannot modify header information puzzler

    - by outofmyleague_lucy
    hi im out of my league i have written what has now become identical code for two sites hosted on different servers. the first worked perfectly and i have therefore used it to compare the second to.from a login for i am directing to a loginaction.php when i include the db_connect.php ie <?php session_start(); include 'db_connect.php'; $user=$_POST['formUser'] $password=$_POST['formPassword']etc ? i am returning cannot modify header information header info already sent, but when i include the content of the db_connect ie <?php session_start(); $connect = mysql_connect("localhost", "user", "pass"); mysql_select_db(db_name, $connection); $user=$_POST['formUser'] $password=$_POST['formPassword']etc ? it works. any ideas? edit - requested error message Warning: Cannot modify header information - headers already sent by (output started at /home/avenncou/public_html/include/db_connect.php:4) in /home/avenncou/public_html/include/loginaction.php on line 14 at line 14 the is a header("Location: {$_SERVER["HTTP_REFERER"]}"); edit - requested db_connect.php <?php $connection = mysql_connect("localhost", "user", "pass");// or die ("Unable to connect!"); mysql_select_db("db", $connection);// or die ("Unable to select database!"); ?> that is all of it (dies commented out in case thats where the error was)!!!

    Read the article

  • Support clicking a link, but sending a POST (vs GET) to the server, without Ajax?

    - by xyld
    I'm thinking this isn't exactly possible, but maybe I'm wrong. I'm simply torn between those who believe that only POST requests should modify data on the server and people that relax the rule and allow GET requests to modify data. Take this situation. Say you have a table, each row is a row in the database. I'd like to allow them to delete the row via a fancy "X" icon as the very last <td></td> element in the row. AFAIK, the only way to send a POST to the server is via a form. But do I really stuff an entire form into the last <td></td> element just to do a POST? Or should I cheat and use an <a href=...></a> tag that sends a GET request? You may be thinking "Do both! Send a POST AND use the <a ...></a> tag! Use fancy javascript + xhr!" And I'll say, oh? And how will that degrade in a zero javascript environment? Maybe we've reached a point when it doesn't make sense to worry about gracefully degrading? I'm not sure. You tell me? I'm new to web development, but I understand most of the concepts involved.

    Read the article

  • Asking browsers to cache as aggressively as possible

    - by balpha
    This is about a web app that serves images. Since the same request will always return the same image, I want the accessing browsers to cache the images as aggressively as possible. I pretty much want to tell the browser Here's your image. Go ahead and keep it; it's really not going to change for the next couple of days. No need to come back. Really. I promise. I do, so far, set Cache-Control: public, max-age=86400 Last-Modified: (some time ago) Expires: (two days from now) and of course return a 304 not modified if the request has the appropriate If-Modified-Since header. Is there anything else I can do (or anything I should do differently) to get my message across to the browsers? The app is hosted on the Google App Engine, in case that matters.

    Read the article

  • Apache Virtual Host setup not working?

    - by user214243
    I had follow this link to setup my first virtual host with apache2 under ubuntu 12.x. Looks easy but is not working. [How to][1] My internal dns can resolv the name, but my browsers is not showing me my webpage. from my lan, Im using ubuntu server no GUI. Every time I point to my virtual host name, I receive apache default page, I cannot see my custom page. My docs are under /home/mydomain/public_html This is my settings: ServerAdmin webmaster@localhost ServerName example.com DocumentRoot /home/example.com/public_html Options FollowSymLinks AllowOverride None Options Indexes FollowSymLinks MultiViews AllowOverride None Order allow,deny allow from all The rest of the file is the same. No, I even try to change the port from *:80 to *:81, restart and don't see apache listen to port 81. My localDNS running in other machine works. What I forget?

    Read the article

  • php HTTP_REFERER header, how to turn off or leave blank

    - by eco_bach
    Hi I'm using the following simple PHP proxy script but am getting a sporadic message at the destination site. I'm thinking that perhaps it may have something to do with the HTTP_REFERER header, although I'm not explicitly defining it. Can anyone tell me how to explicitly turn off or leave the HTTP_REFERER header blank? Thanks in advance! $url = $_GET['path']; readfile($path);

    Read the article

  • x.gif in Apache logs

    - by T. Stone
    I manage a Django site where we host the media on a subdomain. There shouldn't be any requests for media to the main domain. However I keep seeing these requests for "x.gif" showing up in the access logs on the domain that's handled by WSGI (not the media domain). Can anyone explain what this is? X.X.X.X - - [08/Mar/2010:10:04:00 -1000] "GET / HTTP/1.1" 200 3724 X.X.X.X - - [08/Mar/2010:10:04:10 -1000] "GET /x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:04:10 -1000] "GET /x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:04:10 -1000] "GET /x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:04:10 -1000] "GET /x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:04:51 -1000] "GET /x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:04:52 -1000] "GET /x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:04:52 -1000] "GET /x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:04:52 -1000] "GET /x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:04:56 -1000] "GET /contact/ HTTP/1.1" 200 7196 X.X.X.X - - [08/Mar/2010:10:04:58 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:04:58 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:00 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:00 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:00 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:00 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:01 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:01 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:01 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:01 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:02 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:02 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:02 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:02 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:03 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:03 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:03 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:04 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:04 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:04 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:04 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:05 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:05 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:05 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:05 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:06 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:06 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:06 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653 X.X.X.X - - [08/Mar/2010:10:05:06 -1000] "GET /contact/x.gif HTTP/1.1" 404 2653

    Read the article

  • What are the steps to setup git-http-backend w/ Apache on Windows?

    - by Jordan
    I would like setup a Git server using the "Smart-HTTP" approach. However, I'm having difficulties getting it to work in Windows, and I'm new to Apache. My httpd.conf, in part: SetEnv GIT_PROJECT_ROOT "d:/repositories" SetEnv GIT_HTTP_EXPORT_ALL ScriptAlias /git/ "C:/Program Files/Git/libexec/git-core/git-http-backend.exe" <VirtualHost 172.16.0.5:80> <LocationMatch "^/git/.*/git-receive-pack$"> AuthType Basic AuthName "Git Access" Require group committers </LocationMatch> </VirtualHost> Could someone provide the steps to setup a Git server using git-http-backend on Windows?

    Read the article

  • How to use sudo with rcp command to copy files from linux host to HP-UX host?

    - by Justin
    I'm having this issue where when I try to use sudo to rcp some files from a Linux host to an HP-UX host (note that the destination directory requires root access to write to), I get the following error from HP-UX's side: remshd: Login incorrect. I should note that the passwords for the Linux host and the HP-UX host are different. The command doesn't seem to give me a chance to enter the proper HP-UX password and automatically defaults to this error.

    Read the article

< Previous Page | 33 34 35 36 37 38 39 40 41 42 43 44  | Next Page >