Search Results

Search found 112732 results on 4510 pages for 'google plus one'.

Page 288/4510 | < Previous Page | 284 285 286 287 288 289 290 291 292 293 294 295  | Next Page >

  • Getting a lot of '/_' errors from webmaster tools

    - by Vermino
    I'm using a WordPress site and I thought I got all the kinks out of it. For some reason Webmaster Tools is crawling my website and showing a lot of 404 errors which are from /_ like additional pages that I've never created. I just can't figure out what is creating these for Google crawlers and then displaying a 404. My robots.txt is here. My sitemap (created by the Yoast plugin) is here. I have Yoast and Jetpack plugins installed. What could be causing these links to appear

    Read the article

  • Why does the homepage rank in exact match and not in broad anymore? [closed]

    - by Claude Pelanne
    My website was ranking on page one for "walking shoes reviews" in broad and exact. It has disappeared in the rankings for broad and yet still ranks in exact match. No indication in the Webmaster tools of anything wrong yet Google analytics shows no more activity. his occurred sometime after Panda and Penguin. Not sure which because I didn't check closely for a while, this is a fun site for me. I walk for the enjoyment and health benefits so I built this site.

    Read the article

  • Installation of ANYTHING failed

    - by Nervosa
    I got an issue concerning Chrome. It launches now perfectly - still trying to install something else i see - You might want to run 'apt-get -f install' to correct these: The following packages have unmet dependencies: google-chrome-stable : Depends: lib32gcc1 (>= 1:4.1.1) but it is not installable Depends: lib32stdc++6 (>= 4.6) but it is not installable Depends: libc6-i386 (>= 2.11) but it is not installable E: Unmet dependencies. Try 'apt-get -f install' with no packages (or specify a solution). I remember i deleted some folder concerning chrome when it wasn't launchable - don't actually remember what directory exactly. Now - when i try this - "sudo apt-get -f update && sudo apt-get -f install" - i come across an error: sh: 0: getcwd() failed: No such file or directory Seems that deletion was fatal. Got any ideas? Thanks.

    Read the article

  • Help with tracking sub domain

    - by roobus
    I currently have my app's marketing/external website on the root level, e.g. http://example.com My web app itself is hosted at: http://app.example.com What's the best strategy to set-up Google Analytics tracking for both of them? Should I create a separate web property? Also, what's the difference between creating a new web property and a new profile? UPDATE: I would want to be able to track conversion from a page on the root domain to a sign-up page on the app sub-domain.

    Read the article

  • After a domain change, what can I do to recover lost traffic, rankings, impressions etc? [duplicate]

    - by Felix
    This question already has an answer here: How do I rename a domain and preserve PageRank? 3 answers I moved my site to a legacy exact-match domain I purchased about a couple of months ago. I have seen significant reduction in traffic, impressions, and rankings. I did all the right steps/best practices: change of address in GWT, map old site hierarchy and match to new site for 301 redirects etc. Indexation has gone through the Google process: old site has all but dissappeared from he index and new site is indexed, albeit with some 404 errors which I am addressing. Does anyone else who has gone through eh domain change process have any thoughts/advice? Thanks!

    Read the article

  • Does Google pick up anchor text that is in nested elements?

    - by dangerDAN
    When Google looks at anchor text on a website, will it pickup the text if it is inside nested elements? So for example: <a href="http://www.google.com/">Visit Google</a> To: <a href="http://www.google.com/"> <div class="circle"> <span>Visit Google</span> </div> </a> The reason I ask is because I want to use css3 elements for certain links on my website, to style them as circles. But the anchor text needs to be picked up for these links, so I want to know wether or not the above is bad practice in this case.

    Read the article

  • A drop in SERP after following webmaster guidelines [on hold]

    - by digiwig
    So here's a puzzle for all you SEO gurus out there. I recently launched my own site. I had target keywords which were ranking very well for about 1 month, within the top five and even appearing in first place. In an attempt to maintain good positioning, I followed guidelines by adding robots.txt, an xml sitemap redirecting non-www to www redirecting index.php to root domain adding htaccess 301 redirect for old pages I added rich snippets created a google+ account, verified my picture to appear, I went through each of the webmaster issues with duplicate titles and meta descriptions and improved header tag document outlines i even created a few more blog posts to keep the content freshing and moving. So now my website appears on page 2 with my target keywords - and all because I followed the guidelines. What is happening? I see competitors with stagnant content superglued to position 1.

    Read the article

  • Google active la réponse et l'update partielle dans son Data protocol, pour faciliter la mise à jour

    Google active la réponse partielle et l'update partielle dans son Data protocol, pour faciliter la mise à jour de ses APIs Dans le but de booster la rapidité de ses APIs, Google Data Protocol (qui permet aux développeurs d'écrire des applications liées avec les données contenues dans les produits Google) s'est vu attribué deux nouvelles fonctionnalités cette semaine, au stade expérimental : la réponse partielle, et l'update partielle. Conjointement, ces deux fonctions "peuvent significativement réduire les ressources consommées par le réseau, la mémoire et le CPU" dont on a besoin pour travailler avec les APIs de Google. Pour expliquer le rôle de la réponse partielle, l'équipe du Google Data Protocol donne ...

    Read the article

  • how can i track visits to ALL of the subpages of my website COMBINED TOGETHER?

    - by realcheesypizza
    Right now im using statcounter and google analytics. They are great. But my counts are currently separated. Ex: website.com = 1000 visits a day, website.com/about = 50 visits a day, website.com/privacy = 10 visits a day, etc.. How can have a combined count of all of my sub-pages? (mainpage + about page + about 100 other sub-pages ) I can of course manually add them all together, but that's time consuming because there are many pages. I tried placing a separate tracking code in a php include that sits in each of the sub-pages, but it doesnt seem to be working. It seems to require a single URL to create it, which it then only counts the visits from the one url, rather than ALL of them. Ex: website.com) Any help would be appreciated. Hopefully im just missing something very obvious. Thank you!

    Read the article

  • How to improve a single-paged site search result [closed]

    - by Trisism
    Possible Duplicate: How to SEO a Single-Page website I created an online CV of mine a couple of weeks ago and it has had quite a few visits. Now I want to improve the chance it will appear in google search results; however, my web CV is a one-paged site and it contains only internal links (those with hash #) so I can't really submit a sitemap. I could have changed the internal links to normal links to be processed on server-side, but there's no point of doing so. I'm very new to web SEO so I would really appreciate if somebody can show me what should I do with a single-paged site with internal links to be effectively indexed by crawlers.

    Read the article

  • How to Automate Checking for Stolen Content?

    - by Hisoka
    So I know about tools like Copyscape and Google Alerts.. great tools, but it's quite tedious for me to copy and paste an URL or phrase for every one of my pages in my sites. Is there any tool out there that monitors your website and emails you or alerts you whenever someone has stolen content from your site? The only service I know is CopySentry and honestly, it's too expensive for me since I got thousands of pages I want to monitor... Anyone else have this problem? or is it just me? Thanks for any help.

    Read the article

  • Analytics - Total events divided by number of unique pages?

    - by GeekyAndUnique
    I am using Google Analytics events to track keywords on my articles - not necessarily the best system I know but there are too many for variables I can't easily change it right now - and I would like to be able to see how popular each keyword is by dividing the number of page views with a keyword by the number of unique pages. Is there a/what is the best way of doing this? EDIT FOR CLARITY I currently have a system set up where every time somebody loads an article an event is fired for each of the tags/keywords used, with the keyword being the label. I can currently view my view count for each of the keywords by looking at the total events for each label, however I would like to be able to see which keywords are the most popular by dividing the number of times the event has been fired by the the number of different pages it has been fired from.

    Read the article

  • Which MIME type to compress? and what If I omit the `type` attribute from the HTML?

    - by rockyraw
    Per my request, my webhost had turned mod_deflate ON. In my Cpanel I now have an "Optimize Website" button. Inside that menu I could either choose: "Compress all content" or "Compress the specified MIME types" with the following default MIME types: "text/html text/plain text/xml" Which option should I choose and why? If I choose option 2, which types should I add (is there a recommended list with the exact way they should be written)? According to Google recommendations, I have omitted the type="text/css" attributes from all CSS references, as well as the type="text/javascript" attributes from all script references. Would this hinder the "gzipping" process?

    Read the article

  • Need to add 30K new pages to a 10K page website - troubles ahead? (SEO)

    - by Jurga
    We have a situation with a website where we plan to add a huge amount of new pages. The domain is over 10 years old, approximately 10 thousand indexed pages, and the planned addition is approx. 30K new pages. Any idea how we should go about it? Must we schedule a gradual data release? Have you heard of any industry standards as to how many new pages per day / week / month should be added in order to appear natural and not get in trouble with Google? I.e. should we plan a bi-weekly addition of 5K?

    Read the article

  • Google quitterait la Chine dès le 10 avril et l'annoncerait lundi, selon une source proche du dossie

    Mise à jour du 19 03 2010 par Katleen Google quitterait la Chine dès le 10 avril et l'annoncerait lundi, selon une source proche du dossier Une nouvelle source anonyme vient d'évoquer le possible départ de Google avec, cette fois, une date précise de cessation d'activités. D'après un collaborateur de la firme de Mountaiw View, "Google quitterait la Chine le 10 avril, mais Google n'a pas confirmé l'information pour le moment". Ce témoignage a été reccueilli par CBN (China Business News), et annonce une prise de parole officielle de Google pour expliquer l'organisation de ce départ dès lundi. A suivre donc. Mise ...

    Read the article

  • Google Docs : fin de l'export de documents aux « anciens » formats d'Office .doc .xls et .ppt, mais leur support sera toujours assuré

    Google Docs for Business : fin de l'export de documents aux « anciens » formats d'Office .doc, .xls, et .ppt, mais leur support sera toujours assuré par la suite Pas de nouveauté cette semaine pour les Google Apps for Business, mais une annonce qui intéressera les utilisateurs d'anciennes version de Microsoft Office (antérieures à 2007). La suite hébergée de Google ne permettra plus d'exporter des documents aux « anciens » (sic) formats que sont .doc, .xls, et .ppt. Un changement qui prendra effet dès le 1er octobre. Google précise cependant bien que Google Apps for Business continuera de supporter ces formats et qu'il sera toujours possible d'uploader ce type de doc...

    Read the article

  • Should one replace the usage addJSONData of jqGrid to the usage of setGridParam(), and trigger('relo

    - by Oleg
    Hi everybody who use jqGrid! I am a new on stackoverflow.com and it seems to me that a lot of peoples who use stackoverflow.com are not only the persons who have a problem which must be quickly solved. A lot of people read stackoverflow.com to look at well-known things from the other side. Sometime perhaps the reason is a self-training (to stay in the good form) during solving of problems other people. For all these gays, who not want only to solve his problem is my question. I wrote recently an answer to the question "jqGrid display default “loading” message when updating a table / on custom update". During writing of the answer I thought: why he uses addJSONData() function for refresh of data in the grid instead of changing URL with respect of setGridParam() and refreshing jqGrid data with respect of trigger('reloadGrid')? At the beginning I wanted to recommend using of 'reloadGrid', but after thinking about this I understood, that I am not quite sure what the best way is. At least I can't explain in two sentences why I prefer the second way. So I decide that it could be an interesting subject of a discussion. So to be exactly: We have a typical situation. We have a web page with at least one jqGrid and some other controls like combo-boxes (selects), checkboxes etc. which give user possibilities to change scope on information displayed in a jqGrid. Typically we define some event handler like jQuery("#selector").change(myRefresh).keyup(myKeyRefresh) and we need reload the jqGrid contain based on users choose. After reading and analyzing of the information from additional users input we can refresh jqGrid contain in at least two ways: Make call of $.ajax() manual and then inside of success or complete handle of $.ajax call jQuery.parseJSON() (or eval) and then call addJSONData function of jqGrid. I found a lot of examples on stackoverflow.com who use addJSONData. Update url of jqGrid based on users input, reset current page number to 1 and optionally change the caption of the grid. All these can be done with respect of setGridParam(), and optionally setCaption() jqGrid methods. At the end one call trigger('reloadGrid') method of the grid. To construct the url, by the way I use mostly jQuery.param function to be sure, that I all url parameters packed correctly with respect of encodeURIComponent. I want that we discuss advantages and disadvantages of both ways. I use currently the second way, so I start with advantages of this one. One can say me: I call existing Web Service, convert received data to the jqGrid format and call addJSONData. This is the reason why I use addJSONData method! OK, I choose another way. jqGrid can make a call of the Web Service directly and fill results inside of grid. There are a lot of jqGrid options, which allow you to customize this process. First of all, one can delete or rename any standard parameter sent to server with respect of prmNames option of jqGrid or add any more additional parameters with respect of postData option (see http://www.trirand.com/jqgridwiki/doku.php?id=wiki:options). One can modify all constructed parameters immediately before jqGrid makes corresponding $.ajax request by defining of serializeGridData() function (one more option of jqGrid). More than that, one can change every $.ajax parameter by setting ajaxGridOptions option of jqGrid. I use ajaxGridOptions: {contentType: "application/json"} for example as a general setting of $.jgrid.defaults. The ajaxGridOptions option is very powerful. With respect of ajaxGridOptions option one can redefine any parameter of $.ajax request sending by jqGrid, like error, complete and beforeSend events. I see potentially interesting to define dataFilter event to be able makes any modification of the row data responded from the server. One more argument for using of trigger('reloadGrid') way is blocking of jqGrid during ajax request processing. Mostly I use parameter loadui: 'block' to block jqGrid during JSON request sending to server. With respect of jQuery blockUI plugin http://malsup.com/jquery/block/ one can block more parts of web page as the grid only. To do this one can call jQuery('#main').block({ message: '<h1>Die Daten werden vom Server geladen...</h1>' }); before calling of trigger('reloadGrid') method and jQuery('#main').unblock() inside of loadComplete and loadError functions. The loadui option could be set to 'disable' in this case. So I don’t see why the function addJSONData() should be used. Can somebody who use addJSONData() function explain me advantages of its usage? Should one replace the usage addJSONData of jqGrid to the usage of setGridParam(), and trigger('reloadGrid')? I am opened to the discussion.

    Read the article

  • Send email from server to Google Apps email address (same domains)

    - by Orlando
    I'm sending email from a server, let's say domain.com. I also have Google Apps email set up for hosted email, same domain, domain.com. If I get mail sent to me from anywhere else, I receive things just fine. However, if the email originates from my server, it just ends up in /var/mail/root as a delivery error saying the user is unknown. I created a user on the server for the name which is having trouble, [email protected]. Retried sending and it sends, but not to my hosted email at Google Apps. I just receive it at /var/mail/webmaster now. I'm using sendmail. I messed around with /etc/aliases but adding webmaster: [email protected] looked useless (and I was right.) Any help?

    Read the article

  • Google images sometimes terribly slow when using dnsmasq

    - by Joril
    Hi everyone! I am the admin of a small LAN of 10+ computers. I've set up a dnsmasq server for DHCP and DNS resolution, and it's working almost fine.. My problem is that when I try to use Google images, sometimes it takes ages to show the actual images. I get just the textual part of the page (menus and so on) while the images themselves are shown as the still-loading-white boxes.. When I use the DSL router directly as DNS, the site works fine all the time. The problem sometimes presents itself with Google maps too.. The map takes ages to load. Any idea on what I could try to troubleshoot this? (dnsmasq 2.47 on CentOS 5.2 64bit, our outside connection is an asymmetrical 4Mbps DSL)

    Read the article

  • Google images sometimes terribly slow when using dnsmasq

    - by Joril
    Hi everyone! I am the admin of a small LAN of 10+ computers. I've set up a dnsmasq server for DHCP and DNS resolution, and it's working almost fine.. My problem is that when I try to use Google images, sometimes it takes ages to show the actual images. I get just the textual part of the page (menus and so on) while the images themselves are shown as the still-loading-white boxes.. When I use the DSL router directly as DNS, the site works fine all the time. The problem sometimes presents itself with Google maps too.. The map takes ages to load. Any idea on what I could try to troubleshoot this? (dnsmasq 2.47 on CentOS 5.2 64bit, our outside connection is an asymmetrical 4Mbps DSL)

    Read the article

  • SunOne case-insensitive URLs

    - by RoToRa
    It it possible to configure a SunOne web server to automatically redirect all URLs with capital letters to the corresponding lower case URLs? For example, redirect /Example, /eXamPle and /EXAMPLE all to /example. This would have to be for all URLs (or at least a subset excluding a specific prefix) I normally have nothing to do with web server configuration (especially not SunOne). I just need to now if it is generally possible and be pointed to the right direction on how to do this. Thanks.

    Read the article

  • Google Mail: Import labels when switching from POP3 to IMAP

    - by toobb
    My situation is the following: I have a Google Mail email-address and I have been using this address in Thunderbird fetching the emails with POP3 (the emails also remained on the server and where "archived"). In Thunderbird I organized my emails in folders. Now I want to switch to IMAP (with the same email address), but I want to keep the folder structure I created in Thunderbird. I could create a 2nd account in Thunderbird that uses IMAP, and then move my folders from the old account to the new one. But the problem is, that Google Mail probably does not recognize that it already has these moved emails in "All Mails". I will probably end up with two copies of every email. Does someone has good idea how to deal with that problem? Thank you!

    Read the article

  • SPF problems with Google Apps

    - by mahle
    I currently have an SPF record with a hostname of @ that is: v=spf1 mx ip4:x.x.x.243/32 include:_spf.google.com include:amazonses.com ~all I also have another record of" spf2.0/pra mx ip4:x.x.x.243/32 include:_spf.google.com include:amazonses.com ~all We have had a lot of email being bounced back because of spam and now when I go to http://www.kitterman.com/spf/validate.html? and check the "Does my domain already have an SPF record? What is it? Is it valid?" it says no spf record exists. However, when I send an email using our Amazon SES script and check the headers it says it passes the SPF test. Is there something I am missing? Do I need to place that text in quotes ""? Any help would be greatly apprecaited.

    Read the article

  • Anyone know where I can download a copy of Sun Java System Active Server Pages 4.0.3 for Solaris

    - by ewengcameron
    I've contacted Sun regarding this and they have told me that the download is no longer available as Active Server Pages 4.0.3 is now End Of Life. We need to upgrade our server to 4.0.3 to acheive PCI-DSS compliance. Anyone know of a site where I can download older copies of Sun files? Sun offer 4.0.1 and 4.0.2 to download but not 4.0.3 which is going to cause problems come October when Visa stops accepting transactions from non PCI compliant servers. If Sun kept their naming system consistent across versions, the file would be called "sjsasp403-sol-sparc.tar". I know the real solution is to upgrade every site on the server to use a different server language, i.e. PHP, and in the long term, this is our goal but we have over 100 sites requiring upgrading and its not a viable solution to get this done before October.

    Read the article

< Previous Page | 284 285 286 287 288 289 290 291 292 293 294 295  | Next Page >