Search Results

Search found 21508 results on 861 pages for 'google datastore'.

Page 254/861 | < Previous Page | 250 251 252 253 254 255 256 257 258 259 260 261  | Next Page >

  • SEO and external sites that serve responsive images (like Re-SRC)

    - by Baumr
    Re-SRC is a tool that allows you to automatically serve responsive images for your website from their cloud servers. It delivers a new image file each time the browser window (viewport) is resized. To use it in your HTML when linking to an image, you would do the following: <img src="http://app.resrc.it//www.your-domain.com/img/img001.jpg"/> Some more background for SEO considerations: As an example, looking at their demo page's code, the src of the Arc de Triomphe photo — when the browser window is resized to be at a tablet-width — shows this particular file at it's widest. It is found under the following URL: http://app4-uk.resrc.it/s=w560,pd1/ro=h//www.resrc.it/img/demo/demo-image-1.jpg If the viewport is increased to desktop-width, then a smaller image is served in line with the design; see this URL: http://app4-uk.resrc.it/s=w320,pd1/ro=h//www.resrc.it/img/demo/demo-image-1.jpg If I change the viewport to be about half-way between those two, then the image's URL is: http://app4-uk.resrc.it/s=w240,pd1/ro=h//www.resrc.it/img/demo/demo-image-1.jpg In other words, I found that there is a separate file for every 10-pixel increment of the image width. Very cool for saving bandwidth on mobile devices and service responsive/retina images on others, but... Here are two problems I see for SEO: The img on your site, part of your semantic markup, will not be hosted on your site at all, or even a server you control. Any links to these images will pass on "link juice" to Re-SRC's site instead. You are serving a vast array of different image files to different people — some may link to one, others to another size. Then there's the question of what different search engine crawlers will see. Also: There seems to be no fallback option if their servers are down. Do you see any other concerns? Or, perhaps, do you not see those as concerns?

    Read the article

  • Huge surge in direct traffic from one particular town

    - by Jack Lockyer
    Last month I noticed that the direct visits on our site have increased by nearly 150% whilst bounce rate is also considerably up. After drilling down further I can see that we have had nearly 2000 direct visits from one town in Connecticut called Stamford, with a bounce rate of 100%! I have been scratching around for answers but I can only find that it may be to do with our uptime monitoring tool; Pingdom. Does anyone know/have any experience with this kind of issue, any help is appreciated I have just noticed that we are receiving identical traffic in a town in England and a town in Scotland... This definitely makes me think it's to do with our uptime monitoring tool.

    Read the article

  • visit counts in advanced segments not consistant

    - by user671201
    My organization has recently noticed an issue when applying advanced segments to visit counts during different time ranges. With no advanced segments turned on, here are the visit counts for Oct 1st - Oct 4th during the time range Sept 8th - Oct 8th: Oct 1 - 7 Oct 2 - 7 Oct 3 - 8 Oct 4 - 5 Again, with no advanced segments turned on, here are the visit counts for Oct 1st - Oct 4th but I've changed the time range to Oct 1st - Oct 4th. As expected, the numbers are the exact same as above: Oct 1 - 7 Oct 2 - 7 Oct 3 - 8 Oct 4 - 5 Now, I turn on the "Non paid search traffic" advanced segment. Here are the visit counts for Oct 1st - Oct 4th during the time range Sept 8th - Oct 8th: Oct 1 - 0 Oct 2 - 0 Oct 3 - 0 Oct 4 - 2 Here is where it gets weird. I keep the advanced segment on, and change the time range to Oct 1st - Oct 4th. This is what I get for the exact same dates as above: Oct 1 - 4 Oct 2 - 2 Oct 3 - 6 Oct 4 - 5 We've found the same inconsistency in our other GA profiles that get much more traffic (the above numbers come from one of our specialized topic blogs), but the inconsistency is less pronounced where there are more visits. My question is: why are the visit counts different for different time ranges when advanced segments are turned on, but exactly the same when no advanced segments are applied? Is this a GA bug or am I missing something about how the advanced segments work?

    Read the article

  • Slow Resume After suspend having chrome running with many tabs

    - by tUrGoNn
    I usually use chrome and Firefox for browsing. I also open many tabs (around 40 in both some times). The problem I have occurs when I resume the PC after having suspended it: It takes from 2 to 5 minutes sometimes to just get back normally. Does this have to do with memory usage not properly resuming? Is it a bug in Chrome/Firefox or Ubuntu itself? Note that I just upgraded from 10.10 to 11.10 and I was having the problem on both releases, which makes me guess that it has to do with Ubuntu not resuming well if some memory-heavy apps were running before the suspend occured.

    Read the article

  • Build My Own Advertising Network

    - by clifgray
    I have a few ideas that I think would be pretty game changing for online advertising and I would like to build my own network but I don't know where to start. I know it will take a lot of time for major publishers to get on board but I am more curious about the technical side. What language/database model and framework are modern ad networks built on? Basically I want to build an advertising network that registers views per page and allows publishers to manage the look of their own ads and let's the users interact with the ads. Is there any good information on doing something like this or any framework you can suggest to build on? I know this would get complicated pretty fast so if you have suggestions for ad networks that let you customize them heavily I would be glad to hear your suggestions.

    Read the article

  • Chrome keeps crashing after updates

    - by Phil
    I'm a chrome user and an enthusiast but few days ago I made Ubuntu get some updates using Update Manager, then I turned off my notebook and when I restarted it, I tried to start Chrome which after a few seconds turned the screen completely black, and some errors appeared for few seconds and then I got logged off, and had to re-login, and when I opened Chrome again the same thing happened! I've uninstalled it with Synaptic and did complete cleanings before re-installing Chrome but nothing succeeded. Now I can't use either Chromium and Chrome, but Firefox works. It's a very strange thing, never happened to me before and I don't know what to do because I had all my bookmarks syncronized in Chrome! Please help if you can :)

    Read the article

  • pop up html as javascript string instead of hidden div for seo [closed]

    - by user1324762
    Possible Duplicate: How bad is it to use display: none in CSS? I have heard that using display:none or visibility:hidden css properties are not a very good idea for seo purposes. I have about 4 different pop up windows to display and each one has about 20 words inside it. I can create hidden divs. Another option is to store div html elements as javascript string. In this way pop up html elements will be generated from javascript string. This will be still faster than using ajax since the data is static. Is this method absolutely safe for SEO? P.S.: I was just asking about similar question on http://stackoverflow.com/questions/12389075/storing-data-in-javascript-array-for-further-use, but this one is different, it is about static data and about SEO.

    Read the article

  • How do I remove these errors from my blog so as to get adsense approved?

    - by Serenity
    This is the question I asked on SO site earlier, but didn't get satisfactory replies. hoping to find a solution here.. http://stackoverflow.com/questions/12136796/how-can-i-detect-and-correct-these-errors-on-my-blog/12136829#comment16235061_12136829 In web master tools, apart from the errors in the question link above, it is showing a site map error too as in the screenshot below:- Need guidance please...thanks :) Edit -1 EDIT 2 I had 2 SEO plugins on my blog and I would put meta description for each of my article in both plugins that are All in One SEO and Yoast's "Wordpress SEO". Now I removed all article's meta descriptions from "All in one SEO" the other day but STILL web master tool is showing duplicate meta tags and descriptions. Why??

    Read the article

  • Moving from root to www

    - by chris
    I've read tons of questions but I can't find the answer to this seemingly obvious issue. I'm moving my WordPress site from domain.com to www.domain.com so that I can use CloudFlare. I did the change in WP and everything works fine, with domain.com redirecting correctly. Do I need to add the www site in GWT (and remove domain.com) or will it keep tracking the website correctly thanks to the redirections?

    Read the article

  • Chrome not accepting international dead keys 14.04

    - by D3L
    Every other application on 14.04 accepts that I have selected US international with dead keys as my keyboard layout option, and accepts text input as it should. Chrome however fails to recognise what keyboard I have set in system settings and blindly uses "US keyboard". Looking for a solution to force Chrome to accept dead key input. AFAIK it used to work, but something has messed up recently with updates to Chrome

    Read the article

  • Re-indexing website with clean URL's

    - by artsi
    So I have a website with URL's like this: http://www.domain.com/profile.php?id=151 I've now cleaned them up with mod_rewrite into this: http://www.domain.com/profile/firstname-lastname/151 I've fetched and re-indexed my website after the change. What is the best way to make the old dirty ones disappear from search results and keep the clean ones? Is blocking profile.php with robots.txt enough?

    Read the article

  • 301 redirect to 404 page?

    - by Kristian
    Currently i'm migrating www. prefix from my urls and use htaccess to do the job. Since we have new software and cleaned database some of the old urls doesnt exists anymore. Therefore some requests redirect to 404 page. 1. www.domain.com/old-page # htaccess redirect to non-www url, 301 2. domain.com/old-page # page does not exists, 404 Does this method have any SEO issues, or even affect pagerank? Or should i check the page existence before redirecting and show 404 without redirect?

    Read the article

  • Use of another country domain name can influence search engines results?

    - by DontVoteMeDown
    I'm studing a way to create my company domain based on it's name. Consider that my company's name is Another Store and I want to register a domain like anothersto.re - this is just an example. That domain is strictly chosen by marketing. What happens is that my company is stabilished in Brazil and our domain here is .br. The .re domain stands for an island near France so haves nothing to do with my country. If that domain is chosen what it can imply about SEO questions? Did it will have any influence on search engines results considering that they look over user's region? This kind of domain use became common between modern companies - and marketing strategies - and that is why I'm considering it.

    Read the article

  • Custom trackpad mapping doesn't work for all applications

    - by picheto
    I found out I could invert my trackpad scrolling, so as to work more like the OS X "natural scrolling", which I liked better. To do that, I run the following command on startup: xinput set-button-map 11 1 2 3 5 4 7 6 Where 11 is the touchpad id (found with xinput list and xinput test 11). This inverts the vertical and horizontal two-finger scrolling, and works fine in Terminal, Chrome, Document Viewer, etc. However, it doesn't work in Nautilus and some applications such as the Update Manager, as they keep the usual mapping. I'm running Ubuntu 12.04 x64 Why does this mapping work for some applications but not for others? I know there is software I can download to do the same, but this method seemed "cleaner". Thanks

    Read the article

  • Google propose une version professionnelle de son App Engine et des services en Cloud privé grâce à

    Google propose une version professionnelle de son App Engine Et ses services en Cloud privé grâce à une collaboration avec VMWare D'une pierre deux coups ? comme on dit. Lors du Google I/O, VMWare et Google ont annoncé un partenariat qui risque de faire du bruit dans le monde du développement web. Pour tout dire, il en fait déjà beaucoup. Cette collaboration va en effet permettre à Google d'étoffer encore un peu plus son offre de services Cloud spécialement dédiée aux développeurs d'entreprise. Jusqu'ici Moutain View proposait ses produits en mode « Cloud public », autrement dit hébergés sur ses propres serveurs. Mais de nombreus...

    Read the article

  • Quel avenir pour la boite à outils graphique GWT avec le désengagement progressif de Google ? Venez vous exprimer

    Lors du dernier Google IO (https://docs.google.com/presentation...1#slide=id.p18), le leader technique de GWT Ray Cromwell a annoncé le début du désengagement progressif de Google dans le développement de la boîte à outils graphique. Ce désengagement n'est pas surprenant puisque de nombreuses lacunes de la boîte à outils obligent les développeurs à compléter leur code avec des projets communautaires (Cf sondage sur les bibliothèques de composants graphiques). C'est sûremen...

    Read the article

  • Keyword research tool [closed]

    - by sam
    Possible Duplicate: What are some great tools to use for keyword research? in the past ive always choosen which keywords / phrases to go after by checking the traffic in the keyword tool, checking the competions for that keywords, their backlinks and pr and then by reading though all that and going on gut insticnt. I know seomoz has some functionality for this but are there any alternatives you could recomend ?

    Read the article

  • How to conduct A/B split testing with AdSense?

    - by None
    Ok so I have decided to A/B test my AdSense ads. I have run a few tests, but I don't know what conclusion to draw and how to keep track of things. Some specific questions: If I have 2 test units, 1 wins. I test that with a new and so on. How do I find if say the fifth one did better than the first one? How do I keep track of things? Do I let the variables independent of each other, because they certainly are not. In real life, font size can affect CTR even if the colors are different. I can test blue color with red color, and then test Arial font with Georgia, but how do I know which combination is the best? This would result in way too many test units. I tried Googling a lot, but I could not find answers to these questions.

    Read the article

  • "Network is unreachable" When pinging google, can connect to internal computers on debian VM

    - by musher
    Similar to this SU question: "Network is unreachable" when attempting to ping google, but internal addresses work Actually, it's pretty much the same base issue. I went through that thread trying to find a solution, I changed my resolv.conf: before: domain [my work domain] search [my work domain] nameserver [my gateway] nameserver [my gateway2] I changed it to: after: domain [my work domain] search [my work domain] nameserver 8.8.8.8 nameserver 8.8.4.4 However, any time I reboot the computer the resolv.conf gets overwritten to the previous version (the 'before' above). The issues began after I installed virtualbox additions, X server and (specifically) LXDE: Cat of apt history.log: Start-Date: 2014-08-21 10:03:42 Commandline: apt-get install virtualbox-guest-utils virtualbox-guest-dkms Install: x11-xkb-utils:amd64 (7.7+1, automatic), libxaw7:amd64 (1.0.12-2, automatic), xfonts-utils:$ End-Date: 2014-08-21 10:03:56 Start-Date: 2014-08-21 10:18:39 Commandline: apt-get install lxde Install: desktop-base:amd64 (7.0.3, automatic), libgoa-1.0-0b:amd64 (3.12.4-1, automatic), lxmenu-d$ End-Date: 2014-08-21 10:21:52 Start-Date: 2014-08-21 10:26:40 Commandline: apt-get upgrade Upgrade: libio-socket-ssl-perl:am ifconfig on the guest: root@Peridot:~# ifconfig eth0 Link encap:Ethernet HWaddr 08:00:27:89:c9:20 og inet addr:172.31.2.102 Bcast:172.31.2.255 Mask:255.255.255.0 inet6 addr: fe80::a00:27ff:fe89:c920/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:2281 errors:0 dropped:1 overruns:0 frame:0 TX packets:463 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:266507 (260.2 KiB) TX bytes:120554 (117.7 KiB) lo Link encap:Local Loopback inet addr:127.0.0.1 Mask:255.0.0.0 inet6 addr: ::1/128 Scope:Host UP LOOPBACK RUNNING MTU:65536 Metric:1 RX packets:4 errors:0 dropped:0 overruns:0 frame:0 TX packets:4 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:0 RX bytes:240 (240.0 B) TX bytes:240 (240.0 B) The adapter in VBox is a bridged adapter directly onto my ethernet connection; as are my other 2 VMs (which work) Other SU questions I've tried: "connect: Network is unreachable" in VirtualBox VM

    Read the article

  • AdWords : Google s'attaque aux publicités douteuses avec une politique de contrôle et de modération contre les annonces frauduleuses

    AdWords : Google s'attaque aux publicités douteuses Et met en place une politique de contrôle et de modération pour traquer les annonces frauduleuses Via un post publié sur son blog officiel, Google a révélé avoir amélioré sa politique de contrôle et de modération pour lutter contre toutes les annonces AdWords douteuses (vente de produits illicites et nocifs, annonces trompeuses,...). Google AdWords permet aux entreprises de créer leurs annonces publicitaires et de les diffuser sur Google. L'entreprise doit créer une annonce et choisir des mots ou des expressions clés en rapport avec son activité. L'internaute, lorsqu'il effectue une recherche à l'aide de l'un de ces mots clés, voit s'af...

    Read the article

  • Référencement : Google déploie un nouvel algorithme qui pénalise les sites utilisant des noms de domaine avec des mots clés

    Référencement : Google déploie un nouvel algorithme qui pénalise les sites utilisant des noms de domaine avec des mots clés Via un message sur son compte Twitter, le spécialiste de l'équipe anti WebSpam de Google, Matt Cutts, a annoncé le déploiement d'un nouvel algorithme sur le moteur Google. Ce nouvel algorithme vise à affiner les résultats de recherche du moteur Google en pénalisant les sites ayant un contenu de faible qualité malgré que le nom de domaine du site corresponde aux termes de la recherche. En effet, la firme a constaté que des sites Web ont recourt à des noms de domaines qualifiés de « exact-match domain » qui reprennent des termes de recherche susceptibles...

    Read the article

  • Is it possible to have an app in the Play Store that will modify/change the behavior of the Gmail for Android application?

    - by Benjamin Bakhshi
    For example, is it possible for Rapportive (which works on Gmail for web), to work on Gmail for Android. I understand that the UI would be vastly different, but the question remains, is it possible at all to overlay content or change behavior of the Gmail app for android. I have done some research but cannot find the right resources that would tell me if this is possible or not on Android devices. This is a trivial question for Gmail for Web, ie. making a Chrome extension has lots of resources, and services like Rapportive manipulate Gmail apparent ease.

    Read the article

  • will main domains be more seo friendly than subdomains?

    - by C graphics
    Web hosting providers offer services such as hosting multiple domains in one account. Then my concern is about seo friendliness. say the main domain of my account is maindomain.com on which I have added an addon domain say domain2.com. That means cpanel will generate domain2.maindomain.com and the contents of domain2.com will be practically stored into a subfolder in maindomain.com. Now, assume both maindomain.com and domain2.com have same structure both optimized for seo same way. My question is that would maindomain.com links be more seo friendly due to that fact that maindomain.com is the mani domain of my account?

    Read the article

  • Troisième trimestre au delà des prévisions chez Google, son action dépasse les 1000 dollars pour la première fois de son histoire

    Troisième trimestre au delà des prévisions chez Google, son action dépasse les 1000 dollars pour la première fois de son histoire Google a publié des résultats supérieurs à ses prévisions pour son troisième trimestre. Pour la première fois de son histoire, l'action Google est passé au-dessus des 1000 dollars ce vendredi 18 octobre. Rappelons qu'en 2004, l'action valait 100 dollars. De plus, son bénéfice net a augmenté de 36% et la firme enregistre 2,97 milliards de dollars contre 2,18 milliards...

    Read the article

  • Hiding a particulat page from search engines not to index

    - by user702325
    I have a page which i don't want search engines to index or crawl. I am not sure hat should i put in my robots.txt file to tell search engines not to crawl/index that page. The page it itself is getting generated dynamically and do not have a predefined template for it all i know about its URL which is pre-defined and will remain unchanged. I have this page say at www.mysite.com/my-nonindexable-page/ Please suggest what i should do to achieve this.I am using WordPress for my website

    Read the article

< Previous Page | 250 251 252 253 254 255 256 257 258 259 260 261  | Next Page >