Search Results

Search found 31918 results on 1277 pages for 'google maps api'.

Page 147/1277 | < Previous Page | 143 144 145 146 147 148 149 150 151 152 153 154  | Next Page >

  • Should I index my mobile duplicate of my desktop website on google?

    - by Roy
    I have a duplicate of http://he.thenamestork.com with the url http://he.thenamestork.com/mobile - all files are duplicated while the mobile version has a slightly different content. Notice that when I write 'mobile' I only mean regular HTML4 with smartphone friendly CSS. I have a series of redirects (using .htaccess) that allows smartphone users land directly on the mobile versions. But I wonder, shound I index the mobile version as well so those users will be able to get direct, faster links? And what is the proper way of doing that without causing problem in google search? I guess I'm asking if there's a way to get google display regular urls for desktop users and ../mobile/.. urls for smartphone users, and if it is smart SEOwise.

    Read the article

  • How to to let Google know about dynamic content?

    - by Yaniv
    Im looking for the best practice to let Google know about a vast number of dynamically created content. Let's say (I mean - dream) that I'm Facebook, and I want to let Google to index all the users' posts. Sitemap.xml may be the answer for this but they are limited to 50,000 URLs in each site map. I know that I can create 500 sitemaps and create a sitemap for sitemaps, but they are also limited, 25,000,000 URLS sounds quite enough at the moment, but could cause problems in the future. I.E - stackoverflow already has 3 Million posts, probably sitemap is not the solution for them. Creating a page with paging, and links to all the dynamic data. i guess this is what stackoverflow did by creating this page here: http://stackoverflow.com/questions So I think that Option 2 is the answer, but it seems to me that sitemaps might have some added value. So what should i do?

    Read the article

  • Google+ Platform Office Hours for May 2nd, 2012: Hanging out with the Tabletop Forge team

    Google+ Platform Office Hours for May 2nd, 2012: Hanging out with the Tabletop Forge team This week we met with Charles Jaimet, Joshuha Owen and Fraser Cain of the Tabletop Forge team. They showed us their hangout app and shared their experience. Discussion this session on Google+: goo.gl You can learn more about our office hours here: goo.gl Here are some notable moments in this session: 1:50 - Charles explains Tabletop Forge 6:12 - Tabletop Forge Demo begins 7:45 - How do you prevent cheating on dice rolls? 14:07 - A discussion about trust in tabletop gaming 14:57 - Upcoming feature - Fog of war 24:06 - What are some challenges with the Hangouts API that you've overcome? 27:10 - It'd be cool to play a game with a separate on air game view 31:08 - Comments as a source of game material 31:58 - What else is on the roadmap for Tabletop Forge? 35:52 - Will there be a Kickstarter for Tabletop Forge? 36:42 - What do you think about saving game logs to places like Google Drive or Google Docs? 39:07 - The 7 sided die is not something possible in reality. In what other ways have you gone beyond the limits of the physical table top? 43:11 - What was your first game? From: GoogleDevelopers Views: 2401 23 ratings Time: 46:43 More in Science & Technology

    Read the article

  • Why Wikipedia doesn't appear as a referral in Google Analytics' Traffic sources?

    - by Rober
    One of my clients has a website and got not spammy backlinks in a Wikipedia article. When I test it for SEO purposes with Google Analytics (from different IPs), apparently there is no referral information. On the Real-Time view my test visit is visible but with There is no data for this view in the referrals subview. And this visits appear as (direct) / (none) on the Traffic sources view. Wikipedia is not hiding in any way its links origin, since it is shown in the server visits log. Is Google ignoring Wikipedia as a referral? Am I missing anything else? Update: Now it works, several days after the link was active. Maybe something is detecting for how long the link was there so that it doesn't work just from the beggining, as a security measure? Many visits are actually not recorded.

    Read the article

  • Does a lead-to screen with AdSense ad conform to Google's rules?

    - by ElHaix
    Re: Google's ad placement policy I have noticed that when clicking on some Forbes links, I am taken to a screen with an ad in the middle - at the top there is a link to skip the ad. Upon clicking on the skip link I am taken to the article I want to view. I want to implement something similar on my sites, where, when clicking on a search result, a the results window first displays one AdSense add on the screen with a similar UI as what I saw on Forbes. Currently, when a user clicks on a result, a new tab/window opens with the result. What I am proposing is that before the result appears, the screen displays a "Continue to result" link at the top in large letters, and in the center of the page, "Advertisement" with the ad below. This is the only popup that is user initiated and there are no other popups on the site. Navigation elements are not modified in any way. Will I get penalized by Google for implementing this?

    Read the article

  • YouTube API @ Google I/O 2010

    Kuan Yong, Gareth McSorley and I -- representing Product Management, Engineering, and Developer Relations, respectively -- were happy to present a YouTube API session at this year’s Google...

    Read the article

  • What measures can be taken to make sure Google is aware of the existence of a newly created page?

    - by knorv
    Consider a website with a large number of pages. New pages are published regularly. When publishing a new page the website operator wants to get the newly created paged indexed in Google as soon as possible. The website operator wants to minimize the time spent between publication and indexing. Consider the site http://www.example.com/ with hundreds of thousands of pages. The page page http://www.example.com/something/important-page.html is created at say 12:00. I want to get important-page.html indexed as soon as possible after 12:00. Ideally within seconds or minutes. What options are available to try to get Google to index a specific newly created page as soon as possible?

    Read the article

  • How are undocumented functions discovered?

    - by Paska
    Hi all, I've always wondered how do you find undocumented / private API ? Example the Apple undocumented / private API, Play Station, Windows Phone 7, Win32 Kernel, Windows API, hidden callback, etc... What tools do hackers use to find out about private and undocumented functions? Where can I read about peoples' experiences delving into private APIs and reverse engineering techniques which reveal the secrets which are normally explained in API documentation? thanks, A

    Read the article

  • Is Java "dead in the water" as a consequence of Oracle buying Sun and subsequently suing Google

    - by NimChimpsky
    Google has led many useful Java features (guava, gson); now that Oracle has purchased Sun will it effect its future enhancements and utilization as a development language. What exactly, or even approximately, are the legal ramifications? I thought Java was open source and would therefore remain unaffected ... ? Should Google just buy Oracle to get rid of the whole mess, that would be cool wouldn't it ? Do you think this is the beginning of the end for Java as a widely used language ? Its continued success as an open source/free technology is now doubtful?

    Read the article

  • How can I make Google show unit conversions by default?

    - by bUbUKid
    When I search for "4 inches in g" on my Windows Firefox I immediately get a unit conversion done by Google that shows up before the actual search results. On my Ubuntu 12.04 system this does not work though. I tried Firefox and Chromium and have no script blockers installed. I also switched off AdBlock Plus for testing but to no avail. I realize that this is not really Ubuntu doing something wrong but: Are there any settings I can modify to make Google show these results? I use them quite frequently and I believe (though I cannot test it anymore) that this used to work on my last Ubuntu System. Maybe there are some script sources that Ubuntu has disabled by default or something like that?

    Read the article

  • Don’t Like New Google Search Interface? Switch To Old Interface

    - by Gopinath
    Google recently adopted a new user interface layout for it’s search engine. The new layout is very different to classic one, it provides many options on the left side to choose for enhanced search operations. Even though many users like this new interface, there are few who are more comfortable with the classic interface. If you are one among those who wanted to switch back to classic interface, you can access it by using the following http://www.google.com/webhp?hl=all Join us on Facebook to read all our stories right inside your Facebook news feed.

    Read the article

  • Does Google counts backlinks from homepage to inside pages?

    - by SharkTheDark
    I have a site with good PR and my inside pages are getting increase of PR, but they don't have links pointing to them, only from my homepage. Does that means that Google counts ALL links on my homepage, including links to inside pages? Does it calculate inside pages PR with one coming from my domain, my homepage, too? Also, if inside pages that got high PR from homepage have link back to homepage, will that increase homepage PR additionally, since those links should count too? By Google PR algorithm formula, by calculations on Wikipedia and Stanford PR algorithm explanation ( which is originally developed by ) it counts those links, and also it counts after-increase backlink again, making few times circle ( it stops because of d ( 0.85 ) factor. ), but it counts them. Does anyone know is this correct?

    Read the article

  • Tracking logged in vs. non-logged in users in Google Analytics

    - by Justin
    I am building a social media site that is similar is structure to twitter and facebook.com where unauthenticated users who go to https://mysite.com will see a login + sign-up page, and authenticated users who go to https://mysite.com will see their timeline. My question is, what is the best practice (using Google Analytics) for tracking these two different types of users who are viewing completely different content but are visiting the same URL. I tried searching the Google Analytics docs but couldn't find what they suggested for this scenario. Perhaps I just don't know what keywords to search for. Thanks in advance for any help.

    Read the article

  • How to recover organic position in Google results after server down?

    - by ElHaix
    I have several sites that were doing quite well in terms of organic SEO rankings. I have the important sites setup in Google's Webmaster tools. Long story short, the system was down for about two weeks. Now in AdSense and Analytics, I am seeing that the page views are SLOWLY increasing. and I would like to know if there is anything I can do now to try to expedite the process of regaining those positions. Since there were several errors from that server, is it possible that Google will now rank any site from that IP address lower due to those two weeks of errors? Is this something that I just have to let ride out? Thanks.

    Read the article

  • How to interpret Google's "Avg. Page Load Time"?

    - by hawbsl
    Is there any industry rule of thumb for what's considered an unacceptable load time v. an OK one v. a blistering fast one? We're just reviewing some Google Analytics data and getting 0.74 Avg. Page Load Time reported. I guess that's OK. However it would be good if some meatier comparison data were available, or a blog post, or somewhere where there's some analysis of what speeds are generally being achieved by various kinds of sites. Any useful links to help someone interpret these speeds? If you Google it you just get a lot of results dealing with how to improve your speed. We're not at that stage yet.

    Read the article

  • Which one of the Google Analytics stats should be my top priority to improve SEO: visits, pageviews, bounce rate, new visitors, or visit duration

    - by HOY
    My site is getting a lot of traffic from an image (a company logo image) because this image is ranked 1st in Google search results for a company's title. (I have no idea how that happened.) This image is must for my website, but it is not relevant to the site content, so irrelevant people search for the image and find out about my site. I get interesting statistics: Pros: Total Visits & Avg. New Visits Cons: Avg. Page/Visit, Avg. Visit Duration, Bounce Rate I am confused if this image is helpful to my website. I don't know what the balance between those 5 statistics should be. My website is 2 months old, and we are working on SEO at the moment. Edit: here are the details about the image: Search Keyword: arcelik logo Search Site: google.com.tr Search URL

    Read the article

  • How to redirect a international domain to a subfolder on the English site without hurting Google rankings?

    - by ernest1a
    I have two sites: www.example.de www.main.com www.main.com is English version of www.example.de which is in German. I want to keep only www.main.com. For the English version I will keep www.main.com, but for German I want to move it to www.main.com/de. I am wondering what would be best solution for old www.example.de: Redirect everything from www.example.de to www.main.com/de using 301 redirect? Redirect everything from www.example.de towww.main.com/de/page-url-of-old-size.html? So each link actually get own address. Is that necessary or will Google realize where the page belongs on new site even if I redirect everything to home page? Any other solution, maybe just set in Google webmaster tools the new domain or anything like that?

    Read the article

  • When Google gives up recrawling 301 that led to 404?

    - by Easy Life
    I've transferred a domain and made a mistake in the redirects (the URL structure is identical). Even though they went to the new domain, the error caused a 404 when crawled by Google bot. 10 days after I saw and corrected my redirect mistake, and now the site should (hopefully) redirect to proper pages. Q1: The URLs of the 404 pages in the Webmaster Tools all bear the mistake and will never be available at the new site. I marked them as fixed in the tools. Do I need to do something about that, like 301 rewrite them with a condition to fix the error? Q2: Does Google bot attempt to recrawl 301 pages that pointed to a 404?

    Read the article

  • Are my Google Analytics ( 2 domains 1 site) duplicated or unique?

    - by MarcDJay
    We have recently built a new website with a new domain to replace an old website, and on the advice of our IT guys and web dev team have pointed both oldaddress.com's & newaddress.com's a records to the new website. Now, they both share the same google analytics code (UA-12345-1) and as such we have two entries in the Google Analytics dashboard. The problem is I'm still fairly novice with GA and as the reports seem VERY similar (~25k pageviews for each domain), are these figures exclusively for that domain? For example: oldaddress.com 25,400 pageviews newaddress.com 25,600 pageviews Does this mean that in total for this website I have 51,000 pageviews. Hope this is clear enough but let me know if anything needs clarifying. Thanks.

    Read the article

  • Google indexing pages with #! although we don't have any

    - by Benjamin Gruenbaum
    Our company has developed a Single Page Application using AngularJS and its routing. Google indexed our site decently with JavaScript but it did not index some pages very well so we have developed an HTML only version. We have followed the Ajax Crawling Specification posted here and have a <meta name='fragment' content='!'> tag and canonical urls. We expect http://www.example.com/foo/bar to be fetched from http://www.example.com/?_escaped_fragment_=/foo/bar. However, we have found out that when we rolled the AJAX specification we now have all pages indexed twice, once with the JavaScript version as http://www.example.com/foo/bar and once with the new version as http://www.example.com/#!/foo/bar. This is harmful to us since it's duplicate content and also mis-representing out site. I have tried looking for similar questions here and in the Google product forum but could not come up with anything.

    Read the article

  • how to get all group images in facebook using FACEBOOK API

    - by From.ME.to.YOU
    Hello i'm trying to get all the images from a facebook group using facebook API i have a problem i can't get all the photos using $facebook->api_client->call_method('Photos.get', array('subj_id' => $uid)); http://wiki.developers.facebook.com/index.php/Photos.get what i'm using now is the method $albums = $facebook->api_client->photos_getAlbums($uid, NULL); http://wiki.developers.facebook.com/index.php/Photos.getAlbums and then loop for ever album on $facebook->api_client->call_method('Photos.get', array('subj_id' => $uid)); then i add every new array results to my big array $big_array = array_merge($big_array,$result_array_from_that_call); 2 problems occurs here : 1- sometimes this calls fails - i think because of too many calls per second - 2- the request takes a v.long time to process is there a better way to do that? Thanks guys Cheers EDIT :: i tried to get all the images using $facebook->api_client->call_method('Photos.get', array('subj_id' => $uid)); and using $uid as the group ID but that's doesn't work " don't know why maybe because all the images is listed in groups "

    Read the article

< Previous Page | 143 144 145 146 147 148 149 150 151 152 153 154  | Next Page >