Search Results

Search found 32130 results on 1286 pages for 'local search'.

Page 388/1286 | < Previous Page | 384 385 386 387 388 389 390 391 392 393 394 395  | Next Page >

  • 301 redirect from a country specific domain

    - by Raj
    I originally started using a .do domain extension for my site, but later realized that this country specific domain would prevent us from appearing in search results for places outside of the Dominican Republic. We started using a .co domain extension and redirected all requests to the new domain using an HTTP 301. The "Crawl Stats" in Google Webmaster Tools shows me that the .co domain is being crawled, but the "Index Status" shows the number of pages indexed at 0. The "Crawl Stats" for the .do domain says that it's being crawled and the "Index Status" shows a number greater than 0. I also set a "Change Of Address" in Google Webmaster Tools to have the .do domain point to the new .co domain. We're still not appearing in search results at all even for very specific strings where I would expect to find us. Am I doing something wrong?

    Read the article

  • How to get nicer error-messages in this bash-script?

    - by moata_u
    I'm trying to catch any error when run a command in order to write a log-file / report I've tried this code: function valid (){ if [ $? -eq 0 ]; then echo "$var1" ": status : OK" else echo "$var1" ": status : ERROR" fi } function save(){ sed -i "/:@/c connection.url=jdbc:oracle:thin:@$ip:1521:$dataBase" $search var1="adding database ip" valid $var1 sed -i "/connection.username/c connection.username=$name" #$search var1="addning database SID" valid $var1 } save The output looks like this: adding database ip : status : OK sed: no input file But I want it to look like this: adding database ip : status : OK sed: no input file : status : ERROR" or this: adding database ip : status : OK addning database SID : status : ERROR" I've been trying, but it's not working with me. :(

    Read the article

  • SEO Tips - Updating Your Content and Avoiding Duplicate Filters

    If you are just getting into Search Engine Optimization, it's important that you are aware that content is the most important thing in your website in order to get high ratings. If you decide to hire an SEO Consultant, make sure that you explain to them exactly what you want your website to look like and the content that you would like it to feature. Having fresh and relevant content in your website is the key as that will bring web crawlers back frequently. There are different ways of achieving this while avoiding getting your website removed off of a search engine due to duplicate content.

    Read the article

  • How to use Google Analytics as an affiliate to track sales data

    - by lalex
    As an affiliate, how can we get more information on sales? It looks like the goals feature in GA is for those who have control over the receipt page. But we are sending users away using an affiliate link. With event tracking, we've been able to count the clicks and see which links are being clicked the most, but not which ones actually convert. We want to find out the following on each sale: Did the converted user come from search or internal traffic? If it was search, which keyword brought the user to our site (and clicked away and converted)? Is it possible?

    Read the article

  • CSS being Displayed by Google spiders

    - by vipul_vj
    I have written an article HTML Image Tag for the site and it has been indexed by Google. But when I search it, google displays HTML Image Tag - ProgrammingBulls http://programmingbulls.com/html-image-tag-1: content { font-family:verdana; font-size:14px; font-weight:normal } We often use images in a webpage. To insert images in our webpage < img tag is used in. Why is CSS displayed in the google search? I know that CSS and HTML is ignored by Google but due to some reason HTML is being displayed.

    Read the article

  • Why am I getting this message "Some important page is blocked by robots.txt"?

    - by Rounak
    My site's URL is: www.hackinguniverse.org From some day in Google Webmaster tool a message is showing that says "Some important page is blocked by robots.txt". My robots.txt is: User-agent: Mediapartners-Google Disallow: User-agent: * Disallow: /search Allow: / Sitemap: http://www.hackinguniverse.org/feeds/posts/default?orderby=updated Now for other information - I hosted this website on blogger. I have some other sitemaps too where I had only included "/Search" in Disallow like in this robots.txt file. But those sites are ok. I mean no message is showing on those. So why am I getting that message telling that I have blocked some important page via robot.txt?

    Read the article

  • why my site cache 2 time par day?

    - by clarawood
    I have read the FAQs and checked for similar issues: YES My site's URL (web address) is: www.adultxdating.com Description (including timeline of any changes made): I have lost my top search listings from last 4 months. I am still working on this but not getting proper guidance. This site is caching 2 times in 24Hrs. Some times sites will back in to top 10 search listing on 100s keywords, some time its gone out 1000+. anybody can help me why its happening. I have more than 200K+ incoming links and updating the site regularly. Please help. Thanks clara wood

    Read the article

  • localhost .htaccess not working?

    - by diff
    Now i install WordPress then BuddyPress, after the install buddypress i run the site, home page was run, when click the other page link the page was not working. show this error Not Found The requested URL /Repo/website/groups/ was not found on this server. Apache/2.2.17 (Ubuntu) Server at localhost Port 80 After then i check my local htaccess, here is all are ok, but why the problem is showing. Here is my local htaccess .. <Directory /> Options FollowSymLinks AllowOverride all </Directory> <Directory /var/www/> Options Indexes FollowSymLinks MultiViews AllowOverride all Order allow,deny allow from all </Directory> ScriptAlias /cgi-bin/ /usr/lib/cgi-bin/ <Directory "/usr/lib/cgi-bin"> AllowOverride None Options +ExecCGI -MultiViews +SymLinksIfOwnerMatch Order allow,deny Allow from all </Directory>

    Read the article

  • Sharing Large Database Backup Among Team

    - by MattGWagner
    I work on a team of three - five developers that work on an ASP.net web application remotely. We currently run a full local database from a recent backup on all of our machines during development. The current backup, compressed, is about 18 GB. I'm looking to see if there's an easier way to keep all of our local copies relatively fresh without each of us individually downloading the 18 GB file over HTTP from our web server on a regular basis. I guess FTP is an option, but it won't speed the process up at all. I'm familiar with torrents and the thought keeps hitting me that something like that would be effective, but I'm unsure of the security or the process.

    Read the article

  • Manual Directory Submission Service a Long Way to Go But Effective

    Yes you heard it right Manual Directory Submission Service is really a long way to go but very effective for your business. In fact this is a process of registering or submitting your website with its details to an online web directory. The advantage of this listing is that when a customer needs any service or product, he will search for business on a web directory for assistance and when your expertise meet with the keywords that he typed in search bar, you website link will be visible to him and he will click that link from where he will be redirected to your website.

    Read the article

  • install 64 bit as dual boot with 32 bit

    - by profgeorgefhart
    I downloaded Ubuntu 12.10 - bit .iso onto a bootable USB and want to put it on to my 32 bit 12.10 system as a dual boot, using a totally separate 1 Tb disk [which already has data on it]. However, when I go to restart and hit F8 fir BIOS set-up, I can change all the settings except the boot search order. Is there some other way to force the startup to read the USB first? Is the fact that I cannot change the search order to do with the fact that the 32 bit OS was 'Installed in legacy mode' not 'EFI? How can I rectify this? (This might be related -- might I need to update my BIOS for my Intel DX38BT motherboard? I think their last upgrade was 2009.)

    Read the article

  • Actinic and Google Analytics: does it mess with my stats?

    - by tjcss
    My fathers website is built using Actinic - not by me but by a local company - and since he went with them some years ago his traffic never went down, but has stayed more or less consistent which is fine. My question is this; does using actinic somehow confuse analytics? As it shows that 99% of all visitors come "direct", as in, not by organic search. Previous to using Actinic he would get 70 to 85% of new hits exclusively from organic search terms. So I'm wondering if Actinic somehow messes with these new hits and redirects them to a "home" page.. Not sure exactly what I mean but this change in stats is concerning and I'm struggling to find an explanation.

    Read the article

  • How to install texlive2011?

    - by vrx
    I just moved from Windows to Ubuntu. Software Center in Ubuntu does have texlive2009 package. There is no straightforward installation process from TUG, or from Google search result. I found some installation of Texlive2010 but it does not work. Here are the step: download an texlive2011.iso, 2.3GB and save in local drive mount to a virtual drive install perl tk (ok) "sudo ./install-tl" does not work, "perl ./install-tl" shows pre-installation setting but it does not have the write access to /usr/local/textlive/2011/ Finally, I'm stuck Please guide me with a step by step tutorial or suggest any other solution beside typing in the black screen without knowing what is going on behind the Terminal.

    Read the article

  • which status to put for temporarily inactive page

    - by aji
    I was wondering if someone could help me how to manage temporarily inactive website in regards of SEO and search engine. the case is i managed a big ecommerce site, and sometime i need to put down page(s). could be days, could be weeks, could be months, and it depends on our vendor. if my visitors land on the page that been temporarily inactive then i can give them a message that the vendor they looking for is not available at this time and he can check back later OR check another vendor with similar products, but how do i send my message to search engine robots? if i use 301 status and forward URL page to another similar products, then the chance that the current URL being deindex is huge while i still want to use that URL for the future if my vendor want to re-join. any advise will highly appriciated

    Read the article

  • secure offline PC storage accessible through javascript

    - by turbo2oh
    I'm attempting to build a browser-based HTML5 application that has the ability to store data locally on a PC (not mobile device) when offline. This data is sensitive and must be secure. Of course the trick is trying to find a way to be able to access the secure data with Javascript. I've ruled out browser local storage since its not secure. Could this be accomplished with a local database? If so, where could the DB credentials be stored? Javascript obviously doesn't seem like a good option to store them since its user-readable.

    Read the article

  • How to connect to other Ubuntu machine as X Server?

    - by Tibo
    When I want to work on my local Linux machines from windows, I use Putty+Xming and I set the Xming port in Putty connection settings. This way I have a terminal on the machine and if, for example, I run the command 'firefox' a firefox window is opened in Windows (while it is really running on the Linux machine). I want to do the same thing from other Linux machine (they are both Ubuntu machine) connect by SSH and become this connection X Server so any graphic component will be opened on local machine. How can I do that?

    Read the article

  • How to install CREBS on Ubuntu 14.04?

    - by user288556
    I tried installing CREBS on Ubuntu 14.04. unfortunately the repository installation doesn't work, so I decided to download the source code. But after following their instructions(namely running setup.sh as root) I get this error when typing "crebs" in a terminal: Traceback (most recent call last): File "/usr/local/bin/crebs", line 38, in <module> from crebs.main import main File "/usr/local/bin/../share/crebs/lib/crebs/__init__.py", line 6, in <module> from gtk import glade ImportError: cannot import name glade any hint? thank you

    Read the article

  • Using GDI+ vs HTML page

    - by Umesha MS
    I am working on an application which is UI intensive (i.e. we need to customize all the control to look different). For this is I planned to use GDI+ in win32. But one of the suggestion form our team member is, Instead of using GDI+ he asked me to have a web browser control, in that he asked me to display local html.That html page contains customized controls. Form his suggestion I created a web control and used some UI intensive local html pages. That looks good. But my question is, is it is a good idea to use web control to solve this?

    Read the article

< Previous Page | 384 385 386 387 388 389 390 391 392 393 394 395  | Next Page >