Search Results

Search found 14789 results on 592 pages for 'pro backup'.

Page 331/592 | < Previous Page | 327 328 329 330 331 332 333 334 335 336 337 338  | Next Page >

  • will redirect subdomain lose Google SEO links

    - by user29160
    Because of a change in brand, I want to redirect our subdomain.domain.com to a newdomain.com. The content being exactly the same, I was thinking of using a 301 wild card 301 redirect to newdomain.com. I noticed it is not possible to do a redirect in Google webmasters tool as you can with a root domain. Is there a way I can do this redirect without losing all the backlinks referenced with Google? All the best, mark

    Read the article

  • 2 Google Tag Manager containers = double triggers?

    - by fred
    I have a Joomla website with existing analytics done by a GTM tag and I need to add another image tag (1x1 pixel) from a separate GTM container. I set the event name for the additional GTM container to a different name from from the existing GTM container (default 'gtm.js') so that the new tag only fires under the specified event. The new container tested out fine in a blank HTML page, but when it is put in the website, it ends up firing twice. I know that because Firebug showed 2 1x1 pixels being loaded and I mark each request with a randomly-generated UID to distinguish them on the server side. I suspect this being caused by having multiple GTM container tags but want to check whether anyone has run into this problem before? As of now, I could not verify nor fix the double counting problem.

    Read the article

  • How to handle CNAME host redirect to virtual directory?

    - by esac
    I have an internal website and virtual directory http://server2012/logs. I created a CNAME on my DNS server as LOGS - server2012. I would like to set it up so that http://LOGS redirects to http://server2012/logs. Ideally, I would still want it so that all pages appear in the browser as being off from the LOGS URL. So http://LOGS/network.html?site=32 is what is displayed in the browser, but it is really being served from http://server2012/logs/network.html?site=32. I've looked at URL rewrite, but can't seem to get to work.

    Read the article

  • In addition to Google's First Flick Free, should you whitelist search engine bots past a paywall?

    - by tobek
    Our site has subscription-only pages - non-subscribed visitors see a snippet preview. As per Google's FCF requirements, your first 5 hits to a subscriber-only pages with .google. as the referrer, you see the full page. In addition to this, should we whitelist search engine bots so that they can index the full content? I assume this is not required for Google, which can use FCF to index our content, but what about other search engines? Is this considered cloaking? My gut says that whitelisting bots past the paywall is bad practice., but I wanted to confirm - any evidence or references would be amazing.

    Read the article

  • Why would i need extra IP adresses from my web host?

    - by user4524
    I am moving to a new and cheaper host. My old one raised the prices suddenly by a factor 10. Now what I did not like about the old one, was that each time I set up a new website, I had to set up a new account, with the result that I have to pay for a lot of webspace and a lot of bandwith i don't use. Now I am moving over to a new host, who cheaply offers virtual servers. Now I am a n00b when it comes to server tech. I have two questions about this: 1. Could I install all my webistes on this virtual server, provided it is big enough? 2. They offer the possibility of getting more than one IP address, but it costs more. What would be the advantage of this? Would I be able to suffice with one IP address?

    Read the article

  • quickest way to research a set of pages backlinks

    - by JeremyB
    I have a list of 300+ pages (they were chosen based on which pages rank for a keyword I'm interested in) and I want to compile a list all the (known) inbound links to those pages. What's the fastest way to do this? It seems like the only tools out there-- Yahoo Site Explorer, SEOMoz, Majestic, require you to either a) manually export each set of links by hand, or b) get data at the domain level (e.g. Majestic's clique hunter). Does anyone know of any efficient way to do this? I ask because I'm about to write a bunch of code and I don't want to waste my time if there's another tool that will work. I know SEOMoz and Majestic have API's but I'm wondering if there's a more user-friendly option.

    Read the article

  • List of events triggered on pages matching regex

    - by Cubius
    Is there a way to get the grouped list of events (such as in Top events) which were triggered on pages matching a regular expression? I may add the Page secondary dimension in Top events and apply the regex filter but this way I won't get a grouped list. I may apply the filter to Events - Pages report but this way the events will be grouped only inside pages whilst I need global grouping. Any suggestions?

    Read the article

  • More than 5 custom variables across multiple websites using Google Analytics

    - by brakes
    We have multiple websites using the same Google Analytics account number so we can track visitors across multiple websites. One of these websites has set 5 custom variables. We want to introduce a new custom variable to track logged in users for our single sign-on (SSO) system to find out what parts of which website they are accessing. Is this possible or is it a case that all the custom variables have been used up by 1 of the sites?

    Read the article

  • Is there a colloquial term for the web dev equivalent of "Script Kiddies"? [on hold]

    - by Darkcat Studios
    I recently came across a couple of lets say "young people" who were convinced that they were "Web Developers", despite the fact that they were only able to configure a wordpress template, and not even able to edit images properly. So this got me thinking, "back in the day" kids who thought they were hackers but just ran other peoples scripts they had found, were known as Script Kiddies. Do we have a term for these yet?

    Read the article

  • certificate for website login

    - by Mario
    Not sure if this belongs here or at serverfault... I've seen websites where, to login to the website, requires a digital certificate to be installed for the user logging in. As far as I can tell, this certificate is in addition to the website using an SSL certificate (https) I'm just looking to be pointed in the right direction on how to code for this (apache / php hopefully), who issues these certificates (must it be a trusted var or can I ?) or even what to search for via google. -Mario

    Read the article

  • Merging two sites into one, how to redirect from the domain that's going away?

    - by bikeboy389
    I haven't been able to find any existing questions that cover my exact issue, so here goes: My client wants her two sites (domain1.com and domain2.com) rolled into a single, new site under domain1.com. Once the site is ready on domain1.com, DNS for domain2.com would be pointed at the same server as domain1.com. I know how to do an htaccess rewrite rule that would make all domain2.com traffic map to a specific single page or directory within domain1.com. But that's not what the client wants. What she wants is for a bunch of specific pages on domain2.com to map to specific new pages on domain1.com. For example: domain2.com/index.php?pageid=58 GOES TO domain1.com/2011/04/somearticle domain2.com/index.php?pageid=92 GOES TO domain1.com/2011/03/differentname etc. I could put a bunch of 301 redirects in the htaccess on domain1.com, which would work fine. The problem is, the client doesn't want/need specific redirects for ALL the domain2.com pages, and if I just do 301 redirects, anybody who comes looking for a domain2.com page that I haven't built a specific redirect for will get a 404 error. So I need to use 301 redirects for some traffic, and a rewrite rule for any traffic that's not covered in the 301 redirects. How do I do sort of a blending of a rewrite rule and 301 redirects, all in the htaccess file for domain1.com? Is this possible? Is it as simple as putting the 301 redirects in the htaccess file first, then doing the rewrite rule? I'm guessing not.

    Read the article

  • How Does Domain Know Where Your Web Host Is Located [closed]

    - by icu222much
    Possible Duplicate: How Does Domain Know Where Your Web Host Is Located? I just purchased a domain name from RapidNames, and a hosting plan at JustHost. I was told to enter JustHost's name server (ns1.justhost.com) in my domain name's name server field and wait for 24 hours for the process to be complete. I do not understand how RapidNames can find my account on JustHost's server as I am sure I am not JustHost's only customer. I have read the article How DNS Works that John Conde has posted, but I still do not understand the issue. After reading several other articles, I am beginning to understand how it works, but I would still like someone to confirm if I am correct or not. From my understanding, linking your domain name to your web host is a two step process. First, you need to tell your domain name who your web host is. This is done by providing the two DNS server addresses. Secondly, you need to tell your web host which domain names you own by entering your domain names into the domain name manager. As a result, when someone queries your domain name, they will be forwarded to your web host. The web host will look in their database to match the domain name the account's owner, and then serve the appropriate website. I want to confirm if my understanding of how a domain knows where your web host is located is accurate?

    Read the article

  • how did Google Analytics kill my site?

    - by user1813359
    Yesterday I created a google analytics profile for one of my sites and included the JS block in the layout template. What happened next was very strange. Within about 2 minutes, the site had become unreachable. I had been checking the AWStats page for the site when I thought to set up GA. After that had been done, I clicked on the link for 404 stats, which opens in a new tab. It churned for a long while and then showed a nearly blank page, similar to that when Firefox chokes on a badly-formatted XML page, except there was no error msg. But i was logged into the server and could see that that page has a 401 Transitional DTD. Strange! I tried viewing source but it just churned endlessly. I then tried "inspect element" and was able to see an error msg having to do with some internal Firefox lib. Unfortunately, i neglected to copy that. :-( All further attempts to load anything on the site would time out. Firebug's Net panel showed no request being made. Chrome would time out. So, I deleted the GA profile, removed the JS block, and cleared the server cache. No joy. I then removed all google cookies and disabled JS. Still nothing. No luck in any other browser. And now my client couldn't access the site. Terrific. I was able use wget while logged into another server. The retrieved page was fine, and did not contain the GA JS block. However, the two servers are on the same network. (Perhaps a clue.) The server itself was fine. Ping, traceroute looked great. I could SSH in. I tailed the access log and tried a browser request. Nothing. But i forgot to quit and a minute or so later I saw a request from someone else being logged. Later, I could see that requests had been served all day to some people. Now, 24 hours later, the site works once again, but is still unreachable by the client (who is in another city). So, does anyone have some insight into what's going on? Does this have something to do with google's CDN? I don't know very much about how GA works but what I'm seeing reminds me of DNS propagation issues. And why the initial XML error? And why the heck was the site just plain unreachable? What did google do to my site?! Sorry for the length but I wanted to cover everything.

    Read the article

  • Created Custom Report in Google Analytics, Primary Account Doesn't See It?

    - by Anagio
    A client shared access with me to their Google Analytics account. I created a custom report which shows up under Custom Reporting for me. I assumed they would also see this report since it was in their account but they sent me a screen shot showing there's no custom report listed. I have already sent them the shortcut link to the custom report configuration. This seems to be the way to share custom reports along with dashboards in GA now. Do custom reports only appear to the accounts (email) that created them? I would think everyone who had access to the account would see the custom report.

    Read the article

  • How to track opens and pageviews in PDFs?

    - by Osvaldo
    I know how to track clicks in links to pdfs and pfd's downloads. But I need to track how many times a PDF is opened after being downloaded and if possible track how many times certain pages are shown to users. Tracking has to be done without warnings that personal information is being sent somewhere. I do not want readers personal informations, just to know how many opens happened, so this warnings would be inaccurate. Can anyone help by pointing to a tutorial or an example? If you are sure that this can't be done, can you please point to documentation that explains why?

    Read the article

  • Would my domain(register in USA ) or vps(locaton in USA) be banned because of hold pirated resources? [on hold]

    - by Mithril
    I am building a torrent sharing site(like thepiratebay).I don't want to waste time to handle piracy, it is really annoying. I plan to register a domain at godaddy or namecheap, and buy a vps on digitalocean. They are all American company. So I wonder if I buy a domain via a American provider or host my site at on VPS in American.Would they force to ban my domain or vps because of piracy? Is there any good place to register a domain or host a argument site?

    Read the article

  • Duplicate page content and the Google index

    - by Kit Sunde
    I have a static pages with dynamically expanding content that google is indexing. I also have deep links into virtually duplicate pages which will pre-expand the relevant section of content into the relevant section. It seems like Google is ignoring all my specialized pages and not putting them in the index. Even after going through web-masters tools, crawling and submitting them to the index manually. I also use the google API for integrating search on the site, and the deep linked pages won't show up. Is there a good solution for this?

    Read the article

  • How do I fix the paths of my website?

    - by EASI
    I have Joomla 2.5.7 in my client's server updated recently from 1.6 to 1.7. I did not make that site but I am responsible for it now. I prefer to make a site from zero. Now users area clicking on the menu options and when joomla send them to a meta-url like http://iap.pa.gov.br/acervo they get the message 404 The requested URL /acervo was not found on this server. Would that be because they moved joomla folder from the root to root/iap (name of the site)? If it is, what is the configuration to adapt to that new folder?

    Read the article

  • display Google display ads to visitors who have visited certain web sties

    - by Source
    For Google Adwords re-marketing, display ads are shown to visitors have been to your site previously. So when the go to a web site displaying adsense, it is likely that re-marketing ad will be displayed to them. Is there a way to do the same for if a visitor has visited a competitors site. i.e. if a visitor goes to one of my competitor sites, I want the display ads they see, to be mine. Is that possible?

    Read the article

  • Cannot access my own web page

    - by enflam3
    I am developing, learning and experimenting with php,html,javascript,flash and so on. Having web hosting and all of the cpanel, phpmyadmin and other utilities. One day, while updating information, connection between my computer and website just went down. I found out that it is only from this computer, where I cannot access anything. I don't know what is the reason I cannot access website, however this is what I have checked so far: Everything else opens normally, having problem only with my page. Cannot access FTP,cpanel or any kind of information related to the domain and hosting ipconfigs detects IP, but shows request timed out (so its not browser related) Turned off Firewall,AV, Rebooted computer Cleared caches,temp,cookies,histry with CCleaner Checked connectivity with both (wired,wireless) networks ISP has dynamic IP that has been changed about 3 times since issue Checked host file I am out of ideas and understanding what could cause this kind of issue, however couple minutes ago, found out that everything works with proxy server (when adding IP and port to the browsers) Can someone point out what should I check or try to get rid of this problem?

    Read the article

  • best way to import 500MB csv file into mysql database?

    - by mars
    I have a 500MB csv file that needs to be imported into my mysql database. I've made a php file where i can upload the csv file and it analyses the fields n stuff and does the actual importing. but it can only handle small fiels max 5mb. so that's a 100 files and actually pretty slow(uploading) is there another way? I have to repeat this process every month because the data in the file changes every month it's about 12 000 000 lines :D

    Read the article

  • Restricting A Directory Through .htaccess

    - by Whitechapel
    I'm trying to put all of my FTP accounts into a folder on /public_html/ftp and password protect it so search bots can't crawl their private files. I'm also trying to redirect all site traffic from the non-www to www. I keep getting 500 errors when accessing the site, and I need to point it to www.vivalanation.com/ftp to www.vivalanation.com/ftp/, because the /ftp just errors out, you need the trailing slash. Here is my .htaccess in the /public_html/ftp folder: RewriteEngine on RewriteBase / RewriteCond %{HTTP_HOST} !^www\. [NC] RewriteRule ^(.*)$ http://www.%{HTTP_HOST}/$1 [R=301,L] AuthName "FTP Access" AuthType Basic AuthUserFile /home1/vivalst/.htpasswds/public_html/ftp/passwd Require valid-user I created a passwd file in /.htpasswds/public_html/ftp And here is my basic .htaccess in the root of /public_html/: RewriteEngine on RewriteBase / RewriteCond %{HTTP_HOST} !^www\. [NC] RewriteRule ^(.*)$ http://www.%{HTTP_HOST}/$1 [R=301,L]

    Read the article

  • How would I migrate a front page form based site to an IIS7 server? [on hold]

    - by Nick Gilpin
    I have an large html website on a very old server and I'm attempting to move it to a brand new server running IIS7. I've tried moving the entire site to the WWW directory, but then all my forms return "The HTTP verb POST used to access path is not allowed". Additionally, all the solutions I've found online only work on a page by page basis. Is there a simpler/faster way to migrate a large website containing many forms to IIS7? Here is the specific error message: The HTTP verb POST used to access path '/FormServer/Mig/_vti_bin/shtml.dll/admissions/askseaaggie.htm' is not allowed.

    Read the article

  • Links don't work on new server

    - by Beko
    I've recently moved from server with php4 to php5. Now I cant open content in my page. Links look like this "page.com/news-it-is-rainy-day-1.html". But when I try to open link I just get the archive (news.html). I also use smarty template engine. My rewrite rule in .htacces looks like this: Options +FollowSymlinks RewriteEngine on RewriteRule ^news-(.*)\.html$ news.php?do=read&id=$1 RewriteRule ^news.html$ news.php RewriteRule ^index.html$ index.php AddType application/x-httpd-php .xml

    Read the article

  • Removing "www." from domain name and SEO

    - by TecMan
    We are doing a big redesign work on our website, and at least 50% of the website folders will be moved to new places with different names (i.e. many URL will be changed). Sure, Google will need some time to index new pages and we expect our SERP positions will be not so good as they are now for some time. We also have an old idea to remove www from our domain name. It seems, it's the right time to do these two works together with publishing the website with updated contents. Or is it better from SEO perspective first publish the new contents, and only after some time, when our SERP positions will return to prior results, tell Google that the domain name without www is our preferred domain name?

    Read the article

< Previous Page | 327 328 329 330 331 332 333 334 335 336 337 338  | Next Page >