Search Results

Search found 4990 results on 200 pages for 'traffic measurement'.

Page 57/200 | < Previous Page | 53 54 55 56 57 58 59 60 61 62 63 64  | Next Page >

  • Is there a general rule of thumb for which browsers to optimize your site for?

    - by Christian
    I have a site (recently relaunched it with a new design) that I have put off optimizing for ie7 for far too long. I was just never too worried about it. The site is optimized for ie8-10, Firefox, Chrome, Opera, Safari, etc.. Then I asked myself, is it even worth it? I checked traffic over the last couple months before the relaunch and about 1.3% of the traffic is coming from ie7. So, is there a general cuttoff percentage when you would not optimize for a specific browser?

    Read the article

  • Ideas for making money off a website

    - by bradenkeith
    I'm considering acquiring a website that traffics close to 20k a month. It's currently unprofitable, but obviously I would need to change that to justify the cost. It's a support site for a framework similar to CakePHP. A resource that allows people to find plugins for that site. How would you go about expanding this site so that it would be broad enough to raise traffic and how would you make money off of it? Some ideas I've had: 1- BuySellAds.com seem to make ads classy and relevant 2- Team up with other sites in some way (What ways? ... Ecommerce sites that sell plugins (like CodeCanyon), etc) 3- Write tutorials and such to drive more traffic (time consuming, and already heavily saturated for this framework). Are there resources out there that help webmasters in ways like this?

    Read the article

  • Website Stopped Showing From Google Search Results Sunddenly

    - by Aman Virk
    I have a design and development blog http://www.thetutlage.com (1.5 years old), which was doing really well in Google search as I was getting over 70% of my traffic from Google. Now suddenly from last two days it reduced the amount of traffic from 70% to 20% and also when I am trying to search for the exact posts that I can created even after appending my website name to it does not show any results for that. Sample Search Text: JQuery Game Programming Creating A Ping Pong Game Part 1 I have post with exact same title and it does not show it on Google search anywhere. I am totally shocked, I write my own unique content and follow Google guide lines like bible. Also there is no message under my webmasters account stating any problem or error.

    Read the article

  • Over a million COBOL porgrammers in the world?

    - by Lucas McCoy
    I think I heard on a previous StackOverflow podcast that COBOL was used as the programming language for traffic lights (or something like that), so this got me interested. I did a quick Google search and found this little article: Today, Cobol is everywhere, yet largely unheard of by millions of people who interact with it daily when using the ATM, stopping at traffic lights or buying a product online. The statistics on Cobol attest to its huge influence on the business world: There are over 220 billion lines of Cobol in existence, a figure which equates to about 80 per cent of the world’s actively used code. There over a million Cobol programmers in the world. There are 200 times as many Cobol transactions that take place each day than Google searches. I didn't really trust the source seeing as how it's on some random PHPBB forum. So how accurate are these figures? Are there really 220 billion lines of COBOL? I assume a few people/companies still use COBOL, but how many?

    Read the article

  • Understanding the maximum hit-rate supported by a web-server

    - by SNag
    I would like to crawl a publicly available site (and one that's legal to crawl) for a personal project. From a brief trial of the crawler, I gathered that my program hits the server with a new HTTPRequest 8 times in a second. At this rate, as per my estimate, to obtain the full set of data I need about 60 full days of crawling. While the site is legal to crawl, I understand it can still be unethical to crawl at a rate that causes inconvenience to the regular traffic on the site. What I'd like to understand here is -- how high is 8 hits per second to the server I'm crawling? Could I possibly do 4 times that (by running 4 instances of my crawler in parallel) to bring the total effort down to just 15 days instead of 60? How do you find the maximum hit-rate a web-server supports? What would be the theoretical (and ethical) upper-limit for the crawl-rate so as to not adversely affect the server's routine traffic?

    Read the article

  • Will using two different tracking codes affect my SERP

    - by Danny Hefer
    Hello everyone and thanks for your time! I am now facing a problem after a site migration. New site is basically an improved version of old site, with the same content and some extras. After pointing the domain name to the new site, the old site was still online for a while but didn't get any traffic. The new site has its own tracking code. So, old tracking code has age (something like 7 years) but no visitors for a month, but new tracking code is a month old with an acceptable traffic. How to you think google will react if I add old tracking code to new site? Thanks by advance!

    Read the article

  • New versions of Firefox, Fiddler and SVN

    - by TATWORTH
    There are new versions of Fiddler and Tortoise SVN at: http://www.fiddlertool.com/fiddler/version.asp http://tortoisesvn.net/downloads.html  Fiddler is an excellent tool for montoring web service traffic. It also will capture traffic to and from your browser. Firefox needs no introduction from me, just be sure to go in firefox, Help, Check for Updates and make sure you have the latest. Since various browser exploits are about to be publicly discussed, there has been a recent flurry of browser updates, please be sure to get the latest in advance of exploits being made public.

    Read the article

  • Broken links in content reports when tracking subdomains with Google Analytics

    - by Rob Sobers
    I have a tracking code that I use on my main site and my blog, which is on a subdomain: www.example.com blog.example.com I have a single profile in Google Analytics. I use advanced segments to look at traffic to the main site vs. traffic to the blog. Problem 1: When I'm browsing my content reports under Standard Reporting, the "Page" column doesn't show the top-level or sub-domain, so I can't differentiate www.example.com/index.html from blog.example.com/index.html easily. According to the docs, this filter is supposed to make GA prepend the hostname to the page URL in your content reports, but it doesn't seem to work. Problem 2: When I click on the little "Open in new window" icon next to a given page in a content report line, it always assumes the page lives on www.example.com, so I get 404s when the page is actually on blog.example.com. Is there a good solution for these subdomain tracking problems?

    Read the article

  • How frequently Googlebot fetch sitemaps? Is it depending on page rank?

    - by JITHIN JOSE
    How much frequently google fetches sitemaps? I am now working with a high traffic website normally have 30 new posts per minute.But currently it provides sitemaps which links to new 100 posts(3 minutes). Is this method is enough ?. Is Bots fetch sitemaps every 3 minutes?. Did need to change sitemaps to list all 5M posts(indexed sitemaps)?. How this change will effect on traffic and page rank. Is google bot remove urls that previously listed on sitemap but not now?

    Read the article

  • Google Analytics Campaigns Not Tracking E-Commerce

    - by Paul
    I am running email campaigns via MailChimp and tracking the success of my campaigns via Google Analytics. I can successfully see data being tracked for: Reporting > Conversions > Ecommerce (Receiving Data) Reporting > Traffic Sources > Campaigns (Receiving Data) However, I am not receiving any Ecommerce data for the individual campaigns: Reporting > Traffic Sources > Campaigns > Ecommerce (No data) So I see data like: Visits: 18,501 Revenue: $0.00 Everything I have read leads me to believe this should just "work" if Ecommerce is setup. Is there some additional action I need to take for this work? Any help would be appreciated!

    Read the article

  • USB seems to pause system

    - by Marco van de Voort
    I've an application that does some simple measuring, for which it polls a few 100kbs several times a second (8-25 times) The behaviour is not really dependant on chipset (happens on several mobo's intel 965- P55) and OSes (XPsp3 and win7). Also the make of the USB keyboard doesn't seem to matter. I notice that sometimes when an USB kbd is plugged in, the system pauses for say 500-1000ms. (about 900-1000ms on disconnect, and 400-500 on the subsequent connect) It also happens for other USB devices (most notably mice and massstorage devices), but only the first time such device is connected to an installation. This disrupts the measurement and I really would like to get rid on this. I already tried to disable as much as possible. (powersave, teletubby mode (*) etc), and while this helped with the non-USB related disruptions of the measurement, it doesn't help with the USB related ones. (*) fyi, turning off themes (to resp. classic/non-aero), and turning off effects in system solved problems that occured when minimizing/maximizing the app. Any pointers to look into? I'm a bit stuck with this.

    Read the article

  • Keyword Generator Tool Gets Your Ahead of the Competition

    A keyword generator tool provides ideas that website owners and search engine optimizers use for site and engine optimization. Key phrase generators rely on search query popularity from introductory keywords to a more complex keyword search management to drive more traffic to a website. It maximizes prospective and potential high-traffic keywords and integrates it with your sites campaign techniques. Keyword generator tool allows you to manage and add "exact match" and "phrase match" keywords to your lists, also allows you to create misspellings, combine and reverse keywords then automatically calculates the ad group focus score of your keyword lists.

    Read the article

  • Weird issue with iptables redirection

    - by skypemesm
    I am trying to redirect all incoming traffic on UDP port 5060 to port 56790, and all outgoing traffic from 5060 to the port 56789. I used these iptables rules: iptables -t nat -I PREROUTING -p udp ! -s localhost --dport 5060 -j REDIRECT --to-port 56790 iptables -t nat -I OUTPUT -p udp ! -s localhost --sport 5060 -j REDIRECT --to-port 56789 I listen on both ports using RAW SOCKETS after setting the interface to PROMISCUOUS mode using ioctl. I see packets ONLY on 56789 i.e.SENDING side, and I do not see any packets on 56790, while wireshark shows that many packets are delivered to port 5060. Why would this happen? Any ideas? Do you think it's a problem with iptables rules or something to do with raw sockets? [This is ubuntu 10.04 and iptables v1.4.4]

    Read the article

  • Multicasting and VMWare

    - by John Breakwell
    Cracked a Multicasting problem this evening for one of my Canadian Tweeple. They wanted to mulitcast some MSMQ messages to another machine but nothing was arriving in the listening queue. A local queue could be configured to listen to the particular IP address/port in use and messages would arrive, though. Looking at the network traffic, nothing was going onto the wire for the IP address/port pair until they looked at traffic to the VMWare adapter. The machine had a virtual machine to simulate a remote computer and when they changed the setup from NAT to Bridge, multicasting burst into life.

    Read the article

  • Tor and Anlytics how to track?

    - by Jeremy French
    I make a lot of use of Google Analytics, Google has reasonable tracking for location of users so I can tell where users come from. I know it is not 100% but it gives an idea. In the wake of Prism it is possible that more people will make use of networks such as tor for anonymous browsing. I have no problem with this, people can wear tin foil hats while browsing my site for all I care, but it will lead to more erroneous stats. Is there any way to flag traffic as coming from TOR, so I can filter location reports not to include it, and to get an idea of the percentage of traffic which does use it? Has anyone actually tried this?

    Read the article

  • Can I Use A Canonical Tag Instead of a Redirect for Updated Content?

    - by Ewan Heming
    I have some old articles on my blog that get quite a bit of traffic, but are very outdated. I want to remove them from Google's index using the noindex tag, but I'm not sure what the best approach will be to send the same traffic to my new article on the subject without using a redirect (as I want to keep them in my blog archives). I was intending to just put a link at the top of the article pointing to the new one, but was wondering if it was appropriate to use a canonical tag instead; the new article is on the same subject but doesn't contain the same content, so isn't really a copy.

    Read the article

  • Usefulness of blogs for multi language ecommerce site

    - by jawilson
    I have a multi-language ecommerce site for which I am trying to improve its online marketing. There's currently a blog for the english shop, on a subdomain, which doesn't really get very much traffic. I'd been recommended to set up blogs for each of the different language shops and try to generate traffic to each of the blogs. I'm wondering if this is really worth the effort and cost (blog devt costs and ongoing translations required). Does anyone have any experience/advice on this at all? Perhaps where they've used this approach successfully or otherwise?

    Read the article

  • Media Hint Brings Hulu, Netflix, and Pandora to non-U.S. Residents

    - by Jason Fitzpatrick
    Chrome: If you’re outside the United States you know all too well how irritating it can be when you’re denied access to streaming services because of your location. Stop missing out and start streaming with Chrome extension Media Hint. So what’s the secret sauce? The extensions routes your traffic–just the traffic for the streaming service–through U.S.-based proxies so you can enjoy content unavailable in your home country. The extension requires no configuration or registration, simply install it and visit Hulu, Netflix, or Pandora to start streaming. Media Hint is a free extension, Chrome only. Media Hint [via Digital Inspiration] 6 Ways Windows 8 Is More Secure Than Windows 7 HTG Explains: Why It’s Good That Your Computer’s RAM Is Full 10 Awesome Improvements For Desktop Users in Windows 8

    Read the article

  • How can I use smbclient to connect to Windows shares by hostname when a firewall is enabled?

    - by skyblue
    I can't connect to file shares on Windows computers using smbclient -L //hostname when the firewall is enabled. This occurs whether I'm using ufw (which allows outgoing traffic and replies back in with the default configuration) or iptables (where I'm allowing outgoing traffic and replies back in with iptables -A INPUT -m state --state ESTABLISHED,RELATED -j ACCEPT). However, smbclient -L //ip-address works whether the firewall is enabled or not. I also tested this against a Samba server running on Ubuntu and again smbclient -L //hostname does not work when the firewall is enabled, but smbclient -L //ip-address works whether the firewall is enabled or not. For reference, here are the iptables rules I used during testing: *filter :INPUT DROP [0:0] :FORWARD ACCEPT [0:0] :OUTPUT ACCEPT [0:0] -A INPUT -m state --state RELATED,ESTABLISHED -j ACCEPT -A INPUT -p icmp -j ACCEPT -A INPUT -d 224.0.0.251/32 -p udp -m udp --dport 5353 -j ACCEPT COMMIT

    Read the article

< Previous Page | 53 54 55 56 57 58 59 60 61 62 63 64  | Next Page >