Search Results

Search found 7625 results on 305 pages for 'scraper sites'.

Page 23/305 | < Previous Page | 19 20 21 22 23 24 25 26 27 28 29 30  | Next Page >

  • COM+/Desktop Heap errors in IIS affecting sites at random?

    - by tresstylez
    We have a Win2K3 server that is hosting 30+ sites. Each site is configured to have its own unique application pool -- so that we can manually recycle specific sites if needed and not kill sessions for the others. From what I've read, the consequence of this type of setup is that each application pool worker process gets allocated a Desktop Heap (normally 512 kb's) and we limit the number of app pools we can serve. http://blogs.msdn.com/b/david.wang/archive/2006/01/25/security-considerations-of-usesharedwpdesktop-on-iis6.aspx PROBLEM: What we're seeing is that occasionally COM+ errors get triggered, presumably by hitting our 512 kb limit of the desktop heap -- and certain sites become unresponsive (or have errors) until we manually recycle that specific app pool. I know that I can increase the desktop heap limit to 1024, and make other tweaks/tunes, but I've been tasked with finding out what exactly causes one site's heap to max out as opposed to another. It seems that when we start seeing COM+ errors, the sites it affects are random -- small sites or big sites (heavier used). Is it based on process id? Traffic? Any pointers on understanding this a little more would be excellent. Thanks! jg

    Read the article

  • How does one encrypt site passwords for a Dreamweaver export file?

    - by ngreenwood6
    We have over 500 sites that we host. All of their ftp information is in a database. Whenever one of our programmers have to add a site they have to get all the info and set it up. However, I found that you can export them and it has all the info except for one problem. The password is encrypted. I am not trying to hack anything, I want to know how to encrypt our passwords so that we can import them using dreamweavers import feature. Can anyone tell me what encryption they use or a link on how to encrypt. I am not interested in decrypting at all because we already have all of them so it would not do me any good.

    Read the article

  • How do sites avoid SEO issues / legalities with subdomain unique ids?

    - by JM4
    I was looking through a few websites recently and noticed a trend I'm not sure I understand. Sites are creating unique referral URLs for customers in the form of: http://customname.site.com (If somebody were to use http://www.site.com/customname it would function the same way). I can see the sites are using 302 redirects at some point using Google Chrome then doing some sort of htaccess redirect, taking the subdomain name (customname) and applying it as a referral parameter then keeping in session during the entire process. However, there must be thousands of these custom URLs that people are typing in. How are each one of these "subdomains" not treated as separate URLs which in turn are redirected to the same page (in short, generating tons of links all pointing to the same page which Google would normally frown upon)? Additionally, the links also appear on the site themselves as clickable links so I'm not sure how these are not tracked. Similarly, the "unique" url is not indexed or cached in any Google search results. How is this capability handled? It does NOT highlight the referral aspect, but a true example of this is visiting http://sfgiants.com which does a 302 redirect to the much longer proper San Francisco Giants MLB homepage. I am wondering how SFgiants.com is not indexed (assuming that direct shortened link appears on several MLB pages)? 1 - I know these are 302 redirects, I can see this on the sites network flow. 2 - These links do in fact appear on the page itself because in some areas (for example, the bottom of the page may say: send this page to a friend! http://name.site.com/ which in turn would again redirect to something like http://www.site.com?id=name so the id value could be stored in session

    Read the article

  • A frequently updated mixed bag blog OR several seldom updated niche sites?

    - by Melanie
    Background I am a member of the website HubPages where I have about a hundred articles (and I'm always writing more.) Anyway, HubPages revenue model is 40% ad-share for them and 60% ad share for users. While the userbase there is really friendly, the site is REALLY slow, buggy and there is a ton of content on HubPages that is copied from other sources. Upon flagging these articles it takes a ton of time for mods to remove it and it's just generally dragging down my stuff. Furthermore, HubPages was hit really hard by Google's Panda Update: http://www.google.com/search?hl=en&rlz=1B3GGLL_enUS426US426&tbm=nws&q=google+panda& Aside from the temporary problems I would deal with when removing content from HubPages and putting it on my own domain (duplicate content, etc) I have another problem. Which would be the best for my articles? I have tons of articles in a wide variety of niches and would like to do what would help them perform the best. I'm not a huge niche writer and have received wide criticism from the HubPages community for my articles not performing as well as they could because I don't use enough keywords within the text of my articles. I prefer to write more naturally in a way that would appeal to an audience instead of keyword stuff. Anyway, this is aside the point. My Question After removing my articles from HubPages, should I put them on one domain or spread them across multiple domains grouped sort of by topic. For example: a-bunch-of-articles.com OR travel-articles.com and financial-articles.com and knitting-articles.com (I know those domains aren't available, but it's just kind of an example.) Here are the pros and cons of each: a mixed bag site like a-bunch-of-articles.com may not perform as well because of its mixed-bag nature a mixed bag site would be updated far more frequently than several niche sites... some niche sites may be updated so infrequently that a year could pass before one sees a new article a mixed bag site would be like putting all my eggs is one basket, where having several niche sites would spread out my portfolio, so to speak. a mixed bag site would be cheaper, $14 (two year registration) to start out with and hosting and I'm good to go. a mixed bag site wouldn't allow me to easily target keywords, but then again isn't HubPages pretty much a mixed bag site?

    Read the article

  • Is there any negative impact with similar page titles and descriptions on similar sites?

    - by ElHaix
    Currently we have Canadian versions of some websites. We are going to create some American versions, which essentially have everything the same, except the search results are geo-specific to the USA. The format for the results page title and descriptions will remain the same, ie {0} in {1} | Find more {0} etc etc etc... {1}. The search term will most-likely be the same between both sites. Will the relative similarity in the page titles and descriptions between the CDN and USA sites have any negative SEO impact, where the geo location would be the most significant difference?

    Read the article

  • JavaScript serait le langage de programmation le plus populaire selon l'indice RedMonk, basé sur GitHub et les sites d'entraide

    JavaScript est le langage de programmation le plus populaire Dans le classement RedMonk, basé sur GitHub et les sites d'entraide Entre popularité et préférence, quel est le meilleur critère pour classer les langages de programmation ? Le bureau d'analyse RedMonk les classe en fonction de leur popularité au sein de la communauté des développeurs sur les sites GitHub et StackOverflow. Le classement de RedMonk révèle un tout autre résultat comparé au classement par l'indice de TIOBE que nous relevons périodiquement sur Developpez.com. En effet, chez RedMonk on retrouve JavaScript en tête de liste, suivi de près par Java, PHP...

    Read the article

  • My Sites Were Hacked. What To Do?

    - by Vad
    I host multiple domains with this very popular hosting provider and I just went into one of my sites and... I see a black page with message "Hacked by...". I checked and all my sites with the provider are showing this same page. Inside of file system I have seen the hacker placed all default.* and index.* files with this message. So the hacker overwrote all index pages, placed new pages and that is under every, I say again, every folder. Cleaning this up will be close to a most horrible job. What to do (right now I am awaiting the restore of files from hosting provider)? How to prevent this? Whom to blame?

    Read the article

  • SmS Gateways - How do other sites do it? [closed]

    - by chobo2
    Possible Duplicate: Send and Receive SMS from my Website I would love to have a feature on my site that sends Email reminders and SmS(text messages) to people mobile phones. I been searching around and all I am finding is api's that charge money per SmS message(as low as 1cent per message). However even at 1cent per message that is still too much. The amount of money I am charging per year could be servilely eroded by just the Sms messages along. I could of course charge more money for my service or have an add on for SmS messages but I don't think either would work as most people expect it to be free feature and if they have to pay anything that is because of their carrier charging them not the website. How do other sites do it? I guessing companies like google have their own gateway providers or something like that. But how about smaller sites what do they do? I can't see them paying per sms text message.

    Read the article

  • How can I create multiple mini-sites with similar/duplicate content without hurting my search engine rank?

    - by ekpyrotic
    Essential background: I run a small company that lets members of the public post handwritten letters to their local politician (UK-based). Every week a number of early stage bills (called Early Day Motions) are submitted for debate in the House of Commons, and supporters of the motion will contact their local Members of Parliament, asking them to sign the motion. The crux: I want to target these EDMs with customised mini-sites, so when people search "EDM xxx", they find my customised mini-site, specifically targeting that EDM (i.e., "Send a handwritten letter to your MP asking them to sign EDM xxx"). At the moment, all these mini-sites (and my homepage) have duplicate content with only the relevant EDM name, number, and background image changed. (For example, http://mailmymp.com and http://mailmymp.com/edm/teaching-life-saving-skills-at-school-edm-550.php). The question: Firstly, will this hurt my potential search engine ranking? And, if so, what's the best way to target these political campaigns in an efficient manner without hurting my SEO prospects?

    Read the article

  • Is there a (free) reliable place to get statistics from sites, more reliable than Alexa, Quantcast, Compete?

    - by S.gfx
    I mean, seems there's no way. I am just asking in case someone knows of a recent new site being more accurate. I am aware of Alexa's, Compete and Quantcast inaccuracies and/or limited system/range of sites to get their stats. I also know about websitegrader perhaps being a little more accurate (although not sure if that's the data I am after). And read Seomoz tools are reliable. I am yet though looking for a free solution, a 'reliable' Alexa. And not a place depending on a toolbar installation, an easy to trick place, or one with stats way too off, or of a very limited range of sites. I am almost sure there's nothing new, but I wanted to be sure.

    Read the article

  • Google lance Tag Manager, un outil gratuit qui facilite la gestion du suivi et des balises de marketing des sites Web

    Google lance Tag Manager un outil gratuit qui facilite la gestion du suivi et des balises de marketing des sites Web Google vient d'annoncer le lancement de Google Tag Manager, son nouvel outil pour la gestion des différentes balises dans un site Web. Pour mieux monétiser leur site Web et contrôler la manière dont le contenu est utilisé, les gestionnaires de sites ont recours à des outils de suivi de statistiques comme Google Analytics. Pour chaque service, un morceau de code doit être intégré dans chaque page du site. Bien que d'une utilisation relativement simple, la multiplication de ces bouts de code sur une page peut rendre leur gestion fastidieuse. De plus, les requêtes entr...

    Read the article

  • Google va renforcer ses mesures anti-cloacking, un changement important qui pourrait impacter les sites mobiles et les contenus riches

    Google va renforcer ses mesures anti-cloacking Un changement important qui pourrait impacter les sites mobiles et les contenus riches Comme résolution du nouvel an, Google a décidé de resserrer l'étau sur le « cloaking », cette pratique qui consiste à présenter aux moteurs de recherche un contenu différent de celui présenté à l'utilisateur. Souvent utilisée à des fins de spam, cette technique est réprimée par Google et les autres moteurs de recherche. Pourtant, de nombreux webmasters la trouvent utile pour améliorer le référencement de sites conçus avec des technologies encore mal prises en compte par les bots des moteurs de recherche, comme le Flash ou le Rich Medi...

    Read the article

  • What is the most secure environment for multiple CMS sites? [closed]

    - by Brian Gulino
    I wish to run about 50 Joomla or WordPress low-traffic websites on 1 server, or part of a server. Each website will be managed by its own, naive owner who will have be able to access the Joomla or Wordpress backend of the website. I am concerned about security and isolation as my users will periodically get into trouble by not protecting their sites properly. Two alternatives I know of exist: Run one Linux system with multiple websites under Apache. Follow current Joomla and WordPress security tips. Increase the isolation of the individual sites by using mpm-itk, which will allow each website to run as its own user. The alternative to this is to run virtualization software such as the Xen hypervisor. Each site would have its own, virtual Linux system. I lack the experience needed to make this decision and I am asking which path to take. Obviously, there may be other alternatives that I haven't considered.

    Read the article

  • A quel point l'Internet est-il étendu ? Il compterait actuellement 255 millions de sites et 1.97 milliards d'utilisateurs

    A quel point l'Internet est-il étendu ? Il compterait actuellement 255 millions de sites et 1.97 milliards d'utilisateurs La compagnie technologique Pingdom s'est penchée sur la taille de l'Internet, et sur les évènements qui l'ont atteint en 2010. A l'aide de statistiques révélées publiquement par des entreprises spécialisées, ainsi qu'en s'appuyant sur des recherches qu'elle a réalisée elle-même, Pingdom vient de publier des chiffres impressionnant sur le Web. De quoi avoir le vertige. Ainsi, 107 trilliards d'e-mails ont été échangés l'année dernière. Actuellement, la Toile compterait 255 millions de sites Internet pour 88.8 millions de domaines en ".com". 1.97 milliard d'êtres humains seraient des intern...

    Read the article

  • Chrome 10 bogue lorsqu'il s'agit de lire les sites en Flash, êtes-vous touchés par ce problème ?

    Chrome 10 bogue lorsqu'il s'agit de lire les sites en Flash, êtes-vous touchés par ce problème ? Mise à jour du 16.03.2011 par Katleen La semaine dernière, Google a mis son navigateur à jour. Mais la version 10 de Chrome semble être touchée par un bogue. Plusieurs internautes se plaignent en effet de problèmes lors de la consultation de sites en Flash, surtout lors d'affichages simultanés de pages de ce type. A ce moment là, le plugin crasherait en affichant le message suivant : "The following plugin has crashed: Shockwave Flash". Malgré tout, le navigateur continue de fonctionner normalement (sauf pour les pages Flash). Une survenue regrettable alors que Microsoft vient de s...

    Read the article

  • Choose Your Boss : un tiers de recruteurs en plus par mois, le nouveau site d'emploi IT inspiré des sites de rencontres trouve son public

    Choose Your Boss : un nouveau site d'emplois IT qui s'inspire des sites de rencontres Plus de deux cent offres qualifiées de postes disponibles Choose Your Boss est un site original. Il s'inspire de Meetic et autres Attractive World pour mettre en relation développeurs et professionnels de l'IT d'une part et recruteurs d'autre part. Créé par Laurent Chollat-Namy, qui bénéficie d'une expérience professionnelle de 15 ans dans l'IT, celui-ci explique : « Nous avons longuement discuté avec des recruteurs et des informaticiens pour identifier leurs besoins et leurs pratiques. Cette réflexion nous a amenés à réinventer la mise en relation candidat / recruteur en nous inspirant des sites de r...

    Read the article

  • Choose Your Boss : un nouveau site d'emplois IT qui s'inspire des sites de rencontres, plus de 200 offres qualifiées

    Choose Your Boss : un nouveau site d'emplois IT qui s'inspire des sites de rencontres Plus de deux cent offres qualifiées de postes disponibles Choose Your Boss est un site original. Il s'inspire de Meetic et autres Attractive World pour mettre en relation développeurs et professionnels de l'IT d'une part et recruteurs d'autre part. Créé par Laurent Chollat-Namy, qui bénéficie d'une expérience professionnelle de 15 ans dans l'IT, celui-ci explique : « Nous avons longuement discuté avec des recruteurs et des informaticiens pour identifier leurs besoins et leurs pratiques. Cette réflexion nous a amenés à réinventer la mise en relation candidat / recruteur en nous inspirant des sites de r...

    Read the article

  • Choose Your Boss : un nouveau site d'emplois IT qui s'inspire des sites de rencontres, plus de 200 offres qualifiées

    Choose Your Boss : un nouveau site d'emplois IT qui s'inspire des sites de rencontres Plus de deux cent offres qualifiées de postes disponibles Choose Your Boss est un site original. Il s'inspire de Meetic et autres Attractive World pour mettre en relation développeurs et professionnels de l'IT d'une part et les recruteurs d'autre part. Créé par Laurent Chollat-Namy, qui bénéficie d'une expérience professionnelle de 15 ans dans l'IT, celui-ci explique : « Nous avons longuement discuté avec des recruteurs et des informaticiens pour identifier leurs besoins et leurs pratiques. Cette réflexion nous a amenés à réinventer la mise en relation candidat / recruteur en nous inspirant des sites ...

    Read the article

  • How to Configure IIS 7.5 to Host Several Classic ASP Sites using 1 IP Address?

    - by SidC
    We are running Windows Server 2008 R2 with IIS 7.5 and have 4-5 Classic ASP sites to install. The main site is stored in wwwroot and the other sites are stored in folders below wwwroot. We have 1 IP address for the server. How do I configure IIS to allow folks to browse/test the sites before domain names are pointed to te server? When I setup one of the sites in a subfolder of wwwroot and assign a separate port to it, I receive a message stating: Config Error Cannot add duplicate collection entry of type 'add' with unique key attribute 'value' set to 'index.asp' How do I remedy this error and permit IIS to render the site?

    Read the article

  • Can I use a Google Appliance/Mini to crawl and index sites I don't own?

    - by SkippyFire
    Maybe this is a stupid question, but... I am working with this company and they said they needed to get "permission" to crawl other people's sites. They have a Google Search Appliance And some Google Minis and want to point them at other sites to aggregate content. The end result will be something like a targeted search engine. (All the indexed sites relate to a specific topic) The only thing they will be doing is: Indexing Content from the other sites/domains Providing search functionality on their own site that searches the indexed content (like Google, displaying summaries and not the full content) The search results will provide links back to the original content Their intent is not malicious in nature, and is to provide a single site/resource for people to reference on their given topic. Is there anything illegal or fishy about this process?

    Read the article

  • What URLs must be in IE's Trusted Sites list to allow Windows Update?

    - by ewall
    Particularly for servers in which IE Enhanced Security Configuration is enabled, you need to have all the Windows Update/Microsoft Update URLs in your "Trusted Sites" list in order to use the site. (Furthermore, for domain member servers where Group Policy enforces Internet Explorer's list of "Trusted Sites", you don't have the option to edit the Trusted Sites yourself... so all the necessary URLs should be listed in the GPO.) So, what is the full list of URLs I'll need in IE's Trusted Sites? So far I have the following: http(s)://*.update.microsoft.com http://download.windowsupdate.com http://windowsupdate.microsoft.com I seem to remember there being several more...

    Read the article

  • web sites leaking memory? IIS 7.5 Windows server 2008 R2

    - by Charles
    I have several web sites on my windows 2008 server that have been working flawlessly for over a year. Just a few days ago I ran into an issue where my server stopped serving up pages on some of these sites for no apparent reason. I dug into it a little more today and I see that some of my sites (they're all asp.net mvc 3.0 sites), are consuming over 460MB of memory. Like I said, this just started the other day after a very long period of time of no issues at all. I have two questions: 1) is there a way to throttle how much memory is consumed by the w3wp process before I can force it to restart (restart the app pool for a particular site) so that it doesn't keep hogging all of the memory? 2) any ideas what could have caused this to start happening?

    Read the article

  • Does hosting multiple sites on one server hurt your SEO?

    - by MarathonStudios
    I have a handful of (content unrelated) sites with decent PRs and I'm considering hosting them all on the same server. I've heard that if you do this, internal linking between two seperate domains on that server may be seen as less "valid" by Google in PageRank terms (since you obviously own both of the sites as they share an IP address). Anyone have any experience in this? I'd love to save some hosting cash by consolidating, but not at the expense of losing the ability to link my sites together powerfully.

    Read the article

  • I'm seeing a lot of a new format for title elements for major sites: a few tags, and only then the title of the site itself, is this now optimal SEO?

    - by Tchalvak
    I'm seeing a lot of sites go to a format similar to the one now in use by stackexchange sites (i.e. short topic - maybe a long topic - only finally the site's name). Is this optimal SEO? It seems kind weird both because I'm only begun to notice changes in this direction recently, and because it seems like it would make it harder to tell in search results which actual site you're visiting, even if the topic is one that matches. Still, sites like Facebook, StackOverflow, etc probably aren't wrong, so I'm wondering if I should try to make my sites use that format, going forwards...

    Read the article

  • Où se situent les sites web les plus populaires au monde ? Un rapport révèle qu'ils sont concentrés aux États-Unis, la France ferme le top 5

    Où se situent les sites web les plus populaires au monde ? Un rapport révèle qu'ils sont concentrés aux États-Unis, la France ferme le top 5 Un récent rapport de la firme Host Cabi réalisé avec un outil gratuit de géolocalisation de sites web les plus populaires (classés par Alexa), révèle l'emplacement de ceux-ci.Le rapport intitulé « Où sont hébergés les 100 000 sites web les plus populaires au monde ? » permet de se rendre compte qu'ils sont partagés entre 1 204 hébergeurs de 621 villes de...

    Read the article

< Previous Page | 19 20 21 22 23 24 25 26 27 28 29 30  | Next Page >