Search Results

Search found 2676 results on 108 pages for 'spam blocking'.

Page 57/108 | < Previous Page | 53 54 55 56 57 58 59 60 61 62 63 64  | Next Page >

  • Will this sitemap get me de indexed from Google?

    - by heavy rocker dude
    My site's URL (web address) is: http://quantlabs.net/private/sitemap.xml Description (including timeline of any changes made): Will this sitemap get me de-indexed from Google? My new site map just got spidered by Google for some reason. It is located at http://quantlabs.net/private/sitemap.xml, is this in danger of getting me de-indexed from Google's index. Does it look like spam even though it is not meant to be? I am trying to figure the limitation in terms of Google's threshold before they deem it a spammy sitemap. This is sitemap contains automated postings which are different with the stock symbol provided. The amount of postings within the Sitemap are quite a few in a small amount of time.

    Read the article

  • Usefulness of the Backlinks shown in Webmaster Tools

    - by Ewan Heming
    Is the list of links for a site shown in Google Webmaster Tools a complete list or just a sample? I've noticed that the links in there appear to be all the ones I didn't think would have any real value - either because they were nofollow or from irrelivant sites. The few I did think would be some use have never shown up and there's also some links that are sometimes there and sometimes not (such as my linkedIn profile). Does this mean that the missing links don't/no longer carry any value? It almost appears that the list is there for Google to either inform you about problems (there was a useful list there when someone tried to SPAM my site) or mis-imform you about which link-building strategies work or not (to keep people guessing about what works or not).

    Read the article

  • Phishing : une nouvelle technique se répand avec le HTML5, elle contourne le blacklistage des URL malicieuses

    Phishing : une nouvelle technique se répand avec le HTML5 Elle contourne le blacklistage des URL malicieuses Les spammeurs et autres cyber-escrocs se mettent eux aussi au HTML5 pour contourner les mesures anti-spam et anti-phishing de plus en plus répandues et efficaces des navigateurs et les clients de messagerie. Au lieu d'intégrer aux mails des liens HTML classiques vers des pages souvent blacklistées, les spammeurs "modernes" privilégieraient désormais les « attachements HTML ». M86, la firme de sécurité met en tout cas en garde contre la recrudescence de ces menaces. Les liens dans les mail pointent désormais vers des pages HTML jointe, qui contiennen...

    Read the article

  • The Most Important Person Is the One that Keeps Your PC Running [Comic]

    - by The Geek
    Fixing people’s computers usually makes them appreciate you more, though this might be a little too far. Latest Features How-To Geek ETC How To Create Your Own Custom ASCII Art from Any Image How To Process Camera Raw Without Paying for Adobe Photoshop How Do You Block Annoying Text Message (SMS) Spam? How to Use and Master the Notoriously Difficult Pen Tool in Photoshop HTG Explains: What Are the Differences Between All Those Audio Formats? How To Use Layer Masks and Vector Masks to Remove Complex Backgrounds in Photoshop Bring Summer Back to Your Desktop with the LandscapeTheme for Chrome and Iron The Prospector – Home Dash Extension Creates a Whole New Browsing Experience in Firefox KinEmote Links Kinect to Windows Why Nobody Reads Web Site Privacy Policies [Infographic] Asian Temple in the Snow Wallpaper 10 Weird Gaming Records from the Guinness Book

    Read the article

  • How best to take a users signature online? (UK law orientated) [closed]

    - by Ben Griffiths
    Not sure if this is the best place to ask, but I can't seem to find any of the other SE sites that would fit better (unless there's a law one?) I'm building an application that will replace an existing paper based form, and this form would normally be signed by the person filling it in. Looking around, it's hard to find a good definitive resource to explain what I can and cannot accept as far as a signature goes. It looks like some UK government online forms accept just your name typed into a box, but I've also heard you should back up with an email - so that process would be type name into a box along with providing an email address, send out an email, then make them click a link within the email to finally complete the verification. Involving email seems very long winded and leaves the system open to spam filters blocking emails, forgotten emails that just sit in inbox's etc. So, does anyone have any knowledge in this department? Personally, I'd love to just get them to type their name into a box and be done with it!

    Read the article

  • Les ordinateurs sous Linux enverraient proportionnellement plus de spams que les autres, d'après Sym

    Les ordinateurs sous Linux enverraient proportionnellement plus de spams que les autres, d'après Symantec Malgré le peu de parts de marché détenues par Linux, les ordinateurs équipés par ce système serait les plus gros expéditeurs de spam. Symantec et ses chercheurs experts en sécurité informatique ont scrupuleusement étudié les spams envoyés entre novembre 2009 et mars 2010 afin d'identifier le système d'exploitation installé sur l'ordinateur propulsant les pourriels. Cela a été possible en utilisant la méthode dite du fingerprint passif, qui consiste à analyser le trafic réseau d'un hôte à distance, ce qui révèle l'OS de cet hôte. Comme Windows se taille toujours la part du lion sur ce marché (avec p...

    Read the article

  • Error installing gPodder

    - by Ron Webb
    A few weeks ago (newbie alert!) I started using XUbuntu 12.04 with Xfce 4.8. I'm trying to install gPodder Podcast Client (see https://launchpad.net/~thp/+archive/gpodder). I've added the PPA via terminal commands as instructed. When I click the Install button in the Ubuntu Software Centre I get the following error: Package dependencies cannot be resolved This error could be caused by required additional software packages which are missing or not installable. Furthermore there could be a conflict between software packages which are not allowed to be installed at the same time. Details: The following packages have unmet dependencies: gpodder: Depends: python-webkit but it is not going to be installed What do I need to do? Just to make thing more complicated -- I'm not sure, but before I found the launchpad.net link, I think I may have tried to install gPodder from the default Ubuntu repositories (also unsuccessfully). There may be remnants of the previous attempt still installed, which may be blocking the new install. Where/how can I find them?

    Read the article

  • Does the Instant Preview in Google webmaster tools takes Robot.txt in account?

    - by rockyraw
    Is that the way to go If I want to visually see what the googlebot see? I'm trying to check a folder which I have just blocked in my robots.txt. If I fetch the folder as google bot, It fetches ok, so that doesn't tell me nothing about whether the block is working I know there's a tool to check for blocking, but it is dependent on the input of the robots.txt Therefore I've tried the Instant preview, and I don't get a preview for what the bot sees ("pre-render), so I think that means that it's because the robots.txt blocks it; however - I don't see the bot tried beforehand to access my updated robots.txt, so I'm not sure how does it know that this folder is blocked? (it does preview another new folder, that is not blocked)

    Read the article

  • Github Feed affecting my WordPress installation? [on hold]

    - by saul
    Any idea how this fork is affecting my site? I went to verify my website log stats, and realized this may be the cause of a strange redirect constantly happening on my WordPress installation. Here's a line I found on my log: 54.81.91.95 - - [07/May/2014:22:52:08 -0400] "GET /category/selfie/feed/ HTTP/1.1" 200 1826 "-" "feedzirra http://github.com/pauldix/feedzirra/tree/master" And this is the Github fork (or however these are called). https://github.com/feedjira/feedjira/tree/master Basically, I think everytime I update my categories, (selfie in this case), I get redirected to install.php. Probably by triggering some GET function on that feed. to the best of my knowledge, this feed parses all url with this structure, blocking them, kind of like a DDoS attack?? Any ideas how to go about it??

    Read the article

  • Internet keep dropping in Ubuntu 12.04

    - by Kokuryuu
    I am having an issue with my Ubuntu, I keep getting "server not found" errors when I try to go online, I have tried pinging my DNS and Goggle, and it seems like no packets are dropped, but still my Internet is not working Also my other PCs in the house have working Internet connection. I am sure there is no firewall active here, also nothing on my router is blocking, and I checked cables. I have no issues in the same PC in Windows. It is only in Ubuntu. Any ideas?

    Read the article

  • Crafting an effective php/web programmer job post template [closed]

    - by Tchalvak
    I am looking to create a job post to get a satisfactory assistant programmer / templater. Specifically, a php & web programmer. I am, however, afraid of forgetting important things. So, are there resources you can suggest for templates for things to ask and things to tell in a job post for a programmer? Surprisingly, I wasn't able to find similar questions on this site, so there may be duplicate questions out there that I could use but just didn't find. Right now I know that my -requirements- are so generic that they're going to get me in trouble with a spam of applications. e.g. the candidate must know php, must be able to seperate php from html. So I'm looking for criteria that are must-haves, must-mentions, or a general template to try to avoid a "lemon". I also started a gist to work on a job post, comments/edits would be excellent: https://gist.github.com/2906808

    Read the article

  • Where can I find SPF record information for Network Solution hosted email?

    - by Steve Brown
    I have a client who has their e-mail hosted with Network Solutions and some of the e-mails they are sending are going to spam. I've searched Google for information on network solutions and SPF records but all the results have to do with "how to set up an SPF record" - which I understand how to do. The problem is that I can't find any information about what SPF records to use for network solutions hosted e-mail. I even tried lookup up spf records on networksolutions.com but it appears there are none. Where can I find SPF record information for Network Solution hosted email?

    Read the article

  • How to set up a mail server on Linux only for sending admin/debug emails?

    - by ChocoDeveloper
    I need to send server reports to myself from my remote servers, and I don't mind them going to spam, so I don't need SPF, DKIM, etc. I tried using mailutils to send something like this: uptime | mail -s "uptime" [email protected], but the emails don't go through. In /var/mail/root I received a message saying the HELO was invalid. So then I tried also adding -r root@my-ip, and now I don't get any error messages but I don't receive the email either. How can I do this?

    Read the article

  • SEO penalty for landing page redirects

    - by therealsix
    Using ebay as an example- lets say I have a large number of items whose URLs' look like this: cgi.ebay.com/ebaymotors/1981-VW-Vanagon-manual-seats-seven-/250953153841 I want to give my client the ability to put links to these items on their website EASILY, without knowing or checking my URL. So I created a redirect service that will map their identifier with my URL: ebay.com/fake_redirect_service/shared_identifier9918 would redirect to the link above. This works great- my clients can easily setup these links with information they already have, and the user will see the page as usual. So on to the problem... I'm concerned that this redirecting service will have a negative impact on my SEO ranking. Having a landing page redirect you immediately to a different URL seems like something a typical spam site would do. Will this hurt me? Any better solutions?

    Read the article

  • Google ajuste son algorithme pour contrer spams et contenus dupliqués et rejette les critiques sur la qualité de ses résultats

    Google ajuste son algorithme pour contrer spams et contenus dupliqués Et rejette les critiques sur la qualité de ses résultats En réponse aux critiques acerbes ayant récemment remis en question la qualité de ses résultats de recherche, Google entreprend de mettre à jour son algorithme pour contrer les sites de spam et de contenu dupliqué. Ces mises à jour devraient s'attaquer aux sites qui fraudent ou enfreignent les lignes de conduite tracées par Google et ce dans le but de s'accaparer les premières places des résultats. Sont particulièrement visés ceux qui copient le contenu des autres sites et/ou qui disposent d'un taux faible en contenu original. Matt Cutts, en ...

    Read the article

  • Receive anonymous users' input by web upload form or email. Any online service for that?

    - by sja
    Are you aware of any online service or online "platform" allowing users, not previously registered, to upload pairs of picture+comment to a database? It would be a collaborative database of picture+comment pairs. I'm not going wiki or googlegroup, picasa or such because I'd like the user to have the least to do to participate, that is e.g.: take a picture with his phone and email it to an email to an email address. Or go to a web page with an upload form, type in a description, hit OK and that's it. And the goal is also that it be as hassle-less to put up as possible. Yeah I know, it can't programme itself to my requirements :) by I'm suspecting there's a tool or tool combination going a decent way through my needs. Thanks for any info/advice! SJA (NB the final goal is a kind of crowd-sourced census of specific urban items. If you have comment about the potential for spam-overload of my idea, other than "you're doomed", you're welcome!)

    Read the article

  • Thread count in Java game

    - by Taylor Hill
    I'm just curious as to what a reasonable number of threads is for a simple 2D mmo in Java. Is it reasonable to have two threads per connection, one for the input stream and one for the output stream? The reason I ask is because I use a blocking method on the input stream, and a workaround seems unnecessarily complex if I were to try to get around it without adding threads. This is mostly for my own edification; I don't expect to have 5 million people playing it ever, or even 5, but I'm wondering what a good scalable solution is, and if this is reasonable for a small server (<30 connections).

    Read the article

  • Website with over 1 million posts with not much textual content

    - by Far Se
    I've made a website which crawls files from all over the Internet and I feel like Google will ban me if I sent it sitemaps which contain all of these pages (1m+), because they contain only the file name/size/no of downloads and the download link(s). I'm considering this thought because I've made another website like this in the past and Google banned me after one week with the reason: "spam", even it was not (maybe somebody falsely reported me?!). Does someone have an idea about how to keep Google form banning my website? I've seen several other sites like mine and they don't get banned or... anything. And also, should I sent the sitemap or wait until Google indexes every page as it finds them? Thanks in advance :)

    Read the article

  • Scalability of multi-threading in game server

    - by Taylor Hill
    What is a reasonable number of threads for a simple 2D mmo in Java? Is it reasonable to have two threads per connection, one for the input stream and one for the output stream? The reason I ask is because I use a blocking method on the input stream, and a workaround seems unnecessarily complex if I were to try to get around it without adding threads. This is mostly for my own edification; I don't expect to have 5 million people playing it ever, or even 5, but I'm wondering what a good scalable solution is, and if this is reasonable for a small server (<30 connections).

    Read the article

  • Facebook corrige un bogue susceptible d'avoir provoqué une fuite de données de six millions d'utilisateurs

    Facebook corrige un bogue susceptible d'avoir provoqué une fuite de données de six millions d'utilisateursSi vous avez reçu ce weekend un courriel de la part de Facebook vous expliquant que votre compte a été compromis suite à un bogue, sachez que ce n'est pas un SPAM.En effet, vendredi dernier, Facebook a voulu jouer la carte de la transparence en annonçant dans un billet blog avoir été victime d'une panne logicielle qui a été à l'origine de la fuite de données de près de six millions d'utilisateurs. Download Your Information (DYI) est l'outil qui a provoqué le dysfonctionnement. Il est chargé de récupérer les adresses de courriel et numéros de téléphone des utilisateurs dans le cadre de la sécu...

    Read the article

  • Release Notes 12/12/2012

    This past week the CodePlex team worked on several fixes to improve the stability of our TFS infrastructure, including applying TFS 2012 Update 1. We apologize for the recent downtime. We are not completely out of the woods, but we appreciate your patience as we work through the issues. Additional Bug Fixes: Fixed several issues with character encoding within file paths. Fixed issue where the number of pull requests and forks were disappearing after selecting either link. Fixed issue blocking license changes when special characters exist in copyright holder field. Have ideas on how to improve CodePlex? Please visit our suggestions page! Vote for existing ideas or submit a new one. As always you can reach out to the CodePlex team on Twitter @codeplex or reach me directly @mgroves84

    Read the article

  • Is it okay for an application to check for automatic updates in less than 20 hour interval?

    - by FlameStream
    I have a desktop application that has the ability to automatically update itself on the next restart (without the user's consent - but this is another issue altogether). Assuming that the user would never notice anything related to application updating (such as a progress bar, or pop-up requiring restart), and that our server would support the request spam load, is there any reason why it should not check for updates in less than 20 hour interval? The reason I'm asking this is because all applications that I know that have auto-update capability check for update every 20 to 24 hours and at startup. I was just wondering if there was an ethical rule about it, or simply because of the risk of overloading the server.

    Read the article

  • Alt text vs CSS sprites (SEO vs speed)

    - by leeoniya
    I'm reworking our site to reduce HTTP requests and blocking requests by concatenating JS, css, gzipping, loading all JS via LABjs and using CSS sprites for images that were loaded individually via <img> tags before. Progress has been great so far - 5x page load performance improvement. However, we're in the top 5 organic search ranking in google for many targeted keywords and phrases. I'm afraid eliminating so many img tags with alt attributes can hurt our SEO. Does anyone have any experience with alt tag manip/removal and effects on SEO positions? Is previous rank "sticky"?

    Read the article

  • jQuery scrolling images for e-commerce site, what to do about users who disable JS

    - by Livingston Storm
    As the title suggests, I am developing an e-commerce site and I intend of having two jQuery plug ins on the default page, one for scrolling images and the other for the navigation menu. Should I be concerned about making the site work if the users disables JS? Cause if they have it disabled my site would be almost impossible to use with the scrolling images blocking the main content. Plus the CMS I am using, Big Commerce, uses a bit of JS for the products pages, which would also look ridiculous with JS disabled. Anyone have experience with this?

    Read the article

  • Google va renforcer ses mesures anti-cloacking, un changement important qui pourrait impacter les sites mobiles et les contenus riches

    Google va renforcer ses mesures anti-cloacking Un changement important qui pourrait impacter les sites mobiles et les contenus riches Comme résolution du nouvel an, Google a décidé de resserrer l'étau sur le « cloaking », cette pratique qui consiste à présenter aux moteurs de recherche un contenu différent de celui présenté à l'utilisateur. Souvent utilisée à des fins de spam, cette technique est réprimée par Google et les autres moteurs de recherche. Pourtant, de nombreux webmasters la trouvent utile pour améliorer le référencement de sites conçus avec des technologies encore mal prises en compte par les bots des moteurs de recherche, comme le Flash ou le Rich Medi...

    Read the article

< Previous Page | 53 54 55 56 57 58 59 60 61 62 63 64  | Next Page >