Search Results

Search found 2724 results on 109 pages for 'spam filtering'.

Page 57/109 | < Previous Page | 53 54 55 56 57 58 59 60 61 62 63 64  | Next Page >

  • Why do some open source projects do not allow to report issues in a public issue tracker?

    - by linquize
    Why do some open source projects do not allow to report issues in a public issue tracker? Those projects requires the issues to be reported via email, and the issues may be forwarded to people in mailing list. Users may repeatedly report the same issue if there is no public issue tracker, as they have no easy way to know what have been reported before. The project team members need to spend extra time answering those repeated issues. Some projects do have a public issue tracker, but the issues are still reported through email and they are posted only by the project team only after filtering. It does not allow anyone to report directly in issue tracker. (example: SVN) Such arrangement is not transparent nor open, which I think it violates the philosophy of open source. And it is outdated.

    Read the article

  • Will this sitemap get me de indexed from Google?

    - by heavy rocker dude
    My site's URL (web address) is: http://quantlabs.net/private/sitemap.xml Description (including timeline of any changes made): Will this sitemap get me de-indexed from Google? My new site map just got spidered by Google for some reason. It is located at http://quantlabs.net/private/sitemap.xml, is this in danger of getting me de-indexed from Google's index. Does it look like spam even though it is not meant to be? I am trying to figure the limitation in terms of Google's threshold before they deem it a spammy sitemap. This is sitemap contains automated postings which are different with the stock symbol provided. The amount of postings within the Sitemap are quite a few in a small amount of time.

    Read the article

  • What is the general definition of something that can be included or excluded?

    - by gutch
    When an application presents a user with a list of items, it's pretty common that it permits the user to filter the items. Often a 'filter' feature is implemented as a set of include or exclude rules. For example: include all emails from [email protected], and exclude those emails without attachments I've seen this include/exclude pattern often; for example Maven and Google Analytics filter things this way. But now that I'm implementing something like this myself, I don't know what to call something that could be either included or excluded. In specific terms: If I have a database table of filter rules, each of which either includes or excludes matching items, what is an appropriate name of the field that stores include or exclude? When displaying a list of filters to a user, what is a good way to label the include or exclude value? (as a bonus, can anyone recommend a good implementation of this kind of filtering for inspiration?)

    Read the article

  • Displaying a Sorted, Paged, and Filtered Grid of Data in ASP.NET MVC

    Over the past couple of months I've authored five articles on displaying a grid of data in an ASP.NET MVC application. The first article in the series focused on simply displaying data. This was followed by articles showing how to sort, page, and filter a grid of data. We then examined how to both sort and page a single grid of data. This article looks at how to add the final piece to the puzzle: we'll see how to combine sorting, paging and filtering when displaying data in a single grid. Like with its predecessors, this article offers step-by-step instructions and includes a complete, working demo available for download at the end of the article. Read on to learn more! Read More >

    Read the article

  • Noisy audio recording in Ubuntu Linux

    - by Haresh K Miriyala
    I'm unable to record audio without noise on my Ubuntu laptop . It is not a grounding problem . My windows os is able to record voice without the static noise so, my hardware is fine . I read in the Windows forums that the os has software noise filtering mechanism that's automatically enabled. How do I do the same in Ubuntu? If not, please let me know how to filter the static noise. Any help is appreciated. Thanks!

    Read the article

  • Crafting an effective php/web programmer job post template [closed]

    - by Tchalvak
    I am looking to create a job post to get a satisfactory assistant programmer / templater. Specifically, a php & web programmer. I am, however, afraid of forgetting important things. So, are there resources you can suggest for templates for things to ask and things to tell in a job post for a programmer? Surprisingly, I wasn't able to find similar questions on this site, so there may be duplicate questions out there that I could use but just didn't find. Right now I know that my -requirements- are so generic that they're going to get me in trouble with a spam of applications. e.g. the candidate must know php, must be able to seperate php from html. So I'm looking for criteria that are must-haves, must-mentions, or a general template to try to avoid a "lemon". I also started a gist to work on a job post, comments/edits would be excellent: https://gist.github.com/2906808

    Read the article

  • Point me to info about constructing filters (of lists)

    - by jah
    I would like some pointers to information which would help me understand how to go about providing the ability to filter a list of entities by their attributes as well as by attributes of related entities. As an example, imagine a web app which provides order management of some kind. Orders and related entities are stored in a relational database. And imagine that the app has an interface which lists the orders. The problem is: how does one allow the list to be filtered by, for example:- order number (an attribute) line item name (an attribute of a n-n related entity) some text in an administrative note related to the order (text found in an attribute of a 1-1 related entity) I'm trying to discover whether there is something like a standard, efficient way to construct the queries and the filtering form; or some possible strategies; or any theory on the topic; or some example code. My google foo fails me.

    Read the article

  • Why do some open-source projects NOT have a public issue tracker?

    - by linquize
    Why do some open source projects not allow to report issues in a public issue tracker? Those projects require the issues to be reported via email, and the issues may be forwarded to people in a mailing list. Users may repeatedly report the same issue if there is no public issue tracker, as they have no easy way to know what has been reported before. The project team members need to spend extra time answering those repeated issues. Some projects do have a public issue tracker but the issues are still reported through email and they are posted by the project team only after filtering. It does not allow anyone to report directly in issue tracker (example: SVN). Such arrangement is not transparent nor open, which I think violates the philosophy of open source. And it is outdated.

    Read the article

  • Easy QueryBuilder - A User-Friendly Ad-Hoc Advanced Search Solution

    Constructing an easy and powerful QueryBuilder interface becomes more important for complex data grid filtering and accurate reporting services. In this article, I'll discuss how to build a query search engine using ASP.NET AJAX and dynamic SQL. The main goal is to provide an interactive interface to allow users select query attributes, operators, attribute values, and T-SQL operators so that the data context query list can be easily composed and a search engine is invoked.Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Geolocation Blocking and SEO

    - by Mahesh
    I'm in process to block a particular geolocation from accessing my site. But i am not sure if this is going to raise any flags on my website. Basically i'm getting lots of spam from particular state and there are some webmaster copying my content without any attribution or backlink. I feel that is enough and i wish to save my bandwidth on that state or geolocation. So my question is how to block particular geolocation ? Do you have any suggestions for php scripts that help block particular geolocation ? and another question: if there are any SEO disadvantages ?

    Read the article

  • What You Said: How You Deal with Bacn

    - by Jason Fitzpatrick
    Earlier this week we asked you how you deal with Bacn—email you want, but not right now—and you responded. Read on to see the three principle ways HTG readers deal with Bacn. The approach you all took fell into three distinct categories: Filtering, Obfuscating, and Procrastinating. Readers like Ray and jigglypuff use filters: I use Thunderbird as my email client. I have different folders that I filter the email I receive into. The newsletters and other subscribed emails go into a lower priority folder. One word: Filters. I just setup filters for all of this type of mail. Some I let go to inbox, others I let go straight to a folder without seeing it first. Then when I have time or want to go through them, I do. HTG Explains: Is ReadyBoost Worth Using? HTG Explains: What The Windows Event Viewer Is and How You Can Use It HTG Explains: How Windows Uses The Task Scheduler for System Tasks

    Read the article

  • SEO penalty for landing page redirects

    - by therealsix
    Using ebay as an example- lets say I have a large number of items whose URLs' look like this: cgi.ebay.com/ebaymotors/1981-VW-Vanagon-manual-seats-seven-/250953153841 I want to give my client the ability to put links to these items on their website EASILY, without knowing or checking my URL. So I created a redirect service that will map their identifier with my URL: ebay.com/fake_redirect_service/shared_identifier9918 would redirect to the link above. This works great- my clients can easily setup these links with information they already have, and the user will see the page as usual. So on to the problem... I'm concerned that this redirecting service will have a negative impact on my SEO ranking. Having a landing page redirect you immediately to a different URL seems like something a typical spam site would do. Will this hurt me? Any better solutions?

    Read the article

  • CommonFilter 0.3D now released on CodePlex.

    CommonFilter is a subset of the CommonData project, containing just the functions and unit tests for filtering user input. The functions include filters for: Input of upper case and lower case alpha Input of numeric fields Input of text containing HTML markup to check that it only contains permitted markup. The general functions are available both as a form that silently drops non-permitted characters or in a try-parse format....Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • CommonFilter 0.3D now released on CodePlex.

    CommonFilter is a subset of the CommonData project, containing just the functions and unit tests for filtering user input. The functions include filters for: Input of upper case and lower case alpha Input of numeric fields Input of text containing HTML markup to check that it only contains permitted markup. The general functions are available both as a form that silently drops non-permitted characters or in a try-parse format....Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Google ajuste son algorithme pour contrer spams et contenus dupliqués et rejette les critiques sur la qualité de ses résultats

    Google ajuste son algorithme pour contrer spams et contenus dupliqués Et rejette les critiques sur la qualité de ses résultats En réponse aux critiques acerbes ayant récemment remis en question la qualité de ses résultats de recherche, Google entreprend de mettre à jour son algorithme pour contrer les sites de spam et de contenu dupliqué. Ces mises à jour devraient s'attaquer aux sites qui fraudent ou enfreignent les lignes de conduite tracées par Google et ce dans le but de s'accaparer les premières places des résultats. Sont particulièrement visés ceux qui copient le contenu des autres sites et/ou qui disposent d'un taux faible en contenu original. Matt Cutts, en ...

    Read the article

  • How to set up a mail server on Linux only for sending admin/debug emails?

    - by ChocoDeveloper
    I need to send server reports to myself from my remote servers, and I don't mind them going to spam, so I don't need SPF, DKIM, etc. I tried using mailutils to send something like this: uptime | mail -s "uptime" [email protected], but the emails don't go through. In /var/mail/root I received a message saying the HELO was invalid. So then I tried also adding -r root@my-ip, and now I don't get any error messages but I don't receive the email either. How can I do this?

    Read the article

  • Receive anonymous users' input by web upload form or email. Any online service for that?

    - by sja
    Are you aware of any online service or online "platform" allowing users, not previously registered, to upload pairs of picture+comment to a database? It would be a collaborative database of picture+comment pairs. I'm not going wiki or googlegroup, picasa or such because I'd like the user to have the least to do to participate, that is e.g.: take a picture with his phone and email it to an email to an email address. Or go to a web page with an upload form, type in a description, hit OK and that's it. And the goal is also that it be as hassle-less to put up as possible. Yeah I know, it can't programme itself to my requirements :) by I'm suspecting there's a tool or tool combination going a decent way through my needs. Thanks for any info/advice! SJA (NB the final goal is a kind of crowd-sourced census of specific urban items. If you have comment about the potential for spam-overload of my idea, other than "you're doomed", you're welcome!)

    Read the article

  • Where can I find SPF record information for Network Solution hosted email?

    - by Steve Brown
    I have a client who has their e-mail hosted with Network Solutions and some of the e-mails they are sending are going to spam. I've searched Google for information on network solutions and SPF records but all the results have to do with "how to set up an SPF record" - which I understand how to do. The problem is that I can't find any information about what SPF records to use for network solutions hosted e-mail. I even tried lookup up spf records on networksolutions.com but it appears there are none. Where can I find SPF record information for Network Solution hosted email?

    Read the article

  • Website with over 1 million posts with not much textual content

    - by Far Se
    I've made a website which crawls files from all over the Internet and I feel like Google will ban me if I sent it sitemaps which contain all of these pages (1m+), because they contain only the file name/size/no of downloads and the download link(s). I'm considering this thought because I've made another website like this in the past and Google banned me after one week with the reason: "spam", even it was not (maybe somebody falsely reported me?!). Does someone have an idea about how to keep Google form banning my website? I've seen several other sites like mine and they don't get banned or... anything. And also, should I sent the sitemap or wait until Google indexes every page as it finds them? Thanks in advance :)

    Read the article

  • Another website is mirroring and ranks above my site in search results

    - by Marlboro Goodluck
    There is a site of ill-repute known as thedirty which has completely mirrored my site and now has links appearing on Google at the #1 spot using my content. I checked my log files and noticed that this site has been crawling mine for sometime, and also has 10,000 links from their site to mine. I have blocked user access which is referred from this site and reported them as web spam to Google already. I also disavowed the domain. How are they getting top links in Google (even overtaking mine) for such nefarious tactics? What are the steps to completely eliminating an issue such as this?

    Read the article

  • Facebook corrige un bogue susceptible d'avoir provoqué une fuite de données de six millions d'utilisateurs

    Facebook corrige un bogue susceptible d'avoir provoqué une fuite de données de six millions d'utilisateursSi vous avez reçu ce weekend un courriel de la part de Facebook vous expliquant que votre compte a été compromis suite à un bogue, sachez que ce n'est pas un SPAM.En effet, vendredi dernier, Facebook a voulu jouer la carte de la transparence en annonçant dans un billet blog avoir été victime d'une panne logicielle qui a été à l'origine de la fuite de données de près de six millions d'utilisateurs. Download Your Information (DYI) est l'outil qui a provoqué le dysfonctionnement. Il est chargé de récupérer les adresses de courriel et numéros de téléphone des utilisateurs dans le cadre de la sécu...

    Read the article

  • Google va renforcer ses mesures anti-cloacking, un changement important qui pourrait impacter les sites mobiles et les contenus riches

    Google va renforcer ses mesures anti-cloacking Un changement important qui pourrait impacter les sites mobiles et les contenus riches Comme résolution du nouvel an, Google a décidé de resserrer l'étau sur le « cloaking », cette pratique qui consiste à présenter aux moteurs de recherche un contenu différent de celui présenté à l'utilisateur. Souvent utilisée à des fins de spam, cette technique est réprimée par Google et les autres moteurs de recherche. Pourtant, de nombreux webmasters la trouvent utile pour améliorer le référencement de sites conçus avec des technologies encore mal prises en compte par les bots des moteurs de recherche, comme le Flash ou le Rich Medi...

    Read the article

  • Is it okay for an application to check for automatic updates in less than 20 hour interval?

    - by FlameStream
    I have a desktop application that has the ability to automatically update itself on the next restart (without the user's consent - but this is another issue altogether). Assuming that the user would never notice anything related to application updating (such as a progress bar, or pop-up requiring restart), and that our server would support the request spam load, is there any reason why it should not check for updates in less than 20 hour interval? The reason I'm asking this is because all applications that I know that have auto-update capability check for update every 20 to 24 hours and at startup. I was just wondering if there was an ethical rule about it, or simply because of the risk of overloading the server.

    Read the article

  • Ubuntu Audio Pitch Shifting filter

    - by user777305
    I'm currently developing a video player software which intended to be an embeedded player. I'm using Java with VLCJ library for the video player. What i'm looking is a way, something to transform the audio output to make the output sound as oldman or a kid (pitch shifting, i guess is the name). VLC have this when enabling time stretch, but the video play speed is affected (slower to get oldman sound, but fast-forward-alike to get kid voice effect. Is there any solutions for this? I don't find this feature on VLC(J), so i think what i need is the audio output of the ubuntu itself (Ubuntu 12.04) to do this job. something like filtering audio output system wide. Any software or setting to do this? it also need to be controllable via command line to provide realtime effect changing. Thanks in advance.

    Read the article

  • Another website is mirroring my site

    - by Marlboro Goodluck
    Question for you all. There is a site of ill repute known as thedirty which has completely mirrored my site and now has links appearing on Google at the #1 spot using my content. I checked my log file and noticed that this site has been crawling mine from sometime, and also has 10k links from their site to mine. I have blocked user access which is referred from this site and reported them as web spam to Google already. I also disavowed the domain. How are they getting top links in Google (even overtaking mine) for such nefarious tactics? What are the steps to completely eliminating an issue such as this?

    Read the article

< Previous Page | 53 54 55 56 57 58 59 60 61 62 63 64  | Next Page >