Search Results

Search found 25614 results on 1025 pages for 'content filter'.

Page 562/1025 | < Previous Page | 558 559 560 561 562 563 564 565 566 567 568 569  | Next Page >

  • Google not showing any pages from my site in the index after three months [on hold]

    - by Alex Coisman
    Despite having a sitemap and using Google Webmaster Tools, it has been over 3 months and my site has not been added to the Google index at all. Here's the site: www.famouslefthandedpeople.com As far as I know, I have done everything correctly. However, there must be something I am overlooking that is preventing Google from indexing the site. I do not have a robots.txt file, so allow/disallow isn't the issue. Although the content of the site is sparse, it is original and not duplicated internally or externally so Panda/Penguin should not be a problem. I have reviewed the answers at Why isn't my website in Google search results? and I don't think it applies here. If it matters, I am using WordPress to create the site. What other factors should I be looking at in order to troubleshoot this?

    Read the article

  • How to copy the file from source to destination only once at a time?

    - by Viswa
    I have to copy the file from my desktop to my mounted directory. I was using the following command to copy the file from my desktop to mounted directory. os.system("cp -f /home/Desktop/filename /media/folder_1"). It works fine. But the problem is while copying the file from my source to mounted directory(folder_1) if any interruption is happens like network down, then the system continuously keep on trying. It couldn't skip that process. Finally, when the network comes the files are again copy to my mounted directory. Due to this continuous trying, next time i try to move the content it throws "permission denied" error. How do i copy the file only once, if any network issues happen then it will not keep try to copy, instead of that, it throws the error. If you know, Let me. Its very useful to me.

    Read the article

  • GlassFish Community Event @ JavaOne - Save the date!

    - by alexismp
    The interest for having a GlassFish community event at JavaOne is still very strong both inside Oracle and in the community, so this year again we'll be hosting a get together on the Sunday prior to the main event. If you're in town and attending JavaOne, mark your calendars : Sunday 2nd, October 2011 - 12:30pm-4:30pm in the Moscone This will be an opportunity to discuss the community status (adoption of Java EE 6, GlassFish 3.1.x) and hear about future plans, mainly around Java EE 7 and the related GlassFish release(s). We'd also like to have several participants share their deployment stories as well as some time for an free-form unconference format and some team building activity. Of course, beyond all the content shared in slides, this should really also be a good excuse to meet folks from the community and from the core GlassFish team at Oracle. Here's a post on last year's event. And before anybody asks, we are still exploring the party situation :-)

    Read the article

  • Why are new pages not being indexed and old pages stay in the index?

    - by ZakGottlieb
    I currently have a site that was recently restructured, causing much of its content to be reposted, creating new URL's for each page. To avoid duplicates, all of the existing pages were added to the robots file. That said, it has now been over a week - I know Google has recrawled the site - and when I search for term X, it is stil the old page that is ranking, with the new one nowhere to be seen. I'm assuming it's a cached version, but why are so many of the old pages still appearing in the index? Furthermore, all "tags" pages (it's a Q&A site, like this one) were also added to the robots a few months ago, yet I think they are all still appearing in the index. Anyone got any ideas about why this is happening, and how I can get my new pages indexed?

    Read the article

  • Who created that user?

    - by AaronBertrand
    Twitter has provided some great fodder for blog content lately. And twice this week, I've found an excuse to take advantage of the default trace. Tonight @meltondba asked: I'm trying to find who created a user act in a DB It is true, SQL Server doesn't really keep track of who created objects, such as user accounts in a database. You can get some of this information from the default trace, though, since it tracks EventClass 109 (Audit Add DB User). If you run this code: USE [master] ; GO CREATE LOGIN...(read more)

    Read the article

  • How to manually list set of urls for search engines to index

    - by MarutiB
    So I have created a video website which has thousands of videos and thousands of videos get added to it on a daily basis. Here is my problem :- I have created a website which basically loads the skeleton in html and puts all the content through javascript and Ajax. The problem is search engines aren't going anywhere except for the home page. Is there a way say in robots.txt where i give a link to a single html which has links to all these videos ? I agree my site is not accessible for a non-javascript user but stats show that this ratio is very low ( 0.2 % ). Is there a way I can still keep the complete AJAX website and still get each individual videos listed on google ?

    Read the article

  • mailing for categories

    - by nerkn
    There are more than 10 categories in my site, users can register more than one category. My php script prepare contents of each category, according to user preference my script merge those contents for each user. so every user can get what category they want. Problem is I want to send 340+ mails per hour and dreamhost dont allow. What do you suggest? I think to a service like mailchimp but I couldnt find that scenario, do they support category & content etc. Can I use smtp in dreamhost?

    Read the article

  • Entity Framework, Code First: where is the database?

    - by Marko Apfel
    With Entity Framework 5 in Visual Studio 2012 the code first feature could let you come to the question “Where is the automatically created database located?” I run in the question after changing the model which throws during the next run this error: “The model backing the 'MyContext' context has changed since the database was created. Consider using Code First Migrations to update the database (http://go.microsoft.com/fwlink/?LinkId=238269).” Okay – clear I thought “delete the database”. But where is the database and what type is it??? In this constellation the frameworks generates a localDB. You could access this database via SQL Server Object Explorer. For the first time you have to add this localDB. The server name is “(localdb)\v11.0”: And so we could browse through the content of this database. It got the same name like the context class.

    Read the article

  • Improving FAQ SEO with multiple pages?

    - by asdfasdf
    I have a client who has over 200 Question/Answer style content blocks. Neither the questions or answers are very long and most of them have almost the same question but with a word or two differentiating themselves from the rest of the questions. Would SEO be helped or hurt if I would to put each QA on its own page with the title of the page the question being asked etc... Or, would that be considered "farming"? If not, what would be the best way (in SEO world) do present all these QAs? Thanks for any advice..

    Read the article

  • Windows not recognizing its system drive after installing Ubuntu 11.10 alongside with windows 7. What to do?

    - by user53322
    After installing Ubuntu 11.10, it boots perfectly. But when I select Windows 7 from the GRUB menu, it restarts after showing the boot logo. I tried to repair the boot loader but the process failed. Then I decide to repair with system recovery disc. There I realize the system is unable to find any existing system. Then I boot into ubuntu and here I can see all the existing drive (with all content). All drives are still NTFS file system (I have 4 drive: 3 are NTFS another 1 is ext4). Tried to repair with gparted partition tool, but came with no luck. Also tried to reinstall windows but installer don't show any available drive. What to do? (something to do with Ubuntu?)

    Read the article

  • Sept. Chicago IT Architects Group Recap

    - by Tim Murphy
    Thank you to everyone who came out for last night’s presentation.  Hopefully we will have a little better turnout next month when we are back on our regular night.  I will post out the topic and the registration as soon as we get confirmation. For those interested in last night’s presentation you can find the slides here.  I am also planning on making a white paper post here with the full presentation content. See you next month. del.icio.us Tags: Chicago Information Technology Architects Group,Smartphones,Enterprise,Development

    Read the article

  • Pointers to implementing a taxonomy system?

    - by petsagouris
    I am taking part in the development of a PHP CMS (yeah I know you are like, meeh). It is currently using simple tags for tagging content. What I want to do is push the whole thing to using Taxonomy but I am not sure on what to pay attention to, or how to go about structuring the whole vocabularies/terms. I'd love to be able to keep the user interaction close to that of the current simple tags and at the same time allow for more advanced usage by power users. The Drupal and the Wordpress taxonomy systems are being reviewed at the moment and they provide valuable resources. Nevertheless, it could be that someone has some simpler and more appropriate thoughts on this matter.

    Read the article

  • How can I allow search engines to index my invite only website in ruby on rails?

    - by tstyle
    I have a ruby on rails website that will be in invite-only mode for the next couple of months. Currently I have it set up so visits to any page performs an authentication: before_filter :authenticate, :except => [:beta] //authenticate checks for a logged in user But the webpage has a lot of content that I would like to see indexed by search engines, and I was wondering if there's an easy way to allow crawlers to do their work? I am not very knowledgable on SEO related stuff at all, so sorry if this is an suboptimal way to phrase the question.

    Read the article

  • jQuery Mobile Frame Forwarding [on hold]

    - by Nizam
    I have a site that does a standard forward to another site [301 Redirect]. In the redirected site, I detect if the device is a mobile using the following code: if (/Android|webOS|iPhone|iPad|iPod|BlackBerry/i.test(navigator.userAgent)) { window.location.replace("Mobile/Login/Login.aspx") } else { window.location.replace("Apps/Login/Login.aspx") } It works and jQuery mobile makes the site fits device very well. To do so, I use the following code in ASPX page: <meta name="viewport" content="width=device-width, initial-scale=1, maximum-scale=1" /> The problem is that I am wanting to frame forward - instead of standard forward - (there are a lot of advantages), but the site is not fitted for the device anymore, and even the icon I have chosen for my page is not well defined anymore. The code I use to set the icon of page is: <link rel="apple-touch-icon" href="../../Apps/Imagens/Icone.png" /> My site is hosted by Mochahost My question is: Is there anything I could do to make it works?

    Read the article

  • jQuery scrolling images for e-commerce site, what to do about users who disable JS

    - by Livingston Storm
    As the title suggests, I am developing an e-commerce site and I intend of having two jQuery plug ins on the default page, one for scrolling images and the other for the navigation menu. Should I be concerned about making the site work if the users disables JS? Cause if they have it disabled my site would be almost impossible to use with the scrolling images blocking the main content. Plus the CMS I am using, Big Commerce, uses a bit of JS for the products pages, which would also look ridiculous with JS disabled. Anyone have experience with this?

    Read the article

  • How to Determine VPS Hosting Resources Needs for my upcoming Wordpress blog? How much resources should i purchase? [closed]

    - by Ishwar dixit
    Possible Duplicate: How to find web hosting that meets my requirements? Decided to purchase VPS hosting but Getting confused on amount of Resources i need? Wordpress will be used as platform, The blog i want to setup is assumed to have a traffic between 20k - 25k Visits per day with a rate of 5 pageviews per visit... there is No Download Facility provided...the content of the blog will be Text, Images & videos (will be used rarely)... The main question is? For the above requirement: How much RAM will be enough? How much CPU usage i will need? How much Bandwidth will be enough? How much Disk Space? Any other Requirement? Thanx in Advance..

    Read the article

  • New Office VBA site launch on MSDN

    MSDN now features a special site (also accessible via www.iheartmacros.com) dedicated to Microsoft Office VBA - an important part of the Office 2010 value.   The site presents a clear path for technically savvy users to begin learning how to customize the Office applications to meet their needs. This new MSDN site highlights content for: What you can do with VBA How to get started Key blogs VBA tips & tricks Access to a new interactive training tool, the Office Developer Atlas. In addition...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Another website is mirroring my site

    - by Marlboro Goodluck
    Question for you all. There is a site of ill repute known as thedirty which has completely mirrored my site and now has links appearing on Google at the #1 spot using my content. I checked my log file and noticed that this site has been crawling mine from sometime, and also has 10k links from their site to mine. I have blocked user access which is referred from this site and reported them as web spam to Google already. I also disavowed the domain. How are they getting top links in Google (even overtaking mine) for such nefarious tactics? What are the steps to completely eliminating an issue such as this?

    Read the article

  • Multiple Passwords on One Account

    - by user110789
    I'd like to join three ideas into one interesting and sometimes useful feature. There was a question about using multiple passwords earlier this year, but it didn't receive much attention. I'd like to ask the question again after showing an interesting and new way to use the feature. The three original posts I found to be interestingly combined were: (1) Multiple passwords per user (2) http://blog.littleimpact.de/index.php/2009/09/14/automatic-encryption-of-home-directories-using-truecrypt-62-and-pam_exec/ (3) http://www.truecrypt.org/docs/hidden-volume Basically I'd like to login to my account with two passwords and depending on the password I use, I would get a different content in my home directory. In a way I would get a cryptographically hidden account into my system. So the question is, is it possible to allow multiple passwords to log on to Ubuntu/Linux for the same user?

    Read the article

  • Impact of migrating home page from HTTP to HTTPS on search results

    - by 2Stroke SEO
    I've had to change one of my site's home page to HTTPS from HTTP. I had plenty of links coming into the HTTP page and was performing well in Google against many of my targeted search phrases. I did a 301 redirect from the HTTP page to transfer the link juice to the HTTPS page (and to prevent duplicate content issues) but my search rankings have tanked which indicates no link juice has been transferred. My PageRank has vanished - which I'd expect - but I'm really surprised that the SERP rankings fell off the face of the earth. Anyone have any ideas how I can recover from this? I've waited a couple of months since the changes took effect just in case Google was taking time to check it out.

    Read the article

  • Advantages of Search Engine-Friendly Websites

    For every webmaster, System Engine Optimization or SEO only means one thing - tailoring web content to attract more search engine-driven traffic to his website. The higher the rank of a website in the search engine results, the more traffic it gets. Improving the rank of a website means it will have more chances of being visited by more readers and people who may become potential clients or sources of revenue. Regardless of what services or products may be offered, it is essential for for the site to be search engine-friendly as statistics show over 95% of traffic are driven by search engines.

    Read the article

  • On which page(s) to add canonical?

    - by user6211
    I have two pages with same content and same meta title and meta description. they also have very simular url: http://www.mysite.com/new-york http://www.mysite.com/new_york I need first link to be "official". To avoid having duplicated pages, i want to add canonical meta tag in header... but on which page? does it have to be on both of them or only on second? On on first? Can you give me some advice please?

    Read the article

  • Will uploading our .docx files on scribd and embedding the files on our website affect search engine rankings?

    - by user1439968
    We have prepared notes for university students which are on .docx format. And we want it to put on our website for viewing. We tried one option. Uploading the files on scribd and embedding it on our website for viewing on scribd viewer. Will making documents available on srcibd viewer on our website affect search engine rankings ? Will search engines treat it as duplicate content as those are already uploaded on scribd and we are embedding it on our website ? On scribd we have set the uploaded documents as 'private' though. And if it affects, can you suggest any suitable way to make .docx files to be viewed on our website that doesn't affect search engine rankings ?

    Read the article

  • Hide from google while developing

    - by user210757
    I will be building a (wordpress) web site. While I am developing, other team members will be pushing content. I'd like to have it hidden from google while under development. It will be hosted on godaddy. I have thought of not pointing the domain name to it until live and using "preview dns", or buying a static IP during development. Or hosting dev site in a sub-directory ("/dev/") until ready and then moving it up a level. If in the dev directory I'd add htaccess or robots.txt to not crawl. Is any of this a bad idea? Will google penalize for any of this - like search by IP and then associate that with the domain later on? Any better ideas?

    Read the article

  • Web Animations - the next generation

    Web Animations - the next generation The GDL event page for this event: developers.google.com Ask and vote for questions at www.google.com Browser developers are designing the next generation of animation features that will enhance HTML5 with rich, powerful new animation primitives. Designers and web app developers will all benefit from improved methods of animating content. Hear from developers of Google Chrome and Mozilla Firefox how this works, and learn how to try out the new animation features now in your favorite browser. From: GoogleDevelopers Views: 0 3 ratings Time: 30:00 More in Science & Technology

    Read the article

< Previous Page | 558 559 560 561 562 563 564 565 566 567 568 569  | Next Page >