Search Results

Search found 19186 results on 768 pages for 'sharepoint search'.

Page 506/768 | < Previous Page | 502 503 504 505 506 507 508 509 510 511 512 513  | Next Page >

  • Restricting A Directory Through .htaccess

    - by Whitechapel
    I'm trying to put all of my FTP accounts into a folder on /public_html/ftp and password protect it so search bots can't crawl their private files. I'm also trying to redirect all site traffic from the non-www to www. I keep getting 500 errors when accessing the site, and I need to point it to www.vivalanation.com/ftp to www.vivalanation.com/ftp/, because the /ftp just errors out, you need the trailing slash. Here is my .htaccess in the /public_html/ftp folder: RewriteEngine on RewriteBase / RewriteCond %{HTTP_HOST} !^www\. [NC] RewriteRule ^(.*)$ http://www.%{HTTP_HOST}/$1 [R=301,L] AuthName "FTP Access" AuthType Basic AuthUserFile /home1/vivalst/.htpasswds/public_html/ftp/passwd Require valid-user I created a passwd file in /.htpasswds/public_html/ftp And here is my basic .htaccess in the root of /public_html/: RewriteEngine on RewriteBase / RewriteCond %{HTTP_HOST} !^www\. [NC] RewriteRule ^(.*)$ http://www.%{HTTP_HOST}/$1 [R=301,L]

    Read the article

  • Trying to retrace our SEO domain redirect strategy

    - by dans
    An SEO built a copy of my company's e-commerce site on another domain that contained our product's keywords in the name (i.e. as if Levi's built a duplicate site on bluejeans.com)...and then they referenced a lot of the images on the actual website from the other domain (as if Levis.com had images on it referenced like: img src="http://www.bluejeans.com/jeans-front.jpg"), but when you tried to reach the site by typing the name into the browser you would be redirected to the regular website, so the site wasn't really used for any purpose except I guess SEO. Since I didn't think this was doing anything GOOD for us at the time, I deleted the duplicate site and let the hosting on it expire, only to watch our search engine position rankings fall dramatically. Any ideas as to what was going on there? I want to get it back to understand its impact, but I don't know how it was set up. I contacted our host and they have no idea how it was set up. I suspect there was some sort of redirect in play, or something?

    Read the article

  • Are there advantages of using hard coded URLs for localization?

    - by nbolton
    On the Synergy website, localization is detected (and can be overridden) but uses the same URL for all languages. Some websites however, like Wikipedia have language specific subdomains. What are the advantages of having either subdomains or subdirectories (i.e. a specific URL) for each language localization? Also, should it automatically redirect the user to the specific subdomain/subdir based on the language that the browser requests? I suspect that there are advantages, which I'm guessing are: When the website appears in search results for non-English languages, the translated page description will be shown (assuming there is a translation provided by the website). When a user shares a page (e.g. through twitter), it will show in a specific language. Perhaps this is a disadvantage though? Am I correct, if so, are there more advantages?

    Read the article

  • cant boot with graphics in uefi

    - by user205996
    I really searched for information to this problem but none of the guides or threads on forums have helped, either i cant search or this havent come up. No nomodeset, forcevesa, turning brightness up or anything have worked.... I bought the laptop with ubuntu 13.04 preinstalled, bumblebeedrivers were on it but primusrun broke on the ubuntu 13.10 update. Decided to reinstall ubuntu and install windows at the same time, so now I got windows up and running from uefi and have installed ubuntu 13.10, grub works for windows and everything seems like normal. But when I try to boot ubuntu i cant manage to boot into anything else than failsafe rootterminal. Screen just go black or i have a blank terminal flashing i upper left corner.... Dont know anything about X configuration or what to do. If anyone has an idea please let me know. Laptop: Clevo w230st intel i7 haswell intel hd4000 (?) nvidia gtx765m

    Read the article

  • What plug in or module to use with WordPress? [migrated]

    - by Qacro
    I am developing travel website where users can search and book their travel deal. It goes like this: Providers are creating their travel deals (same as some blogger create blog in WordPress); Users book wanted travel deals; Providers, who have their account where they can see if users book their deal, are notified by the email and sms about just booked (sold) deal. Site is going to be developed using WordPress. Is there any plugin or module that I can use to accomplish this, or at least something similar to reconfigure and not to take this process from scratch?

    Read the article

  • How to dual boot ubuntu 13.10 with windows 8 (non UEFI)

    - by user204970
    I just want to ask.. How to dual boot ubuntu 13.10 x64 with windows 8 x64 (non UEFI installation)? And can I use default windows 8 grub? Because I like it so much. Just search for the answer (almost) anywhere but find no answer. I am using ASUS A46CB with Core i3 processor. Thank you for your help.. :) P.S. : I have 3 primary partition (that detected in windows 8 explorer) with MBR partition style

    Read the article

  • error when installing rmagic gem on Ubuntu 12.04 for Ruby on Rails

    - by PeaceDefender
    0 down vote favorite I am trying to get my Ruby on Rails application working and installing RMagic gem. But I get an error that says "Can't install RMagick 2.13.1. Can't find the ImageMagick library or one of the dependent libraries." Altough I have ImageMagick installed with its dependences. Terminal log : http://pastebin.com/vcgkbwZR I even tried to installed RMagic from Synaptic, but I get the same error when I run "bundle" command. Through my search, some people had problem with lib dependencies, but I believed I have them all installed. Another solution I found is this ( error installing RMagick from gem ) a little bit old and not sure if it is going to work for 12.04. BUT, I don't know where those configurations should go.

    Read the article

  • Google not showing any pages from my site in the index after three months [on hold]

    - by Alex Coisman
    Despite having a sitemap and using Google Webmaster Tools, it has been over 3 months and my site has not been added to the Google index at all. Here's the site: www.famouslefthandedpeople.com As far as I know, I have done everything correctly. However, there must be something I am overlooking that is preventing Google from indexing the site. I do not have a robots.txt file, so allow/disallow isn't the issue. Although the content of the site is sparse, it is original and not duplicated internally or externally so Panda/Penguin should not be a problem. I have reviewed the answers at Why isn't my website in Google search results? and I don't think it applies here. If it matters, I am using WordPress to create the site. What other factors should I be looking at in order to troubleshoot this?

    Read the article

  • SQL Server Data Type Precedence

    I am executing a simple query/stored procedure from my application against a large table and it's taking a long time to execute. The column I'm using in my WHERE clause is indexed and it's very selective. The search column is not wrapped in a function so that's not the issue. What could be going wrong? Schedule Azure backupsRed Gate’s Cloud Services makes it simple to create and schedule backups of your SQL Azure databases to Azure blob storage or Amazon S3. Try it for free today.

    Read the article

  • Why are new pages not being indexed and old pages stay in the index?

    - by ZakGottlieb
    I currently have a site that was recently restructured, causing much of its content to be reposted, creating new URL's for each page. To avoid duplicates, all of the existing pages were added to the robots file. That said, it has now been over a week - I know Google has recrawled the site - and when I search for term X, it is stil the old page that is ranking, with the new one nowhere to be seen. I'm assuming it's a cached version, but why are so many of the old pages still appearing in the index? Furthermore, all "tags" pages (it's a Q&A site, like this one) were also added to the robots a few months ago, yet I think they are all still appearing in the index. Anyone got any ideas about why this is happening, and how I can get my new pages indexed?

    Read the article

  • MonoDevelop 2.4 released

    Today we are releasing MonoDevelop 2.4, which includes plenty of new features and improvements. The release highlights are: Workbench usability improvements, which include a new Navigate to Symbol dialog, Solution and Class pad zooming, better pad layout, and many other look & feel improvements. Improvements in project management, with support for standalone assembly projects, and external console support for Windows and Mac. Extensive text editor improvements, including in-line search,...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • IE9

    - by xamlnotes
    Hot dog. IE 9 just hit the download sites this week. I have been running it for a few days and its really sweet. It seems much faster than IE 8 and many other browsers and its got lots of cool features. Some of the ones I really like are: New tab format with one click creation and putting them up top. Cleaner UI Ability to drag a tab off the tab bar and have a new window created Integrated address / search box Support for HTML 5   check out http://www.beautyoftheweb.com/ to see some of the cool features. Pay attention to the HTML 5 samples too. And theres lots more as its just getting off the ground. You can download it from Microsoft. Or you can get the version with hooks into bing and msn. Enjoy.

    Read the article

  • Two different domain for specific languages pointing to one site

    - by user25599
    I developing a client's blog and he needs it to be bilingual (english and spanish). Now what he wants is that users can get to the content based on the domain e.g. Jhon enters www.domain.com and he gets the english version and Juan enters www.elsenordominio.com to get the spanish version. All content will be validated by php so the users and search engines only reads the domain related language. What do i need to use header re-direct or 301? is it bad for seo? will google will google penalize me? I hope you guys can help me and forgive me if my english is not good.

    Read the article

  • Local Small Businesses Are in Need of Your SEO Skills

    Local Small Businesses are in need of your skills to get a web presence that is ranked highly in search engines. What if I told you that you could make significant, even full-time income in one of the biggest online markets, and I told you that this big market is least competitive too. You don't even need to have your own product or your own mailing list for this to work. Here is one of the most important parts, you don't need to sell these products to make profit either, and no this is not affiliate marketing.

    Read the article

  • Release Notes for 8/16/2012

    Below are the release notes for today's deployment. Source Tab Improvements Revamped fork navigation to make it more clear that you are within a fork verses the parent project. Infrastructure Updated CodePlex to ASP.Net MVC 4.0. RTM+1 how's that for a turnaround? Bug Fixes Fixed a set of user experience issues with the in-line diff view experience. Fixed issue with date mismatches between UTC and local time for discussions and issue tracker Fixed issue where new issues would not show up in search results. Changed the default destination branch for pull requests to master. Have ideas on how to improve CodePlex? Please visit our suggestions page! Vote for existing ideas or submit a new one. As always you can reach out to the CodePlex team on Twitter @codeplex or reach me directly @mgroves84

    Read the article

  • I can't log-in to WinSCP using username "root"

    - by Jessyle Ivy
    I can't log-in to WinSCP using username "root". I already change the password of "root" in Ubuntu, and I successfully log-in there. But in WinSCP, it goes like this Search for host... Connecting to host... Authenticating... Using username "root" Authenticating with pre-entered password. Access Denied. and I am need to re-type the password again. By the way I'm using VMware Player for Ubuntu. Thanks!

    Read the article

  • Alt text vs CSS sprites (SEO vs speed)

    - by leeoniya
    I'm reworking our site to reduce HTTP requests and blocking requests by concatenating JS, css, gzipping, loading all JS via LABjs and using CSS sprites for images that were loaded individually via <img> tags before. Progress has been great so far - 5x page load performance improvement. However, we're in the top 5 organic search ranking in google for many targeted keywords and phrases. I'm afraid eliminating so many img tags with alt attributes can hurt our SEO. Does anyone have any experience with alt tag manip/removal and effects on SEO positions? Is previous rank "sticky"?

    Read the article

  • Is there any way to determine who placed an AdWords ad?

    - by geocoin
    I was in the process of discussing with a client using adwords for their next step of promotion and no-one at the company was aware of the system, and certainly hadn't created any adwords campaigns, however in the midst of the discussion and trying a few relevant search queries out, one of the ads in the ad block on the right hand pane was for their website. Is there any way of determining if the ad was somehow generated of is it definitely the case that someone had to consciously take the ad out and if so who might have placed it? They have no referral system so someone outside of the company would get no benefit from running the ad

    Read the article

  • sed problem ....

    - by moata_u
    hello there ... am facing problem in sed command , i was trying write a bash script that do the following : 1. search for the line that contain :@ , 2.then save the line that contained :@ and replace it with new line ....as following : ! /bin/bash echo "Please enter the ip address of you file" read ipnumber find=grep ':@' application.properties # find the line input="connection.url=jdbc\racle\:thin\:@$ipnumber\:1521\:billz" # preparing new line echo sed "s/'${find}'/'${input}'/g" application.properties # replace old with new line **Problem is nothing happen !!!! * I already tried to use "${find}" instead of '${find}'

    Read the article

  • Jeu-concours webmarketing : 2 places à gagner d'une valeur totale de 2 290 euros pour l'événement eM

    Exclusivement pour les membres de Developpez.com : 2 places à gagner pour les deux événements [IMG]http://x-plode.developpez.com/images/emetrics-smx.png[/IMG] C'est à eMetrics Marketing Optimization Summit et à la Search Marketing Expo, à Paris, que se rencontreront les grands acteurs de l´optimisation marketing, les leaders d´opinion, les analystes de marché et les utilisateurs. Venez découvrir l'ensemble des technologies et des techniques existantes, ainsi que les informations et la Business Intelligence indispensables pour optimiser votre marketing en ligne.Deux jours complets de conférences ? des sessions traitant de sujets d´actualités et des stratégies de succès dans les domaines du SEO et PPC ? des experts nationaux et int...

    Read the article

  • What is Rails way to save images?

    - by user
    I develop on iOS, and I'm switching from a PHP backend to Ruby on Rails. The interchange format is JSON. A quick Google search for 'save images in Rails' has nearly every result talking about saving image data as blobs to the database. I might be mistaken, but I'm under the impression that saving image data in a database is a huge waste of time and space (as opposed to saving a link to the file location ('/img/subcat/4656.png'). In PHP, it's pretty standard to receive the data, generate a filename, then save that file to disk, and then update the database with the image's location on disk. Is this the same for Rails, or is there some built-in ActiveRecord image functionality I'm not aware of?

    Read the article

  • How to know an article is copy from other site? [on hold]

    - by cj333
    We have a site. our main servers is for blogs and freedom write. Our custom can submit their article in our site. But for ptotect the copywrite, we do not allow custom just do easy copy and parse a whole article from other site. At least they should write something by themselves. And how to check if the article is a completed copy from other site? google webmasters? google search api? or other better way? Because we always received DMCA notice by google, so I am tring to stopping the articles before they were posted. Thanks.

    Read the article

  • 12.04: I need a new network card - and I think this one may work?

    - by Jason Malone
    I found this list of compatible USB wireless network adaptors: http://www.cyberciti.biz/tips/linux-usb-wireless-compatibility-adapter-list.html Listed here is the Belkin F5D8053. I'm looking for a store to buy a suitable adaptor in locally and I can't find that particular adaptor - but I can see a Belkin F7D4101az. I think this is simply a newer model - but a Google search for "ubuntu belkin f7d4101az" brings no results - not even people trying to get it to work. Nothing. So I didn't want to purchase it without being sure. There's really only one PC shop that sells this sort of stuff around here - so here's a list of all my available options: http://www.pcworld.co.uk/gbuk/wireless-cards-adapters/xx_7088_70095_xx_xx/xx-criteria.html I'd really like to get one today, because I have a lot of work to do and don't really want to have to wait a week for a new adaptor - I spent 6 or 7 hours last night trying to get my Dell 1450 to work with no luck. Any help would be much appreciated.

    Read the article

  • Cloud proxying service

    - by ChristopherJ
    I have an app that mashes up images from Bing image search, it's hosted on Heroku written in rails. The app is client side in javascript, so the mashup is done on an html5 canvas - this means though that if I fetch the images direct from the Bing server, the canvas gets dirty and I can't save it. As a quick work around, i have set up a route on my rails app that simply proxies the request to Bing and passes the result back through. Obviously this is a very poor performance solution and will eat up my dynos very quickly. Can anyone suggest a more suitable option? At the moment I'm thinking maybe Amazon EC2 with apache mod_rewrite rules would be better performing and more cost effective. Is there a cloud service (or an app I could deploy to a cloud service) that would be more appropriate for proxying requests for me so that my javascript can fetch the images without dirtying the canvas?

    Read the article

  • How to install texlive2011?

    - by vrx
    I just moved from Windows to Ubuntu. Software Center in Ubuntu does have texlive2009 package. There is no straightforward installation process from TUG, or from Google search result. I found some installation of Texlive2010 but it does not work. Here are the step: download an texlive2011.iso, 2.3GB and save in local drive mount to a virtual drive install perl tk (ok) "sudo ./install-tl" does not work, "perl ./install-tl" shows pre-installation setting but it does not have the write access to /usr/local/textlive/2011/ Finally, I'm stuck Please guide me with a step by step tutorial or suggest any other solution beside typing in the black screen without knowing what is going on behind the Terminal.

    Read the article

< Previous Page | 502 503 504 505 506 507 508 509 510 511 512 513  | Next Page >