Search Results

Search found 9727 results on 390 pages for 'llblgen pro'.

Page 239/390 | < Previous Page | 235 236 237 238 239 240 241 242 243 244 245 246  | Next Page >

  • Forwarding naked domain to www [duplicate]

    - by mravec
    This question already has an answer here: Why is www. working but no-www is not? 1 answer I successfully mapped my www.mydomain.com to my Google site (as described here). However I would like to forward also my naked domain (mydomain.com) to end up in www.mydomain.com. I assume I should modify 'A' DNS record in my provider but what IP should I point it to?

    Read the article

  • MySQL can only log in as root, even after creating new users with their own database

    - by ionFish
    Problem: I just set up a Debian Wheezy installation for testing, and installed the LAMP packages and PMA. I can log in as root with my pre-defined password, create/edit/delete both databases and users. The problem comes when I create a new user 'something', set a password for it, and grant it all privileges on a table 'something' (same as the username). Upon connecting, it denies access to the user. Details: Host: localhost using MySQL 5.5.24-8 Creating user: CREATE USER 'something'@'%' IDENTIFIED BY '***';GRANT USAGE ON *.* TO 'something'@'%' IDENTIFIED BY '***' WITH MAX_QUERIES_PER_HOUR 0 MAX_CONNECTIONS_PER_HOUR 0 MAX_UPDATES_PER_HOUR 0 MAX_USER_CONNECTIONS 0;CREATE DATABASE IF NOT EXISTSsomething;GRANT ALL PRIVILEGES ONsomething.* TO 'something'@'%'; Checking privileges: GRANT USAGE ON *.* TO 'something'@'%' IDENTIFIED BY PASSWORD '*92F9DAF5F5129554509489FDB6A433510223C799'; Result: Access denied for user 'something'@'localhost' (using password: YES) More Info: I use this same exact procedure for the Squeeze distribution, and it works perfectly. Is there a chance it's because of Wheezy, or something else? I need to continue using Wheezy because of the updated packages (for this test server -- the others work fine), so 'just use Squeeze' is not an option. Note: I HAVE tried flush privileges; to no avail.

    Read the article

  • Is it possible to use two different ErrorDocuments for different paths of a website?

    - by tapwater
    this is my first question on stackexchange and it might be a bit confusing. I currently run PmWiki (sorry, you'll have to google it, new user can only have 2 hyperlinks) at mydomain.com/pmwiki. I have a 404 page and .htaccess set up in my site's document root for 404 pages regarding anything that doesn't have to do with my wiki. By default, PmWiki handles URLs a little confusingly so I had to use this in order to get it to look like mydomain.com/pmwiki/Namespace/Page I had to create a .htaccess in /pmwiki to remove parts of the URL. PmWiki also has a custom 404 (Site/PageNotFound) page that has stopped working, now my site uses the /404.htm page. I noticed this when trying to install this "recipe" to enable case-insensitive URLs. Currently the only way to access Site/PageNotFound is by actually linking to it, and, if you read how that recipe seems to function, this is an issue. Currently mydomain.com/pmwiki/blahblah and mydomain.com/pmwiki/legitimate_namespace_but_lowercase/legitimate_lowercase_page_name both direct to mydomain.com/404.htm. I have to admit I'm very confused, and I apologize if I was unclear in any of this, but I could definitely use some help. Thanks!

    Read the article

  • How can I get stats for what 3rd-party sites have embedded our iframe widget?

    - by Su'
    Say we've produced a widget for other sites to use, like so: <iframe src="http://example.com/whatever.php" frameBorder="0" width="200px" height="300px" scrolling="no"></iframe> The client would like to be able to see within GA who has embedded the thing. Is there some referer information automatically passed that I can look for, or do I need to add something? whatever.php is already loading the analytics Javascript(we're also tracking clicks on an outbound link). [EDIT] Looking around a bit more, I found what seems to be a similar question on SO with an answer saying this can be found, automatically, but I still can't seem to find the information. The question's also old enough the respondent is probably referring to the old interface, though. Maybe someone could explain getting to it in the new look. (I won't likely be able to train this client to switch, deal with the old look, etc.)

    Read the article

  • How to make Google show my site in search result like the following image? [closed]

    - by Samik Chattopadhyay
    Possible Duplicate: What are the most important things I need to do to encourage Google Sitelinks? Currently Google is displaying my site (http://layzend.info) like this in search result, only the link and meta description without any internal page links - But I want to be the search result like the following where the internal links are also displayed - How is it possible? Please help me to make my site more SEO friendly.

    Read the article

  • Difference between Global and Local SEO

    - by user29660
    I have been reading up on SEO techniques in an effort to learn how to do it thoroughly so I can charge my client for the service. To guage my price I have checked out competitor prices and noticed that theres a fair price difference when it comes to guarenteeing a page 1 ranking with global keywords compared to local keywords. So what is the difference in terms of work load and techniques used to justify this price difference? just to clarify, i am looking for technical differences in programming , methodology etc.

    Read the article

  • Somehow Google considers a properly 301'd URL as 200 and is still indexing the new content in old page?

    - by user2178914
    We redirected all the old URL's to new ones properly using htaccess. The problem is Google, somehow is still finding content in the old page(which it shouldn't) and stores it in the cache rather than the new URL. For eg: Old Page- http://www.natures-energies.com/iching.htm New Page- http://www.natures-energies.com/index.php?option=com_content&view=article&id=760 If you type the old URL into the browser it redirects If you fetch the old URL as Googlebot in the webmaster tools the header says 301/permanently redirected. If I try to crawl as any other bot it still says 301 redirected. Even if you click the old link in Google it redirects to the new URL. Only in its cache it shows the old URL and moreover it shows the new content in it! I am stumped on how Google manages to grab the new content and puts in the old URL instead of the new one! One more interesting thing is that if I try a cache for the new page it shows the cache of the new content with old URL! Any help would be appreciated. I am at end of my wits. I think i have tried almost everything. Is there anything that I'm missing to see? You can use this search to find the old url's. Maybe you'll some patterns that i missed. site:www.natures-energies.com inurl:htm -inurl:https|index

    Read the article

  • Restricting URL w.r.t HTTP method and setting different authentication mechanism for each

    - by user31745
    I shall start with an example. I want to restrict to POST requests only for http://path/to/logical/abc.xml and restrict to GET only for http://path/to/logical/def.xml. How do I put constraints like this as the paths are logical and location directive is not supported in .htaccess? The actual problem is to set different authentication type(basic, digest) on diff logical file.for eg. for abc.xml I want to authenticate for Basic type of authentication and def.xml with digest.

    Read the article

  • Is it legal to charge extra fees for copyrighted content on mobile platforms?

    - by Macrow Willson
    this question just came up as we recently bought content from image stock portals. Many of those altered their license agreement in favor of charging more for using in mobile apps. So instead of using their standard licenses, you need to pay an "extended" licenses which multiplies the fee easily by 5-10. That doesn't make sense as the mobile device is just a smaller browser and protects the content even better than a desktop computer. Are those stock agencies allowed to do that, and is it legal at all ? I am not a lawyer but I would even risk to go on with the standard license and wait to be sued in that matter.

    Read the article

  • Making a calendar tool with Google API. Should I use OAuth Service Account?

    - by Goluxas
    I'm creating a calendar tool for a client. He has his work calendars in a Google account, which I have access to. I'll be using the Google Calendar API for the tool, but I'm new to Google APIs and the OAuth stuff. I'll be using PHP. Do I need to use an OAuth2 Service Account, or should I make a regular OAuth2 Web Application Client ID from that Google account? In case this information is helpful, the tool I will be making will do this: 1) Allow my client to fill out a form which will create a new calendar and populate it with several events. 2) Send out an email to a mailing list twice a week listing changes that have been made to these calendars. I'll also be making a page that displays the iframes for each calendar, so his clients may see them even if they do not have a Google account.

    Read the article

  • Will a search engine lower the rank of my page if i have a hidden iframes?

    - by Skurpi
    As a praxis, all external content on our site is put in iframes to lower the risks of any external parties injecting stuff to our users. We also do it to make sure our content shows up before banners, to make the site feel quicker. We now have an external script running which we want to put in an iframe, but it does not have any visible content to go with it so I want to put css "visibility: hidden;" on the iframe. I read in a forum somewhere that search engines will lower the rank of a page, or even drop the page, if a iframe has "the minimal size of 1x1px". Will a search engine lower the rank of my page if I have a hidden (or 1px big) iframe?

    Read the article

  • Can I dynamicaly size a div based on discreet units and still center? [closed]

    - by Dave
    Here's the problem: I have a website I'm working on that, depending on what the user has selected, will pop up a different number of boxes 80px high and 200px wide and currently set to float:left. These boxes are contained within a div that is basically the whole width of the screen minus some 1% margins. So at the moment they all fill in the box and, depending on screen size, occupy a grid of variable height and width. The problem is, if the screen size makes the containing box, say, 700px wide then you end up with 3 boxes per row and a bloody big margin on the right. What I would like to do is center the grid of boxes inside the containing box so that the margins are equal left and right. I suspect this can't be done since it means the containing box needs to set its size by looking both at the size of the user's window as well as the size of its children. It would be easy to do with javascript but I'd prefer not to if that is an option. If it is truly impossible then I will simply script it and let non-js users see a left-justified set of boxes.

    Read the article

  • How to compete in a Saturated Niche? [closed]

    - by jasondavis
    Possible Duplicate: Does the position of the content in the source make a difference to SEO? When entering a hugely over saturated niche such as web-design, is it even possible to compete with the big sites that have been ranked on google's #1 page for years? Also, I have read about how important titles, link anchors, and headings are for SEO and how content is the most important. So let's say we are building a standard header, body, sidebar, footer page. In the the actual markup, would it be better to make sure the main content comes before the sidebar on the page or does this probably not make a difference?

    Read the article

  • Is there any way to determine who placed an AdWords ad?

    - by geocoin
    I was in the process of discussing with a client using adwords for their next step of promotion and no-one at the company was aware of the system, and certainly hadn't created any adwords campaigns, however in the midst of the discussion and trying a few relevant search queries out, one of the ads in the ad block on the right hand pane was for their website. Is there any way of determining if the ad was somehow generated of is it definitely the case that someone had to consciously take the ad out and if so who might have placed it? They have no referral system so someone outside of the company would get no benefit from running the ad

    Read the article

  • How long before Google will update search terms matching my website?

    - by Camran
    I have a website which title I changed about a month ago. The website is a classifieds website which is dynamic, using php. The title changed from "Free classifieds" to "buy and sell free classifieds". The strange part is that after about two weeks the title showed in google search results changed to the new title, BUT when I searched for "buy and sell free classifieds" my website didn't show up at all. I mean I have gone through over 30 pages of search results and my site isn't listed. However, searching for "free classifieds" still display my website at the same position it was before the title change. Any reason for this? How patient should I be? FYI the website has a sitemap submitted and updated, good meta tags and is W3 valid etc etc, so that is not the problem here. Thanks

    Read the article

  • How to move domain from 1&1 to fasthosts with minimum downtime [duplicate]

    - by arcanine
    This question already has an answer here: Changing web hosting companies 1 answer How can we move our domain name from 1&1's control over to our new host with as little affect (if any) on downtime for our domain name. We've tried moving name servers before and our site went dark despite us defining all the old records on to the new name server host, now this time we're going to move the entire domain to another company but we've been bit so many times by this that I'm trying to find out what the process should be, and whether it can be done in a way to prevent down time

    Read the article

  • Page views in Google Analitics are off compared to a similar metric

    - by tiki16
    We have a page where a user can sign a pledge to recycle by clicking a pledge button. A script writes it to a text file which updates the number on the page. In the past 2 days there have been 185 pledges signed but only 63 page views in GA. I trust that they are unique pledges and not just people adding multiple fake names and entering it. Is there anyway to get a better report from Google Analytics?

    Read the article

  • SEO Blog Indexing : Dot Wordpress Versus a Registered Domain?

    - by rumspringa00
    I've used Wordpress for a few of my client's sites, mostly small businesses and ecommerce sites. I have found through Google Analytics as well as the All in One Webmaster plugin that when it comes to social media, using Wordpress is a surefire way of getting your site indexed by Google and occasionally Bing and Yahoo. Since I am a heavy WP user, I'd like to contribute by registering a dot Wordpress domain for my portfolio. When using a WP installation concurrently with a WP domain, e.g. myportfolio.wordpress.com, will the site be more or less likely to be indexed rather a generic myportfolio.com domain? I've seen mixed opinions where people seem to favor a WP domain for URL output where others say that it's a moot point, and that Google will not favor a WP domain over a dot com domain as long as your meta tags are updated and content is keyword optimized. I tend to disagree and believe a WP domian would more likely be indexed and output more URLs over an individual, laconic domain like myportfolio.com. Am I wrong? Thanks in advance!

    Read the article

  • configure postfix to send as different domains

    - by Cerales
    I have postfix configured to receive mail for multiple domains. The primary domain goes through a standard postfix setup, whereas a secondary domain is handed by a /etc/postfix/virtual.db file which maps addresses for the secondary domain to other unix users. Incoming mail works perfectly, but when I send mail as users on the secondary domain, the mail receiver still sees it as coming from [unix username]@[primarydomain].net. Can anyone help me figure out how to configure the virtual postfix domain to send mail through the same domain it receives it as?

    Read the article

  • Appcache and jquery mobile on a CMS powered site?

    - by user793011
    Has anyone used the cache manifest to make a CMS site work offline? I've made a demo with static html files which seems to work fine, so I'm assuming it wouldn't be too hard to achieve the same thing with a CMS. The way that you tell browsers that files have changed (and so need to be downloaded again) is by adding a comment to the cache manifest file so its byte size changes. I'm not quite sure how to do this with a CMS, but maybe some sort of server cron could run periodically? Personally I'm more interested in having a site that works offline rather than achieving ideal performance, so if the file was modified every hour rather than when content actually changed that would be fine for me. If anyone has used appcache with a CMS, has anyone done so with jquery mobile at the same time? What I'm after is a fully native feel to a site that's accessible offline, in other words I want to mimic a native App. My static demo does this perfectly with jquery mobile, so again I would have thought this would be achievable in a CMS.

    Read the article

  • How can I allow search engines to index my invite only website in ruby on rails?

    - by tstyle
    I have a ruby on rails website that will be in invite-only mode for the next couple of months. Currently I have it set up so visits to any page performs an authentication: before_filter :authenticate, :except => [:beta] //authenticate checks for a logged in user But the webpage has a lot of content that I would like to see indexed by search engines, and I was wondering if there's an easy way to allow crawlers to do their work? I am not very knowledgable on SEO related stuff at all, so sorry if this is an suboptimal way to phrase the question.

    Read the article

  • Blank lines between sourcecode [closed]

    - by manix
    I'm so confused with a strange behaviour. Actually I have edited some php files remotely with my PhpDesigner8 (a php editor). Everything goes right, but when my teammates reopen the files that I have edited the source code have blank lines like below: class AdminController extends Controller { function __construct() { parent::__construct(); if (!$this->session->can_admin()) { show_error('Solo para administradores.'); } $this->load->library('backend'); } } Instead of class AdminController extends Controller { function __construct() { parent::__construct(); if (!$this->session->can_admin()) { show_error('Solo para administradores.'); } $this->load->library('backend'); } } Did you have experience these kinds of problems?

    Read the article

  • cookie not being sent when requesting JS

    - by Mala
    I host a webservice, and provide my members with a Javascript bookmarklet, which loads a JS sript from my server. However, clients must be logged in, in order to receive the JS script. This works for almost everybody. However, some users on setups (i.e. browser/OS) that are known to work for other people have the following problem: when they request the script via the javascript bookmarklet from my server, their cookie from my server does not get included with the request, and as such they are always "not authenticated". I'm making the request in the following way: var myScript = eltCreate('script'); myScript.setAttribute('src','http://myserver.com/script'); document.body.appendChild(myScript); In a fit of confused desperation, I changed the script page to simply output "My cookie has [x] elements" where [x] is count($_COOKIE). If this extremely small subset of users requests the script via the normal method, the message reads "My cookie has 0 elements". When they access the URL directly in their browser, the message reads "My cookie has 7 elements". What on earth could be going on?!

    Read the article

  • CLI Shared Hosting Management (scripts to manage web users and hosts) [on hold]

    - by aularon
    I am currently administrating two servers: first one has no control panel, I am creating directory structure, setting permission, configuring different aspects (users/php-fpm pools, nginx hosts..) for each of sites. With more clients, I sat up ISPConfig on my second server, everything is easily handled by ISPConfig Web Interface. However, I am searching for a CLI based solution, i.e. a set of scripts to create and manage hosts. Basically, a method to control ISPConfig from the command line (so I can use it over SSH) would be a good start. Does anybody know of such effort? I searched but all I got was web based solutions. Thanks.

    Read the article

  • What are the minimum steps that I should follow to ensure that my web site is accessible to the disabled?

    - by Tim Post
    I am trying to follow a very important standard that I must admit I have ignored up until recently. I want to make sure that my pages are accessible to a large portion of people that have disabilities. I focus mainly on tutorials that are text and image intensive, but no video / flash or any kind of animations. What is a checklist that I can follow to ensure that many people with disabilities can have a good experience when using my web site, and what disabilities should I be most conscious of? I know that I can't possibly please everyone. I have gone through the W3C guidelines, however I'm not entirely sure what standards apply to me. I'm not building web applications, I'm building mostly wiki like information exchanges, blogs and the occasional forum.

    Read the article

< Previous Page | 235 236 237 238 239 240 241 242 243 244 245 246  | Next Page >