Search Results

Search found 11896 results on 476 pages for 'smart pro'.

Page 259/476 | < Previous Page | 255 256 257 258 259 260 261 262 263 264 265 266  | Next Page >

  • Getting expired domain name - most effective route?

    - by kcdwayne
    There is a domain name I have been wanting for years that was used as a parked page (read: cybersquatted) that has entered into WHOIS's redemptionPeriod stage. The domain has been expired for 61 days now - after contacting the Registar, they informed me that it would stay in redemption for 75 days. After which, it would either be sold to resellers, or sent back to the public registry. (I have since sent a followup message requesting the reseller they use.) My question is: What is the best way to proceed? I know there is at least one competitor that would love to have this name, but I'm unsure if they even know it's expiring. I did not tell the Registar the domain in question, as they seem geared towards cybersquatting, and I do not trust them. Domain Front Running sucks. Should I use a backorder service? Should I just take my chances and try to grab it after75 days? I checked an auctionhouse by manually browsing their expired domains - it wasn't there.

    Read the article

  • How to offer a cookie opt in/out to users?

    - by Darkcat Studios
    I intend to use google analyticts, and as I understand it I will need to offer users the option to opt out of cookies. The question is this: I HATE these constant cookie option boxes, everyone I ask it getting annoyed by them too. Its nice to have the option, but we all know they have been in use for well over a decade. So - how big of a deal do I have to make about the fact that I'm using GoogleAn? can I pop a small link at the bottom of the page, maybe integrate it into "Privacy policies" page, and give people the option to opt out there? This would be very much the "Assume the majority of users don't mind, but at least make the option available" stance. Ironically setting a cookie seems to be the only way I can see to enforce the opt-out! as IP's change.

    Read the article

  • Will a search engine lower the rank of my page if i have a hidden iframes?

    - by Skurpi
    As a praxis, all external content on our site is put in iframes to lower the risks of any external parties injecting stuff to our users. We also do it to make sure our content shows up before banners, to make the site feel quicker. We now have an external script running which we want to put in an iframe, but it does not have any visible content to go with it so I want to put css "visibility: hidden;" on the iframe. I read in a forum somewhere that search engines will lower the rank of a page, or even drop the page, if a iframe has "the minimal size of 1x1px". Will a search engine lower the rank of my page if I have a hidden (or 1px big) iframe?

    Read the article

  • Why won't Webmaster Tools let me set a preferred domain? (says to verify, but it should already be)

    - by Su'
    I've got a domain that I apparently forgot to set a preferred domain for, so I just tried to do it. Webmaster Tools instead popped up a little box: Part of the process of setting a preferred domain is to verify that you own http://www.example.com/. Please verify http://www.example.com/. I'm running into some problems following these instructions: As far as I can tell I already did verify sometime in the past. There's a TXT DNS record with the gibberish Google tells you to set for it that I couldn't have come up with myself. …and nothing is telling me this information is bad. So let's assume the site is somehow not actually verified. All the various methods' instructions start with "click the Manage Site button next to the site you want, and then click Verify this site." That button doesn't even exist on my screens. (It presumably goes away when you successfully verify?) Those instructions were all updated pretty recently, and the DNS one in particular just a couple weeks ago so it seems a bit unlikely they're inaccurate. I am not using Apps, and won't be, so can't try out the verification through there. Note that I also have another domain that I have not done any verification for which is showing the same behavior(no such button, being told to verify when it seems impossible etc.) so something appears to just be broken. I already have a no-www process in place server-side, so we can skip that. I'm just trying to get the box checked off in GWT. If I don't get any resolution, I'll eventually scrap the TXT record and see if the site gets un-verified(or whatever since it thinks it isn't), and see if I can just restart the process. It's not urgent so I'm just trying to figure out if I've gone blind to something or what. Did the button move?

    Read the article

  • SEO Blog Indexing : Dot Wordpress Versus a Registered Domain?

    - by rumspringa00
    I've used Wordpress for a few of my client's sites, mostly small businesses and ecommerce sites. I have found through Google Analytics as well as the All in One Webmaster plugin that when it comes to social media, using Wordpress is a surefire way of getting your site indexed by Google and occasionally Bing and Yahoo. Since I am a heavy WP user, I'd like to contribute by registering a dot Wordpress domain for my portfolio. When using a WP installation concurrently with a WP domain, e.g. myportfolio.wordpress.com, will the site be more or less likely to be indexed rather a generic myportfolio.com domain? I've seen mixed opinions where people seem to favor a WP domain for URL output where others say that it's a moot point, and that Google will not favor a WP domain over a dot com domain as long as your meta tags are updated and content is keyword optimized. I tend to disagree and believe a WP domian would more likely be indexed and output more URLs over an individual, laconic domain like myportfolio.com. Am I wrong? Thanks in advance!

    Read the article

  • How to transfer a website hosted online to my Virtual Host

    - by user831740
    I'm doing some work on a webpage, meaning I gotta modify a few things. As a result of this I installed apache and all the things that come associated with it, and also installed Joomla, and I got everything running well. My problem is, I need to make apache run this website locally, so I downloaded the public_html folder from the FTP server on my website, but I have absolutely no idea how to implement the website in order to make it run on apache. I've read a few guides, but they all tell me how to create a new website instead of helping me host a website already done.

    Read the article

  • Does having over 80% dynamic and rapidly changing content affect SEO?

    - by webmasters
    I have a website that pulls promotions of products from other website. My index page has a structure similar to this: My Brand - Best Promotions Looking for great deals? Check out our top promotions A menu - listing the promotions categories 20 of the latest promotions (the best ones): I list an image; Promotion description (200 chars); Link to the promotion page. Question: More then 80% of my index page (maybe even 90%) is composed of the 20 promotions I list; these promotions change on a daily bases - which dramatically changes the content of my index page. Does the dynamic changing of the index page affect SEO? Should I try to add more static text where I can? (which won't change) Ty

    Read the article

  • XML Parsing Error at 1:1544. Error 4: not well-formed (invalid token)

    - by Steve
    I have installed Joomla 1.5.22 on a new hosting account, which doesn't have a domain yet, so it's public URL is http://cp-013.micron21.com/~annimac/ A message saying: XML Parsing Error at 1:1544. Error 4: not well-formed (invalid token). The source code for this message is: <dl id="system-message"> <dt class="error">Error</dt> <dd class="error message fade"> <ul> <li>XML Parsing Error at 1:1544. Error 4: not well-formed (invalid token)</li> </ul> </dd> </dl> There is nothing in /logs to indicate what the problem is. I have uploaded the following folders from a freshly unzipped copy of Joomla 1.5.22: administrator components includes language\en-GB libraries modules plugins templates\ja_purity xmlrpc and the issue remains. I have no custom or additional plugins, modules, or components installed. If I change templates, the problem remains. What is the problem?

    Read the article

  • Any risks in forwarding incoming requests to my computer? [closed]

    - by pimvdb
    I just installed EasyPHP, which is basically a combination of Apache, PHP and MySQL for dummies. I'd like to make my router forward incoming requests to my computer, so that everyone on the Internet could fetch data from my computer using the Apache server. However, I was wondering if there are many risks in this. For example: Could someone get full control of my computer this way? Is my computer more exposed to secutiry risks? Could I (practically) get DDOSed? The last question might be exaggerating, but I'm just curious. Thanks.

    Read the article

  • I need a form to tie MySQL Records together (i.e., many-to-many relationships).

    - by John R
    I need a HTML-based form to tie database records together. The user scenario that I envision is somthing like this: A super-user does a search of the database and is delivered a table of records based on that search; each record is numbered with the database-record id. There would also be two text fields next to each record. These text fields would allow: the user to make a reference from one record to another. For example, a user could enter into record id #457 that there is a correlation with record ID #242. The user would also describe the type of relationship it is. This could be with a simple integer that represents the type of relationship between the two records. When the super-user hits the submit button, all of these relationships are saved in a mySqljoin table. One option is to give me advice on how to implement and code this myself in PHP. However, before I reinvent the wheel, another option is to lead me to a free script that does something similar.

    Read the article

  • Restricting URL w.r.t HTTP method and setting different authentication mechanism for each

    - by user31745
    I shall start with an example. I want to restrict to POST requests only for http://path/to/logical/abc.xml and restrict to GET only for http://path/to/logical/def.xml. How do I put constraints like this as the paths are logical and location directive is not supported in .htaccess? The actual problem is to set different authentication type(basic, digest) on diff logical file.for eg. for abc.xml I want to authenticate for Basic type of authentication and def.xml with digest.

    Read the article

  • CLI Shared Hosting Management (scripts to manage web users and hosts) [on hold]

    - by aularon
    I am currently administrating two servers: first one has no control panel, I am creating directory structure, setting permission, configuring different aspects (users/php-fpm pools, nginx hosts..) for each of sites. With more clients, I sat up ISPConfig on my second server, everything is easily handled by ISPConfig Web Interface. However, I am searching for a CLI based solution, i.e. a set of scripts to create and manage hosts. Basically, a method to control ISPConfig from the command line (so I can use it over SSH) would be a good start. Does anybody know of such effort? I searched but all I got was web based solutions. Thanks.

    Read the article

  • Where is the best and easiest place to store global website settings [migrated]

    - by Darkcat Studios
    I'm just putting together an internal data system for a client, which is ASP.NET VB, backed by an SQL database, on an in-house IIS7.5 dedicated server. I want to store certain global settings, such as the age limit for news articles, admin contacts etc, in a file NOT within the database (i.e. to avoid unnecessary database query's) Where would the best place to store this be? global.asax? app.config? or a custom XML file? how would i import these at runtime? (probably to session variables) ETA: Also - the settings must be editable from within the site, i.e. an admin section i'm yet to build

    Read the article

  • cookie not being sent when requesting JS

    - by Mala
    I host a webservice, and provide my members with a Javascript bookmarklet, which loads a JS sript from my server. However, clients must be logged in, in order to receive the JS script. This works for almost everybody. However, some users on setups (i.e. browser/OS) that are known to work for other people have the following problem: when they request the script via the javascript bookmarklet from my server, their cookie from my server does not get included with the request, and as such they are always "not authenticated". I'm making the request in the following way: var myScript = eltCreate('script'); myScript.setAttribute('src','http://myserver.com/script'); document.body.appendChild(myScript); In a fit of confused desperation, I changed the script page to simply output "My cookie has [x] elements" where [x] is count($_COOKIE). If this extremely small subset of users requests the script via the normal method, the message reads "My cookie has 0 elements". When they access the URL directly in their browser, the message reads "My cookie has 7 elements". What on earth could be going on?!

    Read the article

  • Looking for open-source solutions for a knowledge sharing website

    - by Bundarr
    There is a need in my workplace to setup a knowledge sharing site, a place where users can discuss projects they are working share documentation ask questions I am looking for a open-source system that answers these needs, and that can be setup in a week, and requires only PHP and MySQL. I am a Wordpress fan and developer and could easily implement such a system in Wordpress, however this system needs to be very simple to use for the technically challenged. Without customization, Wordpress users would still need to login to the "back-end" to post. I like the Stackexchange (OSQA) format, but these do not allow for file uploading out of the box. I do not have experience with Buddy Press, would this be an alternative?

    Read the article

  • Move old website with bad SEO to other domain: redirect or start from scratch?

    - by fox_mulder
    I have an old website with very bad SEO and PR, but still with some visitors per-day. Lately I've rebuilt the app and the concept of it and I'm going to deploy everything in a new domain. I would like to not lose those old visitors but I'm wondering if redirecting that old website to my new one is going to damage the SEO in the new website. Don't know also if it deserves to make the move in the Google Webmaster Tools and Bing or better start from scratch.

    Read the article

  • Appcache and jquery mobile on a CMS powered site?

    - by user793011
    Has anyone used the cache manifest to make a CMS site work offline? I've made a demo with static html files which seems to work fine, so I'm assuming it wouldn't be too hard to achieve the same thing with a CMS. The way that you tell browsers that files have changed (and so need to be downloaded again) is by adding a comment to the cache manifest file so its byte size changes. I'm not quite sure how to do this with a CMS, but maybe some sort of server cron could run periodically? Personally I'm more interested in having a site that works offline rather than achieving ideal performance, so if the file was modified every hour rather than when content actually changed that would be fine for me. If anyone has used appcache with a CMS, has anyone done so with jquery mobile at the same time? What I'm after is a fully native feel to a site that's accessible offline, in other words I want to mimic a native App. My static demo does this perfectly with jquery mobile, so again I would have thought this would be achievable in a CMS.

    Read the article

  • How do I server multiple domains from the same directory and codebase without my configuraton breaking when apache.conf is overwritten?

    - by neokio
    I have 20 domains on a VPS running cPanel. One public_html is filled with code, the remaining 19 are symbolic links to that one. (For example, assets is a directory within public_html ... for the 19 others, there's a symbolic link to that directory in each each accounts public_html dir.) It's all PHP / MySQL database driven, with content changing depending on the domain. It works like a charm, assuming cPanel has suExec enabled correctly, and assuming apache.conf does NOT have SymLinksIfOwnerMatch enabled. However, every few weeks, my apache.conf is mysteriously overwritten, re-enabling SymLinksIfOwnerMatch, and disabling all 19 linked sites for as long as it takes for me to notice. Here's the offending line in apache.conf: <Directory "/"> AllowOverride All Options ExecCGI FollowSymLinks IncludesNOEXEC Indexes SymLinksIfOwnerMatch </Directory> The addition of SymLinksIfOwnerMatch disables the sites in a strange way ... the html is generated correctly, but all css/js/image in the html fails to load. Clicking any link redirects to /. And I have no idea why. I do have a few things in my .htaccess, which work fine when SymLinksIfOwnerMatch is not present: <IfModule mod_rewrite.c> # www.example.com -> example.com RewriteCond %{HTTPS} !=on RewriteCond %{HTTP_HOST} ^www\.(.+)$ [NC] RewriteRule ^ http://%1%{REQUEST_URI} [R=301,L] # Remove query strings from static resources RewriteRule ^assets/js/(.*)_v(.*)\.js /assets/js/$1.js [L] RewriteRule ^assets/css/(.*)_v(.*)\.css /assets/css/$1.css [L] RewriteRule ^assets/sites/(.*)/(.*)_v(.*)\.css /assets/sites/$1/$2.css [L] # Block access to hidden files and directories RewriteCond %{SCRIPT_FILENAME} -d [OR] RewriteCond %{SCRIPT_FILENAME} -f RewriteRule "(^|/)\." - [F] # SLIR ... reroute images to image processor RewriteCond %{REQUEST_URI} ^/images/.*$ RewriteRule ^.*$ - [L] # ignore rules if URL is a file RewriteCond %{REQUEST_FILENAME} !-f # ignore rules if URL is not php #RewriteCond %{REQUEST_URI} !\.php$ # catch-all for routing RewriteRule . index.php [L] </ifModule> I also use most of the 5G Blacklist 2013 for protection against exploits and other depravities. Again, all of this works great, except when SymLinksIfOwnerMatch gets added back into apache.conf. Since I've failed to find the cause of whatever cPanel/security update is overwriting apache.conf, I thought there might be a more correct way to accomplish my goal using group permissions. I've created a 'www' group, added all accounts to the group, and chmod -R'd the code source to use that group. Everything is 644 or 755. But doesn't seem to be enough. My unix isn't that strong. Do you need to restart something for group changes to take effect? Probably not. Anyways, I'm entering unknown territory. Can anyone recommend the right way to configure a website for multiple sites using one codebase that doesn't rely on apache.conf?

    Read the article

  • How long before Google will update search terms matching my website?

    - by Camran
    I have a website which title I changed about a month ago. The website is a classifieds website which is dynamic, using php. The title changed from "Free classifieds" to "buy and sell free classifieds". The strange part is that after about two weeks the title showed in google search results changed to the new title, BUT when I searched for "buy and sell free classifieds" my website didn't show up at all. I mean I have gone through over 30 pages of search results and my site isn't listed. However, searching for "free classifieds" still display my website at the same position it was before the title change. Any reason for this? How patient should I be? FYI the website has a sitemap submitted and updated, good meta tags and is W3 valid etc etc, so that is not the problem here. Thanks

    Read the article

  • How to delete Analytics property from list in Webmaster Tools?

    - by toxalot
    When I look at the Google Analytics Property page in Webmaster Tools (where you associate a Google Analytics web property with a Webmaster Tools site), I see a list of a bunch of properties with weird names. They are test properties from when we were trying things out years ago. I can't view them or change their name in Analytics because they don't exist there. All their profiles were deleted years ago. Is there a way to remove these from the list in Webmaster Tools? Or, alternatively, a way to view them again in Analytics so I can give them a better name, at least. I know this doesn't matter in the big scheme of things, but I hate clutter.

    Read the article

  • Difference between Global and Local SEO

    - by user29660
    I have been reading up on SEO techniques in an effort to learn how to do it thoroughly so I can charge my client for the service. To guage my price I have checked out competitor prices and noticed that theres a fair price difference when it comes to guarenteeing a page 1 ranking with global keywords compared to local keywords. So what is the difference in terms of work load and techniques used to justify this price difference? just to clarify, i am looking for technical differences in programming , methodology etc.

    Read the article

  • What are the minimum steps that I should follow to ensure that my web site is accessible to the disabled?

    - by Tim Post
    I am trying to follow a very important standard that I must admit I have ignored up until recently. I want to make sure that my pages are accessible to a large portion of people that have disabilities. I focus mainly on tutorials that are text and image intensive, but no video / flash or any kind of animations. What is a checklist that I can follow to ensure that many people with disabilities can have a good experience when using my web site, and what disabilities should I be most conscious of? I know that I can't possibly please everyone. I have gone through the W3C guidelines, however I'm not entirely sure what standards apply to me. I'm not building web applications, I'm building mostly wiki like information exchanges, blogs and the occasional forum.

    Read the article

  • mailing for categories

    - by nerkn
    There are more than 10 categories in my site, users can register more than one category. My php script prepare contents of each category, according to user preference my script merge those contents for each user. so every user can get what category they want. Problem is I want to send 340+ mails per hour and dreamhost dont allow. What do you suggest? I think to a service like mailchimp but I couldnt find that scenario, do they support category & content etc. Can I use smtp in dreamhost?

    Read the article

  • How can I allow search engines to index my invite only website in ruby on rails?

    - by tstyle
    I have a ruby on rails website that will be in invite-only mode for the next couple of months. Currently I have it set up so visits to any page performs an authentication: before_filter :authenticate, :except => [:beta] //authenticate checks for a logged in user But the webpage has a lot of content that I would like to see indexed by search engines, and I was wondering if there's an easy way to allow crawlers to do their work? I am not very knowledgable on SEO related stuff at all, so sorry if this is an suboptimal way to phrase the question.

    Read the article

< Previous Page | 255 256 257 258 259 260 261 262 263 264 265 266  | Next Page >