Search Results

Search found 9728 results on 390 pages for 'zee pro'.

Page 105/390 | < Previous Page | 101 102 103 104 105 106 107 108 109 110 111 112  | Next Page >

  • Share on: FB, Tweet, Digg, Linkedin, Delicious, My mother, ... it's just on fashion, or some real value?

    - by Marco Demaio
    Nowadays your site is not in fashion if you don't show at least a couple of share buttons like these: Is this just fashion, or do people actually get something good out of it? When I say "something good" I mostly mean something that you could measure, and not just the feeling that was good. Maybe I can better explain with an example: did you notice (in some way) that many people clicked on those links to share your page/s on those web 2.0 social sites? And in such a case on which social networks did you see they mostly share your pages? BTW I'm not talking about Google PR, i know all web 2.0 social sites use nofollow everywhere and even hidden links, so they are useless by themselves for PR. UPDATE: According to this video, Google's Alter Ego says that they now use in some way data from social sites in ranking. If this is true, it's obvious that the Share on button for FB, Tweet, etc are definitely of some values. But again my question is more about what you noticed in your real experience to be a direct benefit of adding those type of "Share On" links on your webisite? I.e. did you see more traffic coming in form FB, or some users who bought your products because of FB or Twitter? Or any other benefits? Thanks

    Read the article

  • Had anybody earned $0.25+ from each of a captcha (on your website) passing?

    - by vgv8
    I am a real dummy in web monetizing schemes. [ 1 ] informs: "Solve [Media] charges a fee of about 25 cents to 50 cents for each form that is filled out using a Type-In ad [captcha]... the company splits its fees 50-50 with the websites where the ads are placed" Honestly, I cannot imagine that someone (in its proper senses) pasy that much money for just one captcha passed. And how to understand these claims? http://www.solvemedia.com/images/ie9_aboutcalcount.png shows: Why would Microsot pay 0.25-0.5 USD for each entered string "Be part of the Beta"? Has any of webmasters (sysadmins) got those from deployed SolveMedia captchas on their websites? Is it scam? Because if to check the sites mentioned in http://www.solvemedia.com/gallery.html, that is, for ex., http://www.toyotanation.com/forum/register.php?do=register, the latter do not have such captchas. What do I miss? Cited: [ 1 ] Jennifer Valentino-DeVries "An Online Ad That’s Tough to Ignore" WallStreet Journal Blog SEPTEMBER 20, 2010 http://blogs.wsj.com/digits/2010/09/20/an-online-ad-thats-tough-to-ignore/

    Read the article

  • Does having multiple URIs mapping to the same resource help SEO?

    - by Brian Wheeler
    Let's say I have a site with products that have tags, if each resource is available at GET '/products/tagged/:tag_list/:product_permalink' Could that be better for SEO than just one permalink? For example a product tagged "tea" and "coffee" would be available at GET '/products/tagged/tea/:product_permalink' GET '/products/tagged/coffee/:product_permalink' GET '/products/tagged/tea/coffee/:product_permalink' GET '/products/tagged/coffee/tea/:product_permalink' I would imagine that google would appreciate this because it gives multiple URIs with different levels of detail about the product, but I cant really be certain. Anyone have any direct knowledge on the topic? --EDIT-- As John Conde points, this is a horrible idea. What about having the links on my site link to a route such as GET '/products/tagged/:full_tag_list/:product_permalink', and then any time a user changes tags just have a HTTP moved permanently status to the new URL. Therefore duplicate URLs would be highly unlikely and mitigated by the proper response. Would this be better?

    Read the article

  • Is there an 'off the shelf' platform for making a new website similar to Elance?

    - by user17747
    I am interested in developing a website that is similar to 'elance' but for a particular vertical. Is there an 'off the shelf' platform you can recommend for getting started with this, or would I need to develop this web service from scratch? When I write 'platform' I am referring to things such as 'shopify' for e-commerce sites, or 'ning' for social websites. I want to create a multi-merchant professional services site. The site would need to support functionality such as: allowing merchants to open and manage their own profile. merchants would accept payments from customers through the site. file transfers between merchants and customers. merchant ratings by customers.

    Read the article

  • Do I create new site or add to existing site?

    - by nitbuntu
    Hi, Suppose, as an example, I have a website with the address, www.cool-gifts.com and I'm getting regular sales and its a worthwhile site, but no great fireworks. After research I find that there is a great market for '2nd hand stuff' and I'd like to serve that market. Would it be best to add '2nd hand stuff' as an additional category of gifts in my existing site....or, since the 2nd hand stuff is a market in itself, would I be better off investing time and energy bringing up a whole new site (www.used-stuff.com)? If I had employees and financial resources, it probably would be a no-brainer...start a new site. But, what if you are a small guy, with limited resources? So...new site....or add to existing site?

    Read the article

  • Could not log-in properly but shows no error in joomla

    - by saeha
    This is what I did, I added variables in \libraries\joomla\database\table\user.php: var $img_content= null; //contains the blob type data var $img_name = null; var $img_type = null; then I added this code in \components\com_user\controller.php: $file = JRequest::getVar( 'pic', '', 'files', 'array' ); if(isset($file['name'])) { jimport('joomla.filesystem.file'); $fileName = $file['name']; $tmpName = $file['tmp_name']; $fileSize = $file['size']; $fileType = $file['type']; $fp = fopen($tmpName, 'r'); $content = fread($fp, filesize($tmpName)); //$content = addslashes($content); fclose($fp); $user->set('img_name', $fileName); $user->set('img_type', $fileType); $user->set('img_content', $content); } that works fine, but I found this problem in logging in with the new user with an uploaded photo, other user with an empty img_content field could login properly. What happens is when I log-in using the user with uploaded photo, it's not redirecting properly it just return to log-in, but when i log-in through backend using other user which is super admin, i could see that user which appears as logged in. I started saving the images in the database because I am having problem with the directory when I have uploaded the site. I think the log-in was affected by the blob type data in the database. Could that be the problem? What could be the solution? -saeha

    Read the article

  • Symbolic link not allowed or link target not accessible: /var/www on Ubuntu 11.04

    - by Jamie Hutber
    I am getting a 403 when i access http://mayfieldafc.local/ upon looking in the apache logs i am getting [Wed Nov 16 12:32:59 2011] [error] [client 127.0.0.1] Symbolic link not allowed or link target not accessible: /var/www I have what i believe to be the correct permissions set on /var/www. hutber can create and delete files, hutber being my user. I can also execute as program on this folder. in mayfields vhost its: <Directory /var/www/mayfieldafc/docroot> Options +FollowSymLinks AllowOverride None Order allow,deny Allow from all </Directory> I am pulling my hair out not being able to work on my sites with my work ubuntu install. I know of nothing else that could be effecting this. So any ideas?

    Read the article

  • 301 redirecting a blog's RSS feed URL?

    - by Marc Charbonneau
    I moved my personal blog from Wordpress to Ghost this weekend, which changes the RSS feed URL from /feed/ to /rss/. By default Ghost returns a 301 redirect for /feed/, which I've verified by checking the response header and looking at the logs: In Feedly though, new posts aren't being picked up (at least after 24 hours. I'm not sure if they might have a waiting period before updating the URL). What's the correct thing to do in this situation? Do I need to keep /feed/ alive instead of returning a 301? If so, is there a rewrite rule that would let me do this in nginx instead of having to modify the Ghost source code?

    Read the article

  • Should Site Title be Before or After Page Title?

    - by NickAldwin
    Possible Duplicate: Does the order of keywords matter in a page title? Apologies if this is a dupe. I tried searching, but didn't find anything specifically addressing this concern. When creating a large(ish) site, page titles usually reference both the site name and the current page name. However, it seems there are two main conventions: Bob's Awesome Site - Contact Page and Contact Page - Bob's Awesome Site I've looked around, and pages usually use one of the two variants above. Is there any reason to use one over the other? SEO/readability/usability/etc? I've thought about it, and have only come up with: Page first - Differentiates the tab when the browser is crowded with lots of tabs Site first - Immediately see the "parent" site, so to speak; more cohesive experience

    Read the article

  • Can thousands of backlinks from the same site harm PageRank?

    - by Dejan Pelzel
    I just noticed that one particular site has almost 7000 backlinks linking back to our website. The site is something like a news aggregator and for each post they created around 20 (sometimes much more) backlinks back to our page and they basically linked over 400 pages. I am beginning to get concerned that this amount of links might harm our page. They seem to have more backlinks to our page than all the other pages combined and more backlinks that our website has pages. We have seen a massive negative effect going on for quite a while and the PageRank seems to have dropped to None (Not even 0). But I am not sure when and why exactly that happened seeing that PageRank updates take quite a while to appear. The site linking to us is otherwise pretty reputable and doesn't seem to be having any problems with their rank. (PR 6 actually) I was thinking of using the Google disavow tool for this site, but I don't want to make things even worse. Do you think these are harmful? If so, how do I fix this? Thanks :)

    Read the article

  • Should my blog be directly on my website?

    - by steve
    I have my newly launched website at www.slicify.com (redirects to a secure subdomain). I currently have a separate blog on WordPress: slicify.wordpress.com for a couple of reasons: I don't really want to mix my site code (it's a complex ecommerce site written in ASP.Net) with blog code, for ease of maintenance etc. WordPress is already great at blogs - seems silly to reinvent the wheel by trying to integrate blog functionality into my site However is keeping my blog on a separate domain going to hurt me in terms of PageRank or traffic? FWIW: while it's early days, I can see from Google Analytics that a good deal of referral traffic is already coming from my WordPress site to my main site, so at least that seems to be drawing potential users in.

    Read the article

  • What's the difference between cheap and expensive domain registrars?

    - by Joel
    A few years ago I registered a domain with Network Solutions. In recent years I've been using cheaper services such as namecheap, powerpipe etc. Every time that I need to renew some of the older domains with Network Solutions I am surprised at how much expensive they are. What is the reason for the price differences between the services? Why should I use a service like Network Solutions if there are so many companies out there that offer domain registration for a very cheap price?

    Read the article

  • SEO Implications of blog on site versus offsite?

    - by Kelli Davis
    I recently added a blog to one of our company's websites, and was confident that this increase in content on the site would have only positive SEO results. My boss, however, feels that we should have instead located the blog off-site, on a blogging platform such as Wordpress, Typepad, etc., in order to generate a backlink (assuming we'd link from blogging platform back to website.) While I know that backlinks are important for SEO, isn't content creation equally, if not more, important? Granted, I'd be creating content either way, but I figured we'd get more site traffic by having the blog located on our site versus a separate blogging platform. Am I incorrect in my priorities here? Boss's TOP priority is increasing the ranking of our website, so maybe a backlink would be better...? If we do need to relocate the blog to an off-site platform, is there a blogging platform that is more conducive to SEO than others? Is there a platform from which backlinks would be more valuable than others?

    Read the article

  • Microdata Without Reviews

    - by user36562
    I have not been able to find a clear issue on omitting reviews from Microdata. I understand that Microdata values for reviews will default to a certain number when omitted, but I was wondering if it would be correct/acceptable to completely omit the review node completely. I can see where reviews and average "star" ratings would be of help to the end user, especially for things like recipes. However, what if there are no reviews for a product or application? To be completely clear - let's isolate this question to only software applications or extensions. What if a particular piece of software or extension was not featured on an "app store" or other site that provided reviews? Wouldn't the formats still be helpful by providing version number, download URL, compatible software - etc? Sorry for the lengthy background, but I just don't understand why it seems that reviews must be part of a Microdata markup. Or, am I wrong in this assumption?

    Read the article

  • Single paging Meta Tag Duplication [duplicate]

    - by Dominic Reigns
    This question already has an answer here: Should paginated content have unique meta tags? 1 answer How to make Google understand about single paging content issue ? A single page has got above 100 paging ? 1-10 of 199 Go to Page 12345678910 ... like this as mentioned above, Do I need to put fresh meta tag on each page ??? Help me out from this issue ?

    Read the article

  • Will small random dynamic snippets break caching

    - by Saif Bechan
    I am busy writing a WordPress plugin. Now most users have cache plugins installed, they cache the pages. I know also some webservers as nginx have php caching and whatnot. There are also things like memcached. Now I have to admit I do not know exactly how they work, if anyone have some good links on how they work I would be glad. Some links where it's explained simple, not to technical. Now the question. My plugin displays different statistics on posts, they are always different, will this break the caching of the page. To take is a step further, sometimes the statistics of the post needs updating, and there is a small javascript snippet added to the page. Now will these two action result in the page not caching, or am I ok.

    Read the article

  • Groups page is blank in SharePoint 2010 [migrated]

    - by Murali Ramakrishnan
    Sometimes it's very confusing how Sharepoint 2010 group creation works Here's a scenario we have been facing from a long time wrt groups in SharePoint 2010 We had requirement of creating a two custom groups followed by creating a custom site through programmatically, For the most case the scenario works as how it is excepted to work. but, out of 1/100 site creation process the groups creation fails, which means we were able to access the group and users associated with it through programmatically. but, when it comes to UI stand point if you try to access the specific group page from the site permissions page - SharePoint returns a BLANK WHITE Page... BLANK WHITE Page... nothing else... Ain't is this a Sharepoint 2010 issue. or anybody had this problem and fixed it. Kindly share your thoughts

    Read the article

  • Multiple 301 redirects, do search engines/viewers see them all?

    - by Karim
    I've put in place lots of different 301 rules to deal with numerous url changes. And for certain URLS there are 3-4 different 301 redirects landing the visitors to the new URL. I heard that 301 loses pagerank/linkjuice. ALl the 301 are onsite for the same domain. With a mix of php 301s and htaccess 301s. so for instance articles/news.php?id=2 --- articles/blog.php?id=2 [filename change] articles/* --- /* [subdir to root] /blog.php?id=2 --- /title-of-post [mod rewrite url change] so if you were to visit /articles/news.php?id=2 there will be two 301 redirects until you land on the /yellow-wellington-boots/, my question is does google see the intermediate redirects, or just the final page the 301's redirect to.

    Read the article

  • beginner - best way to do a 'Confirm' page? [closed]

    - by W_P
    I am a beginning web app developer, wondering about the best way to implement a "Confirm Page" upon form submission. I have heard that it's best for the script that a form POSTs to to be implemented by handling the POST data and then redirecting to another page, so the user isn't directly viewing the page that was POSTed to. My question is about the best way to implement a "Confirm before data save" page. Do I Have my form POST to a script, which marshals the data, puts in a GET, and redirects to the confirm page, which unmarshals and displays the data in another form, where the user can then either confirm (which causes another POST to a script that actually saves the data) or deny (which causes the user to be redirected back to the original form, with their input added)? Have my form POST directly to the confirm page, which is displayed to the user and then, like #1, gives the user the option to confirm or deny? Have my form GET the confirm page, which then does the expected behavior? I feel like there is a common-sense answer to this question that I am just not getting.

    Read the article

  • What does it mean that hosting ToS DOESN'T allows HOTLINKS and WHY is it like that?

    - by Michal P.
    Free hosting services often doesn't allow to use hotlinks. I'm not sure if I understand it well and what is the reason of such a disaproval. I would like to have pictures for my webpages in Photobucket e.g. which is allowed to have hotlinks and use those pictures using hotlinks on my sites. Is it that what is not allowed? What is a problem for free host server owners to accept such links? Bandwidth of Photobucket is used only as I understand and it is completely legal. I'v read quite enough abt hotlinks, but I can't understand this simple issue.

    Read the article

  • Why is "www.mysite.com" different from "mysite.com"?

    - by sapeish
    In any browser if I use www.mysite.com or just mysite.com the web page is correctly retrieved, but I am having trouble with Google Analytics and Facebook App. Facebook: To be able to get Likes, I create the Facebook App needed and set the site URL to http://mysite.com/. Using their tool http://developers.facebook.com/tools/debug/ when I test my page using http://mypage.com it works but using http://www.mypage.com fails with the message: Object at URL 'http://www.mysite.com/' of type 'website' is invalid because the domain 'www.mysite.com' is not allowed for the specified application id. Google Analytics: To be able to get traffic statistics, I created a Property and a Profile both with the URL http://www.mypage.com and no statistics were gathered in a week, when I changed the configured URL to http://mypage.com statistics where available a few hours later. What should I do to have statistics and likes for both www.mysite.com and mysite.com ??

    Read the article

  • Magento Bulk Product Import + Modules Nightmare

    - by mike
    Have 5,000 products in CSV file File has been re-saved as UT8 file in google documents and exported to CSV from excel File loads perfectly with all fields in demo of magento store manager (except we dont want to buy it:) When trying to upload in regular Magento..keep getting error messages on column duplicates....yes we have hundreds of duplicates as the titles of products in fields correspond with different sizes, etc...no way around it Any solutions around this or any open source software similar to store manager that can do the trick. Ready to give up and go to paid solution such as Big Commerce Also, Uploaded a bunch of free modules/keys...of open source bulk import products from magentocommerce but I cant find them anywhere in the main admin panel to use...there is no menu item for them anywhere??

    Read the article

  • Legal Web Fonts & Licensing

    - by Phill
    So I'm a little confused in regards to legality of using fonts for web. Often from designers I get a PSD file and it uses a special font, and the font is supplied. However attempting to convert the font using: http://www.fontsquirrel.com/fontface/generator Ends up with a message saying it's blocked by Adobe. Not only that sometimes if the font can be converted, it often looks like crap when viewed in a browser. I assume that the generator is primarily for people to convert their own fonts, but if you purchase the use of a font then you can only use it for Web if the terms allows you to? The fonts used in PSD's are often Adobe Fonts, I can't find anything that suggests I can convert those and use them on the web. So I'm wondering if anyone knows the legal rights around using Photoshop supplied fonts on the web? In addition I'm wondering what resources are available (free/paid) that provide fonts that can be used on the web. Free: http://www.fontsquirrel.com http://www.google.com/webfonts Paid: http://www.fontshop.com This is the only one's I've found so far that aren't cartoon type fonts like what's primarily on www.dafont.com

    Read the article

  • my wiki site using mediawiki - databases not found error

    - by Jayapal Chandran
    I had been using mediawiki opensource in my website to display programming articles. Today i tried to access my site but it showed database not found. Mediawiki uses many databases. When i logged in my control panel and checked i can see that most of the databases created by mediawiki is missing so is the reason i am getting this error. I have used mediawiki for two different purposes. It is like two modules. For one the databases are missing and for the other i think the data is corrupt. Do anybody know any issue with mediawiki security or would this be a problem with the webhosting cause we have faced several problems with them initially like before three years and recently the were good. Yet this happened. I have requested to hosting company to look into it and meanwhile i am expecting the help from stackexchange users. How do i check the logs for table deletion?

    Read the article

  • Which token from a long User-Agent should I use in robots.txt?

    - by Gaia
    The definition of User-Agent states that several tokens can be included, as deemed necessary by the client. I want to block certain bots via robots.txt and I am confused as to which part of the User-Agent string to use, especially for more obscure bots. For example: Mozilla/5.0 (compatible; uMBot-LN/1.0; mailto: [email protected])" JS-Kit URL Resolver, http://js-kit.com/ Mozilla/5.0 (compatible; SEOkicks-Robot +http://www.seokicks.de/robot.html Do I use the second token? Can tokens contain spaces, or did the SEOkicks folks forget a semicolon after SEOkicks-Robot? I don't actually intend on making my question specific to a couple bots - I want to know the guideline: which part of UA do I place in robots.txt for these exotic bots with UA as long as a haiku? User-agent: uMBot-LN/1.0 Disallow: / PS: Thank you but I do not need to hear that undesirable bots are better blocked with mod_security. I already have commercial mod_sec rules in place.

    Read the article

< Previous Page | 101 102 103 104 105 106 107 108 109 110 111 112  | Next Page >