Search Results

Search found 14789 results on 592 pages for 'pro backup'.

Page 328/592 | < Previous Page | 324 325 326 327 328 329 330 331 332 333 334 335  | Next Page >

  • http request terminating early

    - by spiderplant0
    I noticed that on some of my sites, images were occasionally not getting downloaded fully. After a bit of investigation it appears that it is not restricted to images - .css, .js etc were also occasionally terminating early. The faults appear to be random. When I use the debug/proxy tool Fiddler2 reports that fewer bytes have been received than were requested. Firebug reports "Image corrupt or truncated". Obviously this is mainly a concern between me and my hosting company. However despite many emails they have not been able to get to the bottom of it. Transfer to another hosting company is obviously an option but is really a last resort. Has anyone seen this kind of thing before or can anyone suggest what might be causing it. Or any apache setting or something that I can ask them to check out. Will apache log this kind of error - they havent been able to provide me with any logs, but if I know exactly where things have been logged, maybe I can prompt them in to action.

    Read the article

  • How can I create a content widget generator?

    - by Richard
    Sites like Buzzfeed offer widgets (Javascript, PHP, WordPress, etc.) for syndicating their content on other sites. Does anyone have any ideas on how I could go about creating/implementing some kind of interactive widget generator that gives users options for customizing their widget? I assume this would require RSS. Check out the Buzzfeed generator to see what I mean http://www.buzzfeed.com/network/widget

    Read the article

  • Best CMS for review-type sites

    - by Pru
    Is there an ideal CMS for making a review site? By review site, I mean like a restaurant review site where you have each entry belonging to different major categories like Cuisine and City. Then users can browse and filter by each or by combination (Chinese Food in Los Angeles, with suggestions of other Chinese restaurants in LA, etc). Furthermore, I'd want it to support other fields like price, parking, kid-friendliness, etc. And to have users be able to filter by those criteria. I've been told that with a combination of custom taxonomies, plug-ins and many clever little queries, that Wordpress 3.x can handle this. But I'm having a heck of a time with it getting into the nitty gritty, and that's where I find the community support is lacking. The sort of stuff you'd think would work in WP, like making one parent category for Cuisine and one for City, don't really work once you get further in and start trying to pull it all together. Then you find these blog posts where people say, "This example shows that one could create a huge movie review site using custom taxonomies..." but when you go and try it you hit all sorts of challenges and oddities that point a big long finger at Wordpress being in fact a blogging platform. The best I came up with was one category for the cuisine and one tag for the city, then I created a couple of custom tag-like taxonomies for the other features. It's quite a mess to try to figure out how to assemble all of that into a natural, intuitive site. I expect a few versions down the road WP will be able to do these sorts of sites out of the box. So I thought I'd take a step back before I run back into the Wordpress fray and find out if maybe there is another platform better suited to this sort of relational content site. Directory scripts in some ways offer many of the features I'm looking for, but I need something more flexible and, hopefully, interactive (comments, reviews). I'm especially looking for feedback from people who've crafted sites like this. Thanks!

    Read the article

  • Website migration from WordPress to a static site and doing 301 redirects without access to existing site?

    - by user3114468
    Currently working on a project that is a hosted on WordPress that is being migrated to a static site. However I presently do not have access to the existing site as it's managed by another developer. The concern is not the lack of having access to content as the site owner has generated very little content (reason for migration) and we were able to do this manually. Rather the concern is to do 301 redirects. The site will not change domains but URLs such as from example.com/?page_id=3 to example.com/services. To add, the site is migrating to new server using same domain name. I thought maybe this could be done via editing permalinks prior to migration and WordPress would update automatically if configured to write on server. But if not configured (as this is not always the case) I do not have htaccess to fix it in case there are suddenly a bunch of 404 errors for every page. Really could use some help on the best procedure to follow in this case. This is the first migration project I've worked on.

    Read the article

  • Help with URL Rewrite

    - by bodesam
    This is the first time i'm doing this and have been doing some research on it. I have a page that selects some info from a database and displays it with a link to a second page that uses the result to query the database, something like this: $sel=mysql_query("select id, title from thetable "); while($row=mysql_fetch_array($sel)) { $id=$row['id']; $title=$row['title']; echo "<a href='more.php?id=$id'>$title</a>"; } The issue is, in the more.php page, instead of more.php?id=5 to show in the address bar, I want something like more/title Secondly, as it obtains in most sites, I want the link on the referring page to show this friendly url on mouse hover not the more.php?id=5 And I notice in most sites some words like 'a', 'and', 'the' etc are usually removed from the url title(even if there originally), moreover how does one handle the situation where more than one record have the same title. How does one go about achieving this url rewrite with htaccess or whatever method is used. Thanks.

    Read the article

  • Hosted News Carousel

    - by user19240
    I'm trying to add a news carousel on my OneHub site. I know there are quite a few plugins that I can use, but I'm looking for a ready-to-use hosted solution where people can log-on and manage the content, and I can simply include as an iFrame widget on our OneHub site. We would prefer to use Expression Engine, since we're using this on some of our other websites. It looks like Expression Engine + JQuery Cycle Rotator might be an option, not sure how easily I can get this onto OneHub though... WordPress has a carousel plugin as well that I'll explore. Anyhow, while I'm trying out these options, I figured I would put a question out there for any thoughts or advice, other solutions, etc. Any help would be greatly appreciated! Thanks, Alan

    Read the article

  • Recommanded cloud/cdn for my website? [closed]

    - by Tom
    Possible Duplicate: How to find web hosting that meets my requirements? Currently we are using Cloudflare (Business) and it's great, except that our website is crushing for around 1 minute every 2 hours on average, and it seems they don't like to give us support. So, we are looking for alternative. I thought about Incapsula. Currently we have around 1mil pageviews daily. Can you guys please recommand for us what to use/do? Thanks!

    Read the article

  • Storing data on server [closed]

    - by Maciekp
    1.How am I supposed to store data on server, using not only: databases,text files and images? And how someone could implement storing data in fb's graph api http://developers.facebook.com/docs/reference/api/ , so when I go to: https://graph.facebook.com/19292868552 it shows me such data(how it can be stored? I guess it's not Mysql database) PS. Link to article: http://jayant7k.blogspot.com/2009/05/how-facebook-stores-billions-of-photos.html <- How can concurrent users writing requests be solved(while storing data in text file).

    Read the article

  • Getting web results URLs in millions [closed]

    - by tereško
    I looked at all sites of SO and couldn't find any suitable to ask this question but posting here as nearest match to scenario After 1 months research I basically give up on getting all URL's from a search results programmatically, I looked at Google Search API to find a way to get millions of search results "URL's" to be specific to a text file or something relative but no success, but I am 100% there must be a way or trick of doing it. Real Question : Is there anyway programmatically or manually I can get 1000+ search results (URLs using search query e.g. "Apple" returns million of results on google and I want as much as possible URLs of them results in a text file)

    Read the article

  • Convert Microsoft Word documents (.doc/x) into HTML files

    - by danie7L T
    Does anybody knows of a good application to get it done quickly and efficiently ? I bought Word Cleaner but the results are merely sufficient and I need go over all the generated html files to clean tons of useless injected tags like <strongH</strong<strongell</strong<strongo </strong<emWor</em<emld</em Most of the articles displayed on a website I manage are based on documents written on MS Word by people how has little idea of what are margins for or ordered/unordered lists, foot/end notes etc and I cannot make them use something else. Does anyone has a tip to help me handle those pages more efficiently than going over them to correct and apply my CSS style ? NB: Just for the record, using "Save as HTML DOC" in Word is faaar worst than Word cleaner

    Read the article

  • Canonical redirection meta tag [duplicate]

    - by sankalp
    This question already has an answer here: How to use rel='canonical' properly 2 answers There are two pages in my website with the same content; only the URL's are different: www.websitename.com and www.websitename.com/default.html. Someone suggested that I should add canonical tags to avoid them being considered as duplicate content. Where should I add canonical tags and why?

    Read the article

  • How do you enhance your websites speed without compromising the design and access?

    - by Thorn007
    How do you enhance your websites load speed without killing the design and accessibility? File compression, CDN, Gzip? What are the best tools for doing so? For example, Google has optimized their site without compromising the design. Also, many website can kill the purity of their images with compression. Is there a way, more or lest best practice, to increase speed without compromising the design and accessibility? Note: sorry for being so vague but I don't know how else to phrase this question.

    Read the article

  • Tracking a single page on another domain in Google Analytics

    - by Ross
    I have access to edit a 'mini-site' hosted on our organisation's parent site. I'd like to track this page using Google Analytics, however I don't have access to the front page so I can't verify this as my domain. Using the tracking code for our main site works, however I don't want this data to be confused with similarly named pages on our site (for example, our mini-site is at /radio, and if we had a /radio at our main site this would be counted as the same). Has anyone been in this situation before? I'd like to just redirect visitors to our mini-site to our main site, seeing as it ranks higher in Google, but I've been told to maintain a separate site with our main features.

    Read the article

  • How to maximize SEO of an internationalized website?

    - by alditis
    I am developing a website that should be multi languages. I would like your opinions about what the best way to do it, considering get the most SEO possible. I have these alternatives: Altarnative 1: Separate domains http://www.miweb.com -- English for default http://www.miweb.com.fr http://www.miweb.com.es http://www.miweb.com.it Altarnative 2: Sub domains http://www.miweb.com -- English for default http://fr.miweb.com http://es.miweb.com http://it.miweb.com Altarnative 3: Sub folders http://www.miweb.com -- English for default http://www.miweb.com/fr http://www.miweb.com/es http://www.miweb.com/it I would like to know about your experiences. I hope the question is okay. Any suggestions, comments, objections or idea is welcome. Thank you.

    Read the article

  • Impact of migrating home page from http to https on search results

    - by 2Stroke SEO
    Guys - I've just had to change one of my site's home page to https from http. I had plenty links coming into the http page and was performing well in Google against many of my targeted search phrases. I did a 301 redirect from the http page to transfer the link juice to the https page (and to prevent duplicate content issues), but my search rankings have tanked which indicates no link juice has been transferred. My PR has vanished - which I'd expect - but I'm really surprised that the SERP rankings fell off the face of the earth. Anyone have any ideas how I can recover this. I've waited a couple of months since the changes took effect just in case Google was taking time to check it out.

    Read the article

  • How to estimate what kind of server resources you will require at launch?

    - by Anonymous -
    How can I go about/what is the best way of estimating what kind of server resources I will require at the launch of a new project. Let's say for instance I'd gathered 10k users email addresses that opted-in to be informed when we launch. I didn't really want to go for a cloud solution at the moment. Should I just get a reasonably-specced dedicated server and go from there, or should I be looking at services like Amazon EC2? Or should I start out with EC2 for the fully-flexible resources, then once things stabilize I'll be able to sort a fixed hosting deal that is adequate elsewhere? Sub-question: What's a good way of estimating what load a particular application will place on the server per user?

    Read the article

  • Customer Reviews on Company Listings [closed]

    - by GSTAR
    I'm not sure if this is the right place but I am after some general advice on a feature I am looking to implement on my website. The website I have is a Wedding Directory. Here, companies can advertise in the form of a directory listing. This listing contains the company details - such as what they do and how you can contact them. There are three packages available - Basic package is free and Silver and Gold packages are charged for. Now in order to further enhance the directory, I want to enable customer reviews for each listing. This is basically whereby customers who have used a particular company's service can write a review (and rate out of 5) of their experience. The dilemma I face is that if customers are to be paying for their listings then surely they will expect to not have content on their listing that would tarnish their reputation (i.e. negative reviews). But at the same time I want to be impartial and help my site visitors to make informed choices on which companies to book when arranging their weddings. Of course I will exercise my ability to remove offensive or fake reviews but I do not want to be removing reviews just to satisfy my clients. Suppose a client pays me a premium fee to have their listing on my front page and at the top of the listings. Now suppose that client gets lost of negative reviews - the fact that they have increased visibility means that they will also get increased bad publicity. This in turn means they won't renew their package and will most likely request the listing to be taken down. So, how can I get the right balance here and keep everyone happy?

    Read the article

  • Facebook Share Button and Counter no longer displaying any Count

    - by donaldthe
    Is it just me or did the Facebook Share button that displays the count of shares and likes just stop working over the past few days? The sharing still works, the count of shares no longer displays. The link that is generated look like this http://www.facebook.com/sharer.php?u= and the JavaScript file on my page is this http://static.ak.fbcdn.net/connect.php/js/FB.Share I haven't changed anything and this has worked for years

    Read the article

  • Show events AND pageviews in Google Analytics

    - by supertrue
    Each page on my site contains a file, and I have Google Analytics set up to track file download events. I would like to see what fraction of users who visit Page X download Page X's file. I can view number of events by page by clicking on Content » Events » Pages. But I can't figure out how to see both events and pageviews (or visits) at the same time. Visits and pageviews are not available in the Secondary dimension dropdown from the Events list, and Events are not available as a Secondary dimension in the regular traffic listing (Content » Site Content » All Pages). I want something like this: Page Pageviews Events 1. /section/mypage 1,000 123 2. /category/anotherpage 867 41 3. /about/download 88 7 Is there a way to get this in Google Analytics?—to view events and pageviews, by page, at the same time?

    Read the article

  • Boilerplate Terms & Conditions for web app? [closed]

    - by Louis Bataillard
    Possible Duplicate: What are some good resources for generating privacy policies and terms of use? I am just about done creating my first web application. Since the application stores some user data, I want to make sure that I can not be held liable should something bad happen to the site. I googled around but I could only find boilerplate T&Cs for websites, not for web apps. Does anybody know where I can find such a boilerplate agreement that I can use? I realize that this won't be 100% security, but it's better than nothing I suppose.

    Read the article

  • Apache FilesMatch regexp: Can it match by the cache buster 10 digit (rails generated) following the filename?

    - by ynkr
    According to the apache FilesMatch docs: The FilesMatch directive provides for access control by filename Basically, I only want to set an expires header for resources that have a 10 digit "cache buster" id appended to the name. So, here is my attempt at such a thing in my httpd.conf <FilesMatch "(jpg|jpeg|png|gif|js|css)\?\d{10}$"> ExpiresActive On ExpiresDefault "now plus 5 minutes" </FilesMatch> And here is an example of a resource I want to match: http://localhost:3000/images/of/elvis/eating-a-bacon-sandwich.png?1306277384 Now obviously my FilesMatch regexp is not matching so I am guessing 1 of 2 things is happening. Either my regexp is wonky or the '?1231231231' cache busting part of the file is not part of what apache considers part of the filename. Can anybody confirm and/or give me a way to cache only those resources that will not persist beyond the next deploy?

    Read the article

  • Need private personal access to ~three PHP pages

    - by Roger
    I would like secure access to the text output by three PHP scripts (the text output is JavaScript and html) . The security level is much less then financial data but important none-the-less. I have considered purchasing AND studying https and SSL certificates. Hostgator charges an extra $2/month for a private ip plus $50+ anually for a certificate. This is more then I want to spend for this project (time + money). Is there a simpler solution that is: less expensive easier to implement. I'm open to different approaches.

    Read the article

  • 301 redirect from a country specific domain

    - by Raj
    I originally started using a .do domain extension for my site, but later realized that this country specific domain would prevent us from appearing in search results for places outside of the Dominican Republic. We started using a .co domain extension and redirected all requests to the new domain using an HTTP 301. The "Crawl Stats" in Google Webmaster Tools shows me that the .co domain is being crawled, but the "Index Status" shows the number of pages indexed at 0. The "Crawl Stats" for the .do domain says that it's being crawled and the "Index Status" shows a number greater than 0. I also set a "Change Of Address" in Google Webmaster Tools to have the .do domain point to the new .co domain. We're still not appearing in search results at all even for very specific strings where I would expect to find us. Am I doing something wrong?

    Read the article

  • Adwords: is it possible to automatically copy changes in one campaign to other campaigns?

    - by Richard
    Is it possible to have identical campaigns except for one thing, and whenever changes are made to one campaign, the same changes are automatically made to others? I want to have identical campaigns running for different time zones, but for the campaigns only to run between particular hours of the day in each time zone. If I have a campaign for each time zone, when I make changes to one campaign, such as adding ad groups, or adding keywords, or changing bid amounts, etc, I want those changes to occur to the other campaigns also. Is this possible? Thanks,

    Read the article

< Previous Page | 324 325 326 327 328 329 330 331 332 333 334 335  | Next Page >