Search Results

Search found 8284 results on 332 pages for 'trusted sites'.

Page 161/332 | < Previous Page | 157 158 159 160 161 162 163 164 165 166 167 168  | Next Page >

  • Redirect from domain to other one

    - by Michal
    I deal with the following problem. I am customer of one domain reseller, which has fair prices, fine administration, and I have all my domains registered by it. Recently I've created a new webpage using a free web service (those sites where you can create a simple webpages with some template after few clicks). This new web page has default address in the following form: "pagename.provider.cz", but I want to use my own domain "pagename.cz". And that is the problem, because the provider would assign domain name to my presentation only if I registered mentioned domain by him. That wouldn't be a problem, but he is three times more expensive then my favorite one. So I am thinking about registering domain "pagename.cz" under my favorite registrator and then making 301 PHP redirect from it to "pagename.provider.cz". Shall this affect (negatively) my domain ranking? Are there any catches which I shoud care about?

    Read the article

  • backup dedicated server runing ubuntu 10.04 and plesk 11.01 prior to update os to uduntu 12.04

    - by timmob
    i would like to backup my dedicated server which is my web server hosting various sites and email, so that I can update the os to Ubuntu 12.04, and basically restore back to 10.04 if things go wrong. I have a local machine that I can install 12.04 onto an then I was going to rsinc between the two, but I am fairly clueless when it comes to linux. I can ssh into the remote server and gain root access. can anyone explain if i need to backup the whole server hard drive or just some of the files? Thanks Timmo.

    Read the article

  • How do I run Microsoft Explorer 8 plus on Ubuntu?

    - by user169487
    First off, you should know I'm a Ubuntu newbie who has become passionate since my strangely low-priced Dell computer showed up with Ubuntu loaded on it and it could do almost everything I need from a computer. But now I've got a problem. I know, it's like asking how to use a cassette player on your computer, but in South Korea EVERY website is made to work with exclusively on MS Explorer PLUS Active X, which Microsoft finally killed with 9. Hence, I have to use 8, as there are some Korean sites I need to access for searching, banking, etc. So I see that the way to go is to load PlayOnLinux, but then I it tells me to delete these. Do I really have to do that? Should I just hit "Install Anyway"? Or would that be just asking for trouble?

    Read the article

  • Are HTTP requests cached? [closed]

    - by nischayn22
    Many HTTP requests are sent repeatedly by browsers on almost every page load, such as requesting the jQuery .js file etc. Since these are already used on too many sites doesn't modern browsers keep a cache for this? I am thinking of a system where the browser has a cached copy of the .js file used very very frequently. On a new request for the .js file, it sends the server a request for a hash of the .js file (provided the server can reply to that) and compares the returned hash with the cached copy's hash... rest is intuitive.

    Read the article

  • What Part of Your Project Should be in Source Code Control?

    - by muffinista
    A fellow developer has started work on a new Drupal project, and the sysadmin has suggested that they should only put the sites/default subdirectory in source control, because it "will make updates easily scriptable." Setting aside that somewhat dubious claim, it raises another question -- what files should be under source control? And is there a situation where some large chunk of files should be excluded? My opinion is that the entire tree for the project should be under control, and this would be true for a Drupal project, rails, or anything else. This seems like a no-brainer -- you clearly need versioning for your framework as much as you do for any custom code you write. That said, I would love to get other opinions on this. Are there any arguments for not having everything under control? Is this sysadmin a BOFH?

    Read the article

  • IP address and SEO

    - by Joel
    Hello, I currently host 5 websites within a dedicated server I own. I have several questions: Does it matter if I host all my sites on 100.100.100.100 (the server's IP for example) or if I split them into 100.100.100.100, 100.100.100.101 ... 100.100.100.104 (that is, each site on its own IP). Does it matter if I use a C-Class for each website? Do search engines really care if your site has its own c-class? Do search engines penalize a website if it moves its IP? Thanks, Joel

    Read the article

  • Strategy for image sizes

    - by MotiveKyle
    I run a site that has a lot of writers that generate quite a few articles a day. I require them to provide two image sizes (one for the big headline image and one as the thumbnail). I've been wanting to change up the site layout a bit, but I am becoming limited by the image sizes for the posts. I have considered just cropping images, but they still need to look nice, and cropping doesn't always generate what I'd want. I'd prefer to just scale down by percentage (as I do with thumbnails). Should I just make the writers provide more images? How do other sites handle this?

    Read the article

  • Domain in PENDINGDELETE, question about drop

    - by kcdwayne
    A domain I want is in the pendingDelete stage according to WHOIS. I have been monitoring it since redemptionPeriod, and it entered into pendingDelete 5 days ago today. After checking a few services (SnapNames, etc), they report it is scheduled to drop on the 11th (7 days, by my calculations). I'm not quite sure what to believe. The domain isn't highly valuable. It is to me and one other company. I can see no backorders placed on the big name sites, so I'm thinking of trying to get it without a backorder service. Any insight as to when it will actually drop? I've read 11AM-2PM PST, but I'm unsure. Thanks.

    Read the article

  • Does setting document.domain via script interfere with Google Analytics?

    - by Seth Petry-Johnson
    I have a site, www.example.com, that displays some secure content from forms.example.com in iframes. To enable cross-frame navigation, pages on both sites use JavaScript to set the document.domain to just "example.com". I am using Google Analytics on www.example.com, but the GA site is not showing any data. It indicates that the tracking code is found (the status icon is a green checkmark), but no data is reported. The GA profile lists the website as "www.example.com". Is this a supported scenario? Is my script interfering with the GA code in some way?

    Read the article

  • How to redirect from HTTPS to HTTP without warning message?

    - by user833985
    I have two web sites: one HTTP site and one HTTPS site. I will validate the credentials in HTTPS environment and will return to HTTP once authorized. The same is working fine in IE but in Mozilla im getting a warning which is given below. Although this page is encrypted, the information you have entered to be sent over an unencrypted connection and could easily be read by a thrid party. Are you sure you want to continue sending this information? How to overcome this warning message? Currently I'm posting from HTTPS aspx page using JavaScript to the HTTP page.

    Read the article

  • BleachBit: How to Completely Clear URL History in Firefox?

    - by tSquirrel
    14.04 / Firefox 29.0 I've been using Bleachbit to clear usage/file history, and for the most part it works great. However, it doesn't seem to clear the website hostnames out of the URL, at all. These addresses are not bookmarked. Also, the total URL isn't preserved, just the hostname. Visit site http://www.bluesnews.com/some_random_URL_string Exit Firefox Run Bleachbit, with ALL Firefox options selected Restart Firefox Check history: completely empty, other than bookmarked sites. www.bluesnews is NOT bookmarked Type "blue" which is Firefox automatically completes as "http://www.bluesnews.com/" Alternate Step #3: Use Firefox's built-in "Clear History" and select ALL entries with a time frame of "Everything". Same result as above. My inquiry in BB forums hasn't been responded to. I found Dan's proposed solution, however changing autocomplete in about:config only turns off the function, it doesn't actually stop storing URLs.

    Read the article

  • Naming interfaces for persistent values

    - by orip
    I have 2 distinct types of persistent values that I'm having trouble naming well. They're defined with the following Java-esque structure, borrowing Guava's Optional for the example and using generic names to avoid anchoring: interface Foo<T> { T get(); void set(T value); } interface Bar<T> { Optional<T> get(); void set(T value); } With Foo, if the value hasn't been set explicitly then there's some default value available or pre-set. With Bar, if the value hasn't been set explicitly then there's a distinct "no value" state. I'm trying to optimize the names for their call sites. For example, someone using Foo may not care whether there's a default value involved, only that they're guaranteed to always have a value. How would you go about naming these interfaces?

    Read the article

  • How to report abuse to website hosting company (GoDaddy)

    - by lgratian
    I'm not sure if this is the right place to ask such a question... Let's say that a website posted a picture of me, without my consent, and I want to be removed (it's something private, could compromise my career if it's seen by someone that shouldn't). I sent them an email asking nicely that they should remove it, but they didn't respond and the picture is still there. Using 'Whois' I found that the website is hosted by GoDaddy. Is there a way (an email address, for ex.) to report to GoDaddy that one of the sites they're hosting does something illegal and to force them to remove the photo? I searched the site and found nothing about such a thing. Thnaks in advance!

    Read the article

  • Do they ask too much on this job?

    - by user58404
    I am looking for web developer job and this job description caught my eyes. I am not sure how much they offer but I was wondering if anyone here meets all of their requirements? To me, that's a lot of knowledge. 2 to 4+ years experience building web sites and applications in a professional environment Strong working knowledge of HTML5 and CSS3 Strong working knowledge of JavaScript, jQuery, AJAX Working knowledge of Ruby on Rails or similar MVC framework Working knowledge of ExpressionEngine, Wordpress or similar CMS Experience administering a LAMP-based server Experience with cross-platform and cross-browser website testing Comfortable working with version control (preferably Git) Proficient with Adobe Photoshop, Illustrator, and Fireworks Comfortable working on a Mac Self-starter with excellent time-management skills with the ability to meet challenging deadlines Ability to work independently with minimal supervision Desire to work on a small team Bonus Skills: Experience deploying to Heroku or similar PaaS provider. Experience developing Facebook applications A strong sense of design Cool open source projects (send us your Github account!) Advanced working knowledge of server administration and website deployment. Java and/or .NET experience

    Read the article

  • How Web Optimization Services Work to Increase Your on the Internet Reputation

    SEO is a symbol of search engine optimization and that is the key to success from the enterprise. No site has meaning if it seriously isn't properly promoted. Anytime any web surfer is in look up of any certain merchandise, providers or data he makes use of the easiest method of browsing as a result of search engine optimization and this is habit of many individuals to only search straight into 5 or six major sites for their goal. No person has time to seem directly into 100 pages of internet search engine as there is no need to have when he finds in major pages.

    Read the article

  • Kaspersky decouvre un virus "fileless" qui s'installe dans la RAM et désactive le contrôle des accès utilisateurs de Windows

    Suite aux indications d'un chercheur indépendant qui « souhaite garder l'anonymat », le laboratoire Kaspersky a découvert un nouveau virus. Celui-ci se propageait via les annonces du réseau publicitaire Russe AdFox.ru, présentes sur des sites d'informations populaires. Ce qui rend ce virus particulier est son mode opératoire. Si l'inclusion d'une iFrame renvoyant vers un site contenant du code malicieux (hébergé sur un site .eu) est classique, la stratégie d'exploitation utilisée est par contre rarissime. Le virus utilise une vulnérabilité critique de la machine virtuelle Java (CVE-2011-3544, pour laquelle il existe un patch depuis six mois) ; mais contrairement aux a...

    Read the article

  • Do I need multiple accounts in Facebook for each of my product site?

    - by John
    I've a dozen sites which include for-profit ones as well as for charity. For each site I've created a Facebook company/charity account. After creating those accounts it dawned on me that I could as well have created a new page for each of my site from my personal account only even if a site has multiple product pages. What'll be the right strategy? Also as per Facebook terms we can have only single personal account. I do have single personal account only but for each site I've created only company pages. I hope I'm not violating the facebook terms.

    Read the article

  • Photos - do I really need to look for the author and ask his permission when posting them on my site?

    - by user6456
    When I find a photo somewhere on the internet, without any explicit information of whether I can re-publish it on my own website, without any hint of who is the owner/author of that photo, can I still do it? I'm puzzled here cause I've seen like millions of websites, often very big, that repost photos, most probably found via google and it's VERY unlikely they bothered to look for and contact the author of that photos. Is every one of that sites likely to be sued at any moment? What about the case of forums and content provided by users - there is virtually no way of prevention here.

    Read the article

  • How does delicious.com avoid being sued for copyright infringement?

    - by Stanish
    With the recent redesign of delicious.com, they've added a much more graphical home page. The site continues to be a service for people to bookmark and share websites they come across on the web. The delicious home is now made up of images taken from those linked sites. See for yourself at http://delicious.com I would like to know what in the law allows them to do this, considering the images represent the main content of the page, and they clearly do not own copyright to those images? I know there is some leeway given to search engines where it is considered fair use to use a small portion of the content if the aim is to lead people to the originating site. Does that apply here?

    Read the article

  • How to pass information across domains to ask for newsletter only once?

    - by Michal Stefanow
    Lets assume following scenario, I have two sites: example1.com example2.com When user visits 1 there is a prompt "please signup to a newsletter". Same thing happens when user visits 2. However when navigating from 1 to 2 I don't want signup form to be shown. My first thought were 3rd-party cookies, but it seems that they are blocked / not working: http://stackoverflow.com/questions/4701922/how-does-facebook-set-cross-domain-cookies-for-iframes-on-canvas-pages?rq=1 http://stackoverflow.com/questions/172223/how-do-i-set-cookies-from-outside-domains-inside-iframes-in-safari?rq=1 Another thought is to append #noshow for each URL but that would require some work - for instance a script that would intercept click / tap events and modify URL structure depending on the address. (but that seems hacky) I wonder if you know a robust well-established solution to this issue? Thanks

    Read the article

  • Is there a way to make money from indie downloadable games? [closed]

    - by AShelly
    It appears that there are ways to make money with flash games through portal and aggregator sites and embedded ads. But I do my programming in C and C++. I've started a prototype which relies on a few existing C++ SDK's. The game would have to be downloadable. Is this just a labor of love, or are there any ways to make money from this type of game? Does anyone pay for shareware anymore? What other options are there?

    Read the article

  • Website (X)HTML Code Change Detection [closed]

    - by 0pt1m1z3
    I am looking for an enterprise-grade service or a tool that can be used to scan / fingerprint websites and notify when major XHTML code changes are detected. The tool should be able to continuously scan thousands of websites and determine the percentage of HTML code that has been modified since the last run. And then either save the data where it can be easily accessed or send periodic notifications. I know of services like ChangeDetect.com, but they don't do markup only changes and instead focus on everything, including content. We don't really care about presentation content, because a lot of sites we need to cover are updated frequently with content.

    Read the article

  • Make the Web Fast: Automagic site optimization with mod_pagespeed 1.0!

    Make the Web Fast: Automagic site optimization with mod_pagespeed 1.0! Ask and vote for questions at: bit.ly mod_pagespeed is an open-source Apache module that automatically optimizes web pages and resources on them: images, CSS, JavaScript, and much more. In this episode, we'll catch up with Joshua Marantz, the tech lead of the project at Google and talk about the history of mod_pagespeed, its fast growing adoption (130K+ sites!), technical architecture and how it works under the hood. Finally, we'll talk about the upcoming 1.0 release milestone for the project. If you're curious about mod_pagespeed, then this is definitely the show you won't want to miss! From: GoogleDevelopers Views: 0 0 ratings Time: 00:00 More in Science & Technology

    Read the article

  • What payment processors fullfill this requirements and what are their pluses and minuses? [on hold]

    - by Sharen Eayrs
    Accept credit card Allow me to automatically credit a customers' account. We're not selling e-book. We're selling a credit to our account. So it's important that customers do not get credited twice. Easy to program and integrate with our sites. Have affiliate programs. Transfer money to bank accounts quickly Accept merchants from many countries No monthly fee is a plus. I am thinking of using 2co.com avangate.com clickbank.com Some people reccomend https://stripe.com/us/features (is it easy to implement) http://www.paymentwall.com/ There are so many payment processor I am very confused. Can anyone tell me pluses and minuses of those ones. or perhaps others. What would be the plus and minuses of those 3 that you know off

    Read the article

  • How do I make a subdomain of a subdomain?

    - by joe
    I have a server that hosts mydomain.com. Then I have a server that hosts subdomain1.mydomain.com. On the subdomain1.mydomain.com server, how do I make subdomains? For instance I want to create subdomain2.subdomain1.mydomain.com I have my subdomain1.mydomain.com site in /home/USER/public_html and it is working just fine. But When I try to create subdomain2.subdomain1.mydomain.com, it says Server not found in the browser. I have tried so many configurations in my /etc/hosts and /etc/apache2/sites-available/subdomain2.subdomain1.mydomain.com, without any luck Can I get some detailed instructions maybe? I have looked everywhere and every place tells you different, with no luck. Ubuntu Server 11.10 w/ the Lamp Stack option

    Read the article

< Previous Page | 157 158 159 160 161 162 163 164 165 166 167 168  | Next Page >