Search Results

Search found 27295 results on 1092 pages for 'cross site'.

Page 81/1092 | < Previous Page | 77 78 79 80 81 82 83 84 85 86 87 88  | Next Page >

  • Catch all DOM events on a page (without Firebug et. al.)

    - by Morten Bergfall
    How would I go about creating a cross-browser script, that would intercept all events firing on a page/browser window/DOM-tree (regardless of browser)? What I'm hoping to accomplish is basically to get a better understanding of the different handling of events in different browsers; I know the basic theory, but need to see to believe...

    Read the article

  • Is there a C++ graphing library?

    - by thyrgle
    Is there a C++ graphing library that can display visual graphs (such as hyperbolas and parabolas and linear equations) based on the equation it is given and that is cross platform? Or am I just asking for too much...

    Read the article

  • Client-side Twitter API method using XMLHttpRequest?

    - by user359239
    Does (or did) Twitter ever have a method in their API that allowed client-side HTTP requests without producing any cross-domain issues? I have a site that used the following url: http://search.twitter.com/search.json?q=from%3adaleyjem&rpp=3&page=1&callback=? It once worked, but now it doesn't. Is there something else I should be using, or do I have to make a request to my own server-side code?

    Read the article

  • looking for a proxy library

    - by gwegaweg
    i need to set up a proxy for cross-domain javascript. the problem is, i tried the simple fgets() , i discovered, javascripts are ignored and non-functional. is there a proxy library that handles all these ? it can be in any scripting language.

    Read the article

  • Moving .JS files to a CDN: How to manage AJAX requests?

    - by pagewil
    I am thinking of moving my static .JS files to a CDN such as Amazon S3 for performance reasons. As my PHP files and mySQL DB remain on my primary hosting domain what is the best way to manage my JS AJAX requests if they are now cross domain? Currently they look like this within my .JS file (with relative paths): $.ajax({ type: "POST", url: "/myNearbyPhpFile.php", data: {data:someData}, success: function($r){} });

    Read the article

  • Another getcwd question...

    - by rubenvb
    Sorry to keep hammering on this, but I'm trying to learn :). Is this any good? And yes, I care about memory leaks. I can't find a decent way of preallocating the char*, because there simply seems to be no cross-platform way. const string getcwd() { char* a_cwd = getcwd(NULL,0); string s_cwd(a_cwd); free(a_cwd); return s_cwd; }

    Read the article

  • SWT-like GUI toolkit for C

    - by user345409
    Hello, do you know any cross-platform gui toolkit like swt for C (using default widgets in each operating system = right pics on eclipse.org/swt) ? There is an implementation of swt for D language called DWT but I need it for C or C++. Thanks.

    Read the article

  • IE Problem in jQuery

    - by venkatachalam
    I am using the jQuery plugin Cross Slide. My page is working fine in Chrome and Firefox. But in IE 7, I get: Debug error as Object expected on line 1: $(document).ready(function() { $('#image').crossSlide({sleep:4,fade:1},[{src:'images/1.jpg'},{src:'images/2.jpg'}]); }); How can I fix bug in IE 8 & 7?

    Read the article

  • How do I remove a site from IIS7 using JavaScript?

    - by Petteri Vilmusenaho
    Hi, I've been searching the Internet for a way to remove a site from IIS7 using JavaScript. I've found, and used, a lot of examples on how to create a site and applications using JavaScript but not a single example on how to delete a site! For example, I've figured out how to create a site using the examples found at http://www.iis.net/ConfigReference/system.applicationHost/sites. Does anyone know how to remove a site from IIS7 using JavaScript? Please help! /Petteri

    Read the article

  • Use mod_rewrite to force users to homepage when entering a site?

    - by scotru
    Is it possible to use mod_rewrite to force all users entering a site (either through a link from another site, or by typing a URL in the address bar) to be redirected to the homepage? From the homepage (or any page within the site), users should then be able to access other pages in the site. But all users would be forced to enter the site through the homepage. Can this be done with mod_rewrite (or without using a scripting language)?

    Read the article

  • How to stop Feedreader fetching content from my site using iFrame?

    - by Wei Kai
    As you all can see from the picture below, my site's content is duplicated by FeedReader (using iFrame) and indexed at Google. When I clicked at the FeedReader link, it uses some sort of iFrame to draw content from my site live. At the meantime, my site traffic has dropped significantly, but I not sure if this is the reason. https://lh4.googleusercontent.com/-hc4pVwHvQoo/UGGcwVyRqYI/AAAAAAAAAIc/9m04UOwmfEk/s1600/1.PNG https://lh3.googleusercontent.com/-ljj6dV7xTik/UGGc0x4GiZI/AAAAAAAAAIk/3mZ6HiCiQ2w/s1600/2.PNG What can I do to prevent Feedreader to fetch my content to their site? Any help would be much appreciated. By the way, I'm using wordpress as my CMS. I have also highlighted this issue to FeedReader 2 days ago, but yet to get any reply from them.

    Read the article

  • Wikileaks : le créateur du site arrêté et ses avoirs gelés, les cyber-militants s'organisent et OVH s'exprime dans une nouvelle lettre ouverte

    Wikileaks : le créateur du site arrêté et ses avoirs gelés, les cyber-militants s'organisent et OVH s'exprime dans une nouvelle lettre ouverte Mise à jour du 07/12/10 Wikileaks est en train de passer du statut de site polémique à celui de site traqué. Une position qui, habituellement, attire de multiples sympathies dans la sphère d'Internet. Et Wikileaks ne fait pas exception. De nombreux internautes, y compris ceux qui n'adhèrent pas spécialement à la philosophie du site, ont décidé de réagir aux tentatives visant à le mettre hors-ligne. Il existe par exemple à présent de nombreux miroirs (et donc d'hébergeurs v...

    Read the article

  • What bots are really worth letting onto a site?

    - by blunders
    Having written a number of bots, and seen the massive amounts of random bots that happen to crawl a site, I am wondering if the goal of the site allowing bots is for the potential for the bot to send real traffic back to the site if there is any reason to allow bots that are not known to be sending real traffic back, and how to spot these "good" bots; based on how they ID themselves, IPs they come from, behaviors, etc.

    Read the article

  • How do I get google to see keywords on a one page web application site?

    - by David
    I'm going to have to link to the web site to explain this, http://www.diagram.ly, it's a free service, so I hope this doesn't break advertising rules. Basically, it's a one page web application, I don't want to create a web site for it. Some background text loads and if JavaScript is enabled, the web application itself then loads. The problem is that Google only seems to be picking up the title of the page and the text on the footer, so the site only appear on Google search for very limited text (based on the title and meta description mostly). I was hoping that search engines would pick up on the background text and index that. The text is factual, not keyword stuffed. Yahoo seems to pick up the text, just not Google. Does anyone have any experience of how Google would view such a site and where I could put the text for a better result? Edit I should mention that Google Webmaster Tools lists the site keywords as "Component, diagramly, feed, mxgraph, share and twitter". Basically the footer and little else.

    Read the article

  • My parked domain was de-indexed by Google - what to do?

    - by Programmer Joe
    I have a question about how to handle my domain. In a nutshell, I bought a domain last year from Go Daddy. My intention was to launch a real site with this domain and I have spent the last year working on my site. For the last year, I have been using the default Go Daddy page display for an up and coming site. When I first bought this site, it was indexed by Google - you could search for "alphabanter" and my site would show up on the search result page for Google. Several months ago, it seemed Google de-indexed my domain and if you type "alphabanter," my domain no longer shows up on the list of search results. However, if you search for "www.alphabanter.com", that's the only way it shows up in the search results for Google. Anyways, I am about to launch my site for real. However, I don't quite know if I can get my site back into Google's index. I have a few questions: 1) Was my domain permanently penalized by Google and removed from their index just because it was a parked domain? I don't believe I have done anything abusive other than using the Go Daddy default page for almost a year because my site was not ready. 2) Should I just launch my site, put a few backlinks to my site, and hope that Google indexes my site again? 3) Should I submit my site to Google at Google submit your content I assume getting Google to reconsider my site is the last option if none of the above works.

    Read the article

  • How to block FeedReader from fetching my content to their site?

    - by Wei Kai
    As you all can see from the picture below, my site's content is duplicated by FeedReader and indexed at Google. When I clicked at the FeedReader link, it uses some sort of iFrame to draw content from my site live. This forms some sort of content duplication to me, and I believe it does harm to my site. (Stackoverflow doesn't allow me to post image due to new account, pls click at the link above to see the picture, million thanks to you.) What can I do to prevent Feedreader to fetch my content to their site? I know robots.txt can perform such function, but I don't know how to do it. Any help would be much appreciated. I have also highlighted this issue to FeedReader 2 days ago, but yet to get any reply from them.

    Read the article

  • Is Wordpress a good CMS for a Event Site? [closed]

    - by Roland
    I plan on building a gallery/exhibition event site. so Locations are usually always the same an fall into 3 categorys (gallery, offspace, institution). then there is the Exhibition title the date and the participating artists. So I was wondering if Wordpress could handle such a site. it should be very data driving though, so all the information is in a list view on one site and can be ordered and queryed (which artists took part in which exhibitions and so on) Please tell me the cons and pros of using Wordpress for such a site and problems I could run into if I might plan to broaden the scope later on. thanks!

    Read the article

  • What if you've been asked to develop a site and the client later introduces Ts&Cs that you'll breach whilst doing your job?

    - by Matt Lacey
    Disclaimer : this is all made up. Honest. And it represents no clients or employers living or dead, blah blah blah, etc. [Allegedly] As part of a website I've built, I've now been provided the Terms and Conditions of site usage to display on the site. These terms--which must be agreed to to access the site--include my (or any visitor to the sites) compliance with a number of clauses. Many of these clauses refer to general computer use and are not tied specifically to use of the site. Some of these clauses refer to things I have had to previously do as a legitimate part of my job and would expect to have to do again. When I've raised similar issues previously my line manager has said just to ignore it but that doesn't seem to be the professional thing to do. So, what do I do? Abiding by the terms would mean that I could no longer work on the project and would cause issues with my employer and the owner of the business the site is being created for. Ignoring them could lead to possible future issues with the business owner and is not something I'm necessarily happy with (the deliberate breaking of a legal contract). Neither option is one I'd choose and could have major consequences. Any thoughts?

    Read the article

  • How can I view localized versions of my site?

    - by Max Vernon
    We are adding internationalization to our site. We are getting the client's IP address from the headers and looking it up against the IP2location database to get the client's country. Several of our clients reported seeing a blank page over the weekend. We'd like to be able to get screenshots or use a browser from many different countries on an ongoing basis for testing code changes. I need to know what the site looks like when accessed from various countries since there are several elements that vary by country. I've used Tor and Vidalia, along with the Tor customized Firefox browser however it appears the CSS is getting mangled. I have also used http://webpagetest.org to check the site, however the screenshot it gives is too small to be really useful. Is there a site or a service I can use to get screenshots or interact with my website from various countries?

    Read the article

  • Should I give preferential treatment to proxy users on my ecommerce site?

    - by Question Overflow
    I am setting up an ecommerce site that caters to a worldwide audience. I would imagine that visitors would come from everywhere, and for whatever reasons, some would be connecting through proxy servers. My site uses a server that is configured to rate limit connections from the same ip address to protect itself from a DOS attack. So, if a proxy server is heavily used by my visitors, then it would appear to be a DOS. This is problematic in a sense that it is hard to tell whether the users are genuinely browsing my site or if a DOS is taking place. So my question is, should I give preferential treatment to proxy users on my ecommerce site? If yes, how should this be done. If not, why not?

    Read the article

< Previous Page | 77 78 79 80 81 82 83 84 85 86 87 88  | Next Page >