Search Results

Search found 17138 results on 686 pages for 'facebook actionscript api'.

Page 130/686 | < Previous Page | 126 127 128 129 130 131 132 133 134 135 136 137  | Next Page >

  • Trouble Letting Users Get to Certain Sites through Squid Proxy

    - by armani
    We have Squid running on a RHEL server. We want to block users from getting to Facebook, other than a couple specific sites, like our organization's page. Unfortunately, I can't get those specific pages unblocked without allowing ALL of Facebook through. [squid.conf] # Local users: acl local_c src 192.168.0.0/16 # HTTP & HTTPS: acl Safe_ports port 80 443 # File containing blocked sites, including Facebook: acl blocked dst_dom_regex "/etc/squid/blocked_content" # Whitelist: acl whitelist url_regex "/etc/squid/whitelist" # I do know that order matters: http_access allow local_c whitelist http_access allow local_c !blocked http_access deny all [blocked_content] .porn_site.com .porn_site_2.com [...] facebook.com [whitelist] facebook.com/pages/Our-Organization/2828242522 facebook.com/OurOrganization facebook.com/media/set/ facebook.com/photo.php www.facebook.com/OurOrganization My biggest weakness is regular expressions, so I'm not 100% sure about if this is all correct. If I remove the "!blocked" part of the http_access rule, all of Facebook works. If I remove "facebook.com" from the blocked_content file, all of Facebook works. Right now, visiting facebook.com/OurOrganization gives a "The website declined to show this webpage / HTTP 403" error in Internet Explorer, and "Error 111 (net::ERR_TUNNEL_CONNECTION_FAILED): Unknown error" in Chrome. WhereGoes.com tells me the URL redirects for that URL goes like this: facebook.com/OurOrganization -- [301 Redirect] -- http://www.facebook.com/OurOrganization -- [302 Redirect] -- https://www.facebook.com/OurOrganization I tried turning up the debug traffic out of squid using "debug_options ALL,6" but I can't narrow anything down in /var/log/access.log and /var/log/cache.log. I know to issue "squid -k reconfigure" whenever I make changes to any files.

    Read the article

  • Google API to figure how many times a keyword is searched.

    - by Gal
    Is there a Google API tool to get how many times a specific term was searched between a certain interval (month/day/year)? From my understanding, what I'm looking for is similar to Google Keywords Tool, but I prefer not to fill in their captcha everytime, and I want to know only the statistics that suit the entered term. Does Google/others offer an API to do this?

    Read the article

  • Crystal Reports API - chart: "for all records" or "for each record"?

    - by Epaga
    Is there any way to determine whether a chart in Crystal Reports 2008 (using either the RAS SDK or the older RDC API) is set to display values "for each record" or "for all records"? I can get access to a CrystalDecisions.ReportAppServer.ReportDefModel.ChartObject but can't find any API there to access which type of chart it is - "for each" or "for all".

    Read the article

  • FBML in Rails views using Facebooker

    - by ewakened
    Hi all, I have successfully wired up a Facebook Connect application and everything is working fine. I can sign new users up with Facebook, or I can link existing users with Facebook. No problems there. However, now I am trying to add an Invite page, where a user can see which of their Friends have the application, and then show them a Facebook FMBL Multi Select form to invite friends into it. Here is the documentation for that FBML tag. http://wiki.developers.facebook.com/index.php/Connect/Integrating_an_Invite_Form_into_Your_Website#Creating_a_Full_Multi-friend_Selector_Request_or_Invite However, I cannot get this to work within my Rails view. Do I need to setup something special to use FBML in my Rails views? THIS IS NOT A CANVAS Application, meaning I want people to come to my site and not visit the page through facebook, but I feel like I am missing something. Thank you! Kent

    Read the article

  • How can I get the plain-text content of a Google Doc via Google Doc API in Silverlight?

    - by Edward Tanguay
    I would like to use the Google Doc API to retrieve the plain-text content of a Google Doc document (not a spreadsheet) via the Google Docs API. In the documentation I can find information about accessing Spreadsheets and how to "create/upload/copy" documents: http://code.google.com/apis/documents/overview.html Does anyone have a C# code example to do this in Silverlight, something like this: //PSEUDOCODE: WebClientGoogleDoc proxy = new WebClientGoogleDoc (); proxy.DownloadStringCompleted += new DownloadStringCompletedEventHandler(proxy_DownloadStringCompleted); proxy.DownloadProgressChanged += new DownloadProgressChangedEventHandler(proxy_DownloadProgressChanged); proxy.DownloadStringAsyncWithGoogleDocId("dc7gj86r_20dn2csqg3");

    Read the article

  • Do Websites need Local Databases Anymore?

    - by viatropos
    If there's a better place to ask this, please let me know. Every time I build a new website/blog/shopping-cart/etc., I keep trying to do the following: Extract out common functionality into reusable code (Rubygems and jQuery plugins mostly) If possible, convert that gem into a small service so I never have to deal with a database for the objects involved (by service, I mean something lean and mean, usually built with the Sinatra Web Framework with a few core models). My assumption is, if I can remove dependencies on local databases, that will make it easier and more scalable in the long run (scalable in terms of reusability and manageability, not necessarily database/performance). I'm not sure if that's a good or bad assumption yet. What do you think? I've made this assumption because of the following reason: Most serious database/model functionality has been built on the internet somewhere. Just to name a few: Social Network API: Facebook Messaging API: Twitter Mailing API: Google Event API: Eventbrite Shopping API: Shopify Comment API: Disqus Form API: Wufoo Image API: Picasa Video API: Youtube ... Each of those things are fairly complicated to build from scratch and to make as optimized, simple, and easy to use as those companies have made them. So if I build an app that shows pictures (picasa) on an Event page (eventbrite), and you can see who joined the event (facebook events), and send them emails (google apps api), and have them fill out monthly surveys (wufoo), and watch a video when they're done (youtube), all integrated into a custom, easy to use website, and I can do that without ever creating a local database, is that a good thing? I ask because there's two things missing from the puzzle that keep forcing me to create that local database: Post API RESTful/Pretty Url API While there's plenty of Blogging systems and APIs for them, there is no one place where you can just write content and have it part of some massive thing. For every app, I have to use code for creating pretty/restful urls, and that saves posts. But it seems like that should be a service! Question is, is that what the website is? ...That place to integrate the worlds services for my specific cause... and, sigh, to store posts that only my site has access to. Will everyone always need "their own blog"? Why not just have a profile and write lots of content on an established platform like StackOverflow or Facebook? ... That way I can write apps entirely without a database and know that I'm doing it right. Note: Of course at some point you'd need a database, if you were doing something unique or new. But for the case where you're just rewiring information or creating things like videos, events, and products, is it really necessary anymore??

    Read the article

  • Would a Centralized Blogging Service Work?

    - by viatropos
    If there's a better place to ask this, please let me know. Every time I build a new website/blog/shopping-cart/etc., I keep trying to do the following: Extract out common functionality into reusable code (Rubygems and jQuery plugins mostly) If possible, convert that gem into a small service so I never have to deal with a database for the objects involved (by service, I mean something lean and mean, usually built with the Sinatra Web Framework with a few core models. My assumption is, if I can remove dependencies on local databases, that will make it easier and more scalable in the long run (scalable in terms of reusability and manageability, not necessarily database/performance). I'm not sure if that's a good or bad assumption yet. What do you think? I've made this assumption because of the following reason: Most serious database/model functionality has been built on the internet somewhere. Just to name a few: Social Network API: Facebook Messaging API: Twitter Mailing API: Google Event API: Eventbrite Shopping API: Shopify Comment API: Disqus Form API: Wufoo Image API: Picasa Video API: Youtube ... Each of those things are fairly complicated to build from scratch and to make as optimized, simple, and easy to use as those companies have. So if I build an app that shows pictures (picasa) on an Event page (eventbrite), and you can see who joined the event (facebook events), and send them emails (google apps api), and have them fill out monthly surveys (wufoo), and watch a video when they're done (youtube), all integrated into a custom, easy to use website, and I can do that without ever creating a local database, is that a good thing? I ask because there's two things missing from the puzzle that keep forcing me to create that local database: Post API RESTful/Pretty Url API While there's plenty of Blogging systems and APIs for them, there is no one place where you can just write content and have it part of some massive thing. For every app, I have to use code for creating pretty/restful urls, and that saves posts. But it seems like that should be a service! Question is, is that the main point of a website? Will everyone always need "their own blog"? Why not just have a profile and write lots of content on an established platform like StackOverflow or Facebook?

    Read the article

  • Google Books API - How do I separate Book from Book Edition?

    - by RadiantHex
    Hi, I'm using Google Book Search API to add missings bits and pieces to my database. Problem is that The API gives me back a list of book editions, and not reference to the original book itself. The data I'm trying to get is this: Original title: The Hobbit Original year of publication: 1937 Can anyone help? Just in case anyone asks: I'm developing this in Python.

    Read the article

  • Google search ajax api is to restrictive. Any alternatives?

    - by boomhauer
    The google search ajax api is terrific, and the .net wrapper available on codeplex makes using it from a .net project very simple. However, the api itself is crippled so that it only returns 64 results per query. Not very useful for many applications. Ignoring the likely TOS problems, are there known .net libraries that can query the root google website and scrape the results into a resultset? I'm assuming this could result in much larger result counts than the ajax version enabled.

    Read the article

< Previous Page | 126 127 128 129 130 131 132 133 134 135 136 137  | Next Page >