Search Results

Search found 2482 results on 100 pages for 'syndication feed'.

Page 26/100 | < Previous Page | 22 23 24 25 26 27 28 29 30 31 32 33  | Next Page >

  • Facebook IOS SDK: Error in Publish Story Dialog

    - by lividsquirrel
    I've successfully set up the "DemoApp" project from the Facebook IOS SDK to use my "OKC ThunderCast" Facebook application. I have also configured another "Tester" application from scratch to successfully use the Facebook SDK and publish stories to my news feed. However, in my production application, I always get this result when calling the "dialog" method. The full description of the error message is "Error on line 52 at column 17: Opening and ending tag mismatch: div line 0 and body" Here's a detailed walkthrough of all of my code to make sure nothing is missed. 1) A UIViewController calls the "authorize" method NSArray *fbPerms = [NSArray arrayWithObjects:@"read_stream", @"offline_access", nil]; [[FacebookSingleton sharedInstance].facebook authorize:fbPerms delegate:self]; Note: The FacebookSingleton is a class I wrote that always returns a single instance of the "Facebook" class. I am using it successfully in other applications. 2) Safari is opened and the user is successfully authenticated and authorized 3) The application is called back and the "handleOpenUrl" method is called, which calls the "fbDidLogin" method of the UIViewController - (BOOL)application:(UIApplication *)application handleOpenURL:(NSURL *)url { Facebook *fb = [FacebookSingleton sharedInstance].facebook; return [fb handleOpenURL:url]; } 4) The same UIViewController handles the "fbDidLogin" event, and calls the "dialog" method - (void)fbDidLogin { [[FacebookSingleton sharedInstance].facebook dialog:@"feed" andDelegate:self]; } I also have the necessary "URL Schemes" and "URL Types" entries in the .plist file. To my eyes, I am using exactly the same code in the "DemoApp", "Tester", and production applications. But while the DemoApp and Tester work, I always see this HTML error in the feed dialog in my production application. Has anyone seen a similar issue? Could it be related to the Facebook "Bundle ID" setting in the Facebook application settings? Is there some build or .plist setting that is different? I have invested a great deal of time into troubleshooting with no success in several weeks. Thanks in advance...

    Read the article

  • How to parse rss from a php page, using jQuery/jFeed?

    - by ricebowl
    I'm trying to fumble my way through parsing rss sensibly, using jQuery and jFeed. Because of the same origin policy I'm pulling the BBC's health news feed into a local page (http://www.davidrhysthomas.co.uk/play/proxy.php). Originally this was just the same proxy.php script as available in the jFeed download package, but due to my host's disabling allow_url_fopen() I've amended the php to the following: $url = "http://newsrss.bbc.co.uk/rss/newsonline_uk_edition/health/rss.xml"; $ch = curl_init(); curl_setopt($ch, CURLOPT_URL, $url); curl_setopt($ch, CURLOPT_HEADER, 0); curl_setopt($ch, CURLOPT_RETURNTRANSFER, 1); $data = curl_exec($ch); echo "$data"; curl_close($ch); Which seems to generate the same/comparable contents as the original fopen on my local machine. Now that seems to be working, I'm looking at setting the jFeed script up to work with the page and, to my embarrassment, don't see how. I understand that, at the least, this should work: jQuery.getFeed({ url: 'http://www.davidrhysthomas.co.uk/play/proxy.php', success: function(feed) { alert(feed.title); } }); ...but, as I'm sure you anticipate, it doesn't. What non-output there is, is available for your perusal here: http://www.davidrhysthomas.co.uk/play/exampleTest.html. And I honestly don't have a clue what to do about it. If anyone could offer some pointers, tips, hints, or, at a pinch, a quick slap around the cheeks and a 'pull yourself together!' it'd be much appreciated... Thanks in advance =)

    Read the article

  • Having difficulties in ending Michael Hartl's tutorial. Help?

    - by konzepz
    Following Michael Hartl's (amazing) Ruby on Rails Tutorial, on the final section, I get the following errors: 1) User micropost associations status feed should include the microposts of followed users Failure/Error: @user.feed.should include(mp3) expected [#<Micropost id: 2, content: "Foo bar", user_id: 1, created_at: "2011-01-12 21:22:41", updated_at: "2011-01-12 22:22:41">, #<Micropost id: 1, content: "Foo bar", user_id: 1, created_at: "2011-01-11 22:22:41", updated_at: "2011-01-12 22:22:41">] to include #<Micropost id: 3, content: "Foo bar", user_id: 2, created_at: "2011-01-12 22:22:41", updated_at: "2011-01-12 22:22:41"> Diff: @@ -1,2 +1,2 @@ -#<Micropost id: 3, content: "Foo bar", user_id: 2, created_at: "2011-01-12 22:22:41", updated_at: "2011-01-12 22:22:41"> +[#<Micropost id: 2, content: "Foo bar", user_id: 1, created_at: "2011-01-12 21:22:41", updated_at: "2011-01-12 22:22:41">, #<Micropost id: 1, content: "Foo bar", user_id: 1, created_at: "2011-01-11 22:22:41", updated_at: "2011-01-12 22:22:41">] # ./spec/models/user_spec.rb:214 2) Micropost from_users_followed_by should include the followed user's microposts Failure/Error: Micropost.from_users_followed_by(@user).should include(@other_post) expected [#<Micropost id: 1, content: "foo", user_id: 1, created_at: "2011-01-12 22:22:46", updated_at: "2011-01-12 22:22:46">] to include #<Micropost id: 2, content: "bar", user_id: 2, created_at: "2011-01-12 22:22:46", updated_at: "2011-01-12 22:22:46"> Diff: @@ -1,2 +1,2 @@ -#<Micropost id: 2, content: "bar", user_id: 2, created_at: "2011-01-12 22:22:46", updated_at: "2011-01-12 22:22:46"> +[#<Micropost id: 1, content: "foo", user_id: 1, created_at: "2011-01-12 22:22:46", updated_at: "2011-01-12 22:22:46">] # ./spec/models/micropost_spec.rb:75 Finished in 9.18 seconds 153 examples, 2 failures Seems like mp3 is not included in the feed. Any ideas on how to fix it? Or where to look for possible errors in the code? I compared the files with Hartl's original code; seems exact. Thanks.

    Read the article

  • Getting Feedburner Subscriber With CURL

    - by Eray Alakese
    I'm using FeedBurner Awareness API. XML data like this : <rsp stat="ok"> - <!-- This information is part of the FeedBurner Awareness API. If you want to hide this information, you may do so via your FeedBurner Account. --> - <feed id="9n66llmt1frfir51p0oa367ru4" uri="teknoblogo"> <entry date="2011-01-15" circulation="11" hits="18" reach="0"/> </feed> </rsp> I want to get circulation data (11) . I'm using this code: $whaturl="https://feedburner.google.com/api/awareness/1.0/GetFeedData?uri=teknoblogo"; //Initialize the Curl session $ch = curl_init(); //Set curl to return the data instead of printing it to the browser. curl_setopt($ch, CURLOPT_RETURNTRANSFER, 1); //Set the URL curl_setopt($ch, CURLOPT_URL, $whaturl); //Execute the fetch $data = curl_exec($ch); //Close the connection curl_close($ch); $xml = new SimpleXMLElement($data); $fb = $xml->feed->entry['circulation']; echo $fb; echo "OK"; But, returned data is blank. There isn't any error. Only return OK . How can i solve this ? EDIT : echo $data; returning blank, too.

    Read the article

  • Advance: Parsing XML into another XML page using only javascript or jquery; Can't use PhP, Java or MySQL

    - by UrBestFriend
    Current site: http://cardwall.tk/ Example of intended outcome: http://www.shockwave.com/downloadWall.jsp I have an embeded flash object that uses XML/RSS (Picasa) to feed itself pictures. Now I created my own XML/RSS feed so that I can add additional XML tags and values. Now here's my big problem: enabling search. Since I'm not relying on Picasa's API anymore to return custom RSS/XML for the user's search, how can I create xml from another xml based on the user's search queries using only JavaScript and Jquery? Here is the current code: <script type="text/javascript"> var flashvars = { feed : "http%3A%2F%2Frssfeed.ucoz.com%2Frssfeed.xml", backgroundColor : "#FFFFFF", metadataFont : "Arial", wmode : "opaque", iFrameScrolling: "no", numRows : "3", }; var params = { allowFullScreen: "true", allowscriptaccess : "always", wmode: "opaque" }; swfobject.embedSWF("http://apps.cooliris.com/embed/cooliris.swf", "gamewall", "810", "410", "9.0.0", "", flashvars, params); $(document).ready(function() { $("#cooliris input").keydown(function(e) { if(e.keyCode == 13) { $("#cooliris a#searchCooliris").click(); return false; } }); doCoolIrisSearch = function() { cooliris.embed.setFeedContents( '** JAVA STRING OF PARSED RSS/XML based on http%3A%2F%2Frssfeed.ucoz.com%2Frssfeed.xml and USER'S SEARCH INPUT** ' ) }); <form id="searchForm" name="searchForm" class="shockwave"> <input type="text" name="coolIrisSearch" id="coolIrisSearch" value="Search..." class="field text short" onfocus="this.value='';" /> <a id="searchCooliris" href="#" onclick="doCoolIrisSearch();return false;" class="clearLink">Search Cooliris</a> </form> <div id="gamewall"></div> So basically, I want to replace cooliris.embed.setFeedContents's value with a Javastring based on the parsed RSS/XML and user search input. Any code or ideas would be greatly appreciated.

    Read the article

  • Gdata JavaScript Authsub continues redirect

    - by Krustal
    I am using the JavaScript Google Data API and having issues getting the AuthSub script to work correctly. This is my script currently: google.load('gdata', '1'); function getCookie(c_name){ if(document.cookie.length>0){ c_start=document.cookie.indexOf(c_name + "="); if(c_start!=-1){ c_start=c_start + c_name.length+1; c_end=document.cookie.indexOf(";",c_start); if(c_end==-1) c_end=document.cookie.length; return unescape(document.cookie.substring(c_start, c_end)); } } return ""; } function main(){ var scope = 'http://www.google.com/calendar/feeds/'; if(!google.accounts.user.checkLogin(scope)){ google.accounts.user.login(); } else { /* * Retrieve all calendars */ // Create the calendar service object var calendarService = new google.gdata.calendar.CalendarService('GoogleInc-jsguide-1.0'); // The default "allcalendars" feed is used to retrieve a list of all // calendars (primary, secondary and subscribed) of the logged-in user var feedUri = 'http://www.google.com/calendar/feeds/default/allcalendars/full'; // The callback method that will be called when getAllCalendarsFeed() returns feed data var callback = function(result) { // Obtain the array of CalendarEntry var entries = result.feed.entry; //for (var i = 0; i < entries.length; i++) { var calendarEntry = entries[0]; var calendarTitle = calendarEntry.getTitle().getText(); alert('Calendar title = ' + calendarTitle); //} } // Error handler to be invoked when getAllCalendarsFeed() produces an error var handleError = function(error) { alert(error); } // Submit the request using the calendar service object calendarService.getAllCalendarsFeed(feedUri, callback, handleError); } } google.setOnLoadCallback(main); However when I run this the page redirects me to the authentication page. After I authenticate it send me back to my page and then quickly sends me back to the authenticate page again. I've included alerts to check if the token is being set and it doesn't seem to be working. Has anyone has this problem?

    Read the article

  • Why can’t PHP script write a file on server 2008 via command line or task scheduler?

    - by rg89
    I created a question on serverfault.com, and it was recommended that I ask here. http://serverfault.com/questions/140669/why-cant-php-script-write-a-file-on-server-2008-via-command-line-or-task-schedul I have a PHP script. It runs well when I use a browser. It writes an XML file in the same directory. The script takes ~60 seconds to run, and the resulting XML file is ~16 MB. I am running PHP 5.2.13 via FastCGI on Windows Server Web edition SP1 64 bit. The code pulls inventory from SQL server, runs a loop to build an XML file for a third party. I created a task in task scheduler to run c:\php5\php.exe "D:\inetpub\tools\build.php" The task scheduler shows a time lapse of about a minute, which is how long the script takes to run in a browser. No error returned, but no file created. Each time I make a change to the scheduled task properties, a user password box comes up and I enter the administrator account password. If I run this same path and argument at a command line it does not error and does not create the file. When I right click run command prompt as an administrator, the file is still not created. I get my echo statement "file published" that is after the file creation and no error is returned. I am doing a simple fopen fwrite fclose to save the contents of a php variable to a .xml file, and the file only gets created when the script is run through the browser. Here's what happens after the xml-building loop: $feedContent .= "</feed"; sqlsrv_close( $conn ); echo "<p>feed built</p>"; $feedFile = "feed.xml"; $handler = fopen($feedFile, 'w'); fwrite( $handler, $feedContent ); fclose( $handler ); echo "<p>file published</p>"; Thanks

    Read the article

  • php facebook delete feeds

    - by meAtStackOverflow
    Hi there, I have manage to grab all the photos posted to my app wall. Now I am trying to add a function to my facebook app where it will delete all the feeds posted on the wall. <?php include_once 'facebook.php'; include_once 'config.php'; $feed = $facebook->api('/me/feed'); foreach($feed['data'] as $feeds) { echo "Feeds secion", "<br />"; echo $feeds["picture"], "<br />"; echo "<img src='{$feeds['picture']}' />", "<br />"; $target_path = "uploads/"; $target_path = $target_path . basename($feeds["picture"]); $content = file_get_contents($feeds["picture"]); file_put_contents($target_path, $content); } ? This is how i grab the photos and push them into my server. Please guide me on how i could delete all the feeds. I have went through some online examples but i could not figure out how to go about doing it. I kept receving errors. Thanks!

    Read the article

  • What is the best way to properly test object equality against an array of objects?

    - by radesix
    My objective is to abort the NSXMLParser when I parse an item that already exists in cache. The basic flow of the program works like this: 1) Program starts and downloads an XML feed. Each item in the feed is represented by a custom object (FeedItem). Each FeedItem gets added to an array. 2) When the parsing is complete the contents of the array (all FeedItem objects) are archived to the disk. The next time the program is executed or the feed is refreshed by the user I begin parsing again; however, since a cache (array) now exists as each item is parsed I want to see if the object exists in the cache. If it does then I know I have downloaded all the new items and no longer need to continue parsing. What I am learning, I think, is that I can't use indexOfObject or indexOfObjectIDenticalTo: because these really seem to be checking to see that the objects are using the same memory address (thus identical). What I want to do is see if the contents of the object are equal (or at least some of the contents). I've done some research and found that I can override the IsEqual method; however, I really don't want to iterate/enumerate through the entire cache contents table for every newly parsed XML FeedItem. Is iterating through the collection and testing each one for equality the only way to do this or is there a better technique I am not aware of? Currently I am using the following code though I know it needs to change: NSUInteger index = [self.feedListCache.feedList indexOfObject:self.currentFeedItem]; if (index == NSNotFound) { }

    Read the article

  • Oracle WebCenter Quiz

    - by Michael Snow
    Quiz: How many of the following business necessities can you accomplish with Oracle WebCenter? a) Employee On-boarding b) Policies & Procedures c) Regulatory Compliance d) Sales Enablement Dashboards e) Secure Deal Collaboration f) Document & IP Management g) Accounts Payable h) Records Management i) Claims Processing j) Marketing and Brand Management k) Call Center & HelpDesk l) Contract Management m) Collaborative Content Contribution and Sharing Environment n) Enterprise Application, Desktop and Office integration o) Share Content Across Intranet And Extranets p) Combine Content In Composite Applications q) Subject Matter Expert Location r) Personalize Recommendations of Spaces, Documents, Wikis, Blogs, and Topics s) Collaborative Community Websites t) Marketing Driven Websites u) Strategic Web Experience Management v) Online Engagement Optimization w) Create Targeted Online Experiences x) Manage Interactive Social Experiences y) Optimize Multi-Channel Customer Experiences z) End-User Personalization & Syndication aa) ALL OF THE ABOVE!!!  (HINT: CHOOSE THIS ONE!!) bb) NONE OF THE ABOVE Learn More - Join us for a Webcast   Do More with Oracle WebCenter – Expand Beyond Content Management

    Read the article

  • The Best Free RSS Readers for Keeping Up With Your Favorite Websites

    - by Lori Kaufman
    RSS stands for “Really Simple Syndication” or “Rich Site Summary.” It’s a document specification that allows you to easily collect and organize web-based news and information from websites. RSS reader applications allow you to follow your favorite websites without having to visit each one separately. For more information about RSS, see our article defining RSS and describing how you can benefit from using it. We’ve collected some links to good desktop and web-based, free RSS applications. 6 Ways Windows 8 Is More Secure Than Windows 7 HTG Explains: Why It’s Good That Your Computer’s RAM Is Full 10 Awesome Improvements For Desktop Users in Windows 8

    Read the article

  • Adding my podcast to my Facebook fan page

    - by Donald Burr
    I've set up a Facebook fan page for my podcast, Otaku no Podcast. I'd love to add a Flash based player to the fan page that can play the latest episode of my podcast. Or, at the very least, a link to the latest episode on my website (which has its own Flash-based audio player). My podcast's website of course exports a valid RSS feed. I've tried several different podcast player/RSS feed display applications including Podcast Pickle (which has a facebook app), but none of them appear to work and/or are maintained any more. Podcast Pickle used to work for me a long time ago, but is no longer working for me. Any ideas?

    Read the article

  • iPhone: "there was a problem downloading this podcast"

    - by brianegge
    I have several podcasts which work perfectly fine in iTunes, but when I attempt to download additional episodes on the iPhone I get the message "there was a problem downloading this podcast" once the download completes. All I can find points to my iPhone not having enough disk space. The settings tells me I have 1GB available. I've also tried deleting some content and receive the same message. Originally, I thought this might be due to the RSS feed indicating a different file size than the media file, but I manually verified its correct. A podcast which I can always reproduce this problem has the following feed: http://www.podquiz.com/rss.php I can reproduce the problem on both WiFi and 3G.

    Read the article

  • Tips on Managing Podcast Subscriptions

    - by Ben Griswold
    I listen to a silly number of technical podcasts. I listen to enough of them that it is literally impossible to keep up. I nearly gave up and started dropping feeds from my subscription list when I heard Craig Shoemaker talk about his Polymorphic Podcast fast feed. The idea is he provides the same content at a higher speed so you can listen to his complete show in 3/4th the time. I tried it out with his recent jQuery Secrets with Dave Ward interview and I was shocked with the feed quality. It was a super clear, understandable conversation which only took a fraction of the time commitment. I experimented a bit and played the normal recording at 2x speed on my iPhone and the quality was once again just fine. But now I'm saving half of the time. I'm curious as to how you might manage your podcast subscriptions. Can you offer any tips or advice on how to get the best bang for your buck when it comes to technical podcast listening?

    Read the article

  • Mac/iPhone:Streaming video file to iPhone

    - by user34157
    Hi, I am redirecting my question to superuser from stackoverflow: I have a http streaming link which gives me .flv streaming feed. I want to convert that and access in my iPhone program. How can i do that? I want to have a desktop software like VLC and input this streaming feed URL and convert to iPhone supported and stream again to iPhone. I tried VLC with H.264 and Mpeg-1 audio, but seems to be it doesn't give the supported format, so as iPhone program doesn't play the video. Could someone please guide me how can i setup a desktop software which can stream iPhone supported file? Thanks in advance.

    Read the article

  • Streaming video file to iPhone

    - by user34157
    I have a http streaming link which gives me .flv streaming feed. I want to convert that and access in my iPhone program. How can i do that? I want to have a desktop software like VLC and input this streaming feed URL and convert to iPhone supported and stream again to iPhone. I tried VLC with H.264 and Mpeg-1 audio, but seems to be it doesn't give the supported format, so as iPhone program doesn't play the video. Could someone please guide me how can i setup a desktop software which can stream iPhone supported file?

    Read the article

  • Finding Missing UDP Frames Using Wireshark + Custom Dissector (for CQS)

    - by John Dibling
    How do you use Wireshark to identify missing UDP frames? I have written a custom dissector for the CQS feed (reference page). One of our servers gaps when receiving this feed. According to Wireshark, some UDP frames are never received. I know that the frames were sent because all of our other servers are gap-free. A CQS frame consists of multiple messages, each having its own sequence number. My custom dissector provides the following data to Wireshark: cqs.frame_gaps - the number of gaps within a UDP frame (always zero) cqs.frame_first_seq - the first sequence number in a UDP frame cqs.frame_expected_seq - the first sequence number expected in the next UDP frame cqs.frame_msg_count - the number of messages in this UDP frame And I am displaying each of these values in custom columns, as shown in this screenshot: A typical CQS log will consist of millions of rows, so I can't just eyeball it. Is there any way I can get Wireshark to tell me which frames are missing?

    Read the article

  • Setting up podcasting for a non-tech user

    - by Force Flow
    I have a user who wants to start making podcasts, but they only have basic skills when it comes to technology. So, I was trying to get a process together that would be easy for them to follow. To upload files (the mp3's and rss feed files), I have an explorer shortcut for their FTP space. To record the podcast, I was going to either use audacity or PodProducer. For the RSS feed, I was looking for a podcast RSS generator of some sort. In my search for this, I've come across a lot of dead links and a lot of paid tools, so I haven't come up with anything too useful. Is there a free, reliable webservice or windows-based tool available that folks like to use?

    Read the article

  • Re-downloading accidentally deleted podcasts in iTunes

    - by Matt
    I deleted some previously deleted podcast episodes in iTunes 9 and I'd like to re-download them. (I don't want to manually re-download them because then they won't show up in the podcast's list in the iTunes Podcast section.) I've read http://blog.krisgielen.be/archives/54 as well as Way to get old podcasts in iTunes where the suggestion is to delete the iTunes entries for the podcasts I'd like to recover, then hold shift while re-opening that podcast's little triangle icon. Here's the problem though: holding shift and re-opening the triangle tells iTunes to refresh the RSS feed, and it's possible that the RSS feed no longer contains the item you want. What I'm really looking for is a means of getting the "Get" button (that normally appears on un-downloaded podcasts) to reappear on the episodes I've previously downloaded but deleted.

    Read the article

  • Feeding the kernels entropy source from other machines and/or increasing its maximum size

    - by David Spillett
    We have has a little trouble with a small box that acts as a VPN end-point and mail relay for our network, caused by the available entropy for /dev/random being too low (which causes TLS connection attempts by exim to fail). The machine doesn't do anything else, so the normal feed into the entropy pool (interrupt timings from things like disk access) is not enough. As a quick hack I've set a looping script that reads from /dev/hda at a couple of Mbyte/sec which keeps it topped up. Other than buying a hardware RNG, is there a clean way of piping data for entry from elsewhere, such as a copy of the data our file server uses for its entropy source? I've spotted several tips for using rng-tools to feed it from /dev/urandom on the same machine but that "feels dirty". Also, is it possible to increase the maximum pool size? It currently seems to max out at 3585.

    Read the article

  • Pointcast Alternatives?

    - by kellyllek
    I used to love Pointcast in the early days of the internet; it was a screensaver that, like RSS feeds, brought in all your favorite sources with news articles and stock quotes along with various other info and turned it into a visually interesting screensaver. It was before broadband though and I heard so many work places had to ban it that the company eventually folded. But I wondered if there are any modern day equivalents? I certainly have feed readers and some awesome desktop gadgets that bring in various things like news and weather, and I don't want a whole screensaver like pointcast. But I wondered if anyone out there had other suggestions worth taking a look at; making some kind of feed display of interest, or some other way to turn favorite content into more animated desktop background, etc...

    Read the article

  • .htaccess redirect, from old dirty URL to a clean new URL with parameters

    - by JustAnil
    I have the following 2 links, I'm not great with .htaccess rules yet. Old URL: http://www.mywebsite.org.uk/donate/donate.php?charity_id=885&project_id=18111 New URL: http://new.mywebsite.org.uk/donation/to/885/18111 I want all the traffic coming from the old URL to the new url (including the parameters charity_id & project_id). I'm trying to learn .htaccess rules, but finding the tutorials online to be kinda vague. I'd really like a simple explanation on the .htaccess rules. (Give a man a fish, feed him for a day, teach a man to fish, feed him for a lifetime). The correct answer will be the answer with a simple and useful explanation (along with the rules if possible!).

    Read the article

  • Internet Explorer cannot download rss.php

    - by davethegr8
    Whenever I try to go to an RSS feed in IE7 directly, I get a very strange error. IE 7 tries to download the page, but fails. It throws an error, Internet Explorer cannot download rss.php from www.example.com. Internet Explorer was not able to open this Internet site. The requested site is either unavailable or cannot be found. Please try again later. But when I click a link to view the same feed, it doesn't throw an error but displays the XML instead. Are there any ways I caan fix this, that don't involve OS hacks? Also, if I try to put an xsl stylesheet on it, it tries to download the file no matter what and still errors.

    Read the article

  • iPhone and Mac Twitter App Not Parsing HTML in Feeds

    - by otakrosak
    I've come across this problem where my tweets on the iPhone and Mac Twitter app are not parsing the HTML properly. In any browser, the HTML is parsed fine. It only happens to the apostrophe character (') and on the iPhone and Mac Twitter app. Also, I'm using the dlvr.it service to push my Drupal blog posts to my Twitter feed. My initial guess was that it was the RSS generated by Drupal, but if that were the case, then the feed displayed on the browser would be affected too. Any ideas anyone? Any help would be very much appreciated. P.S: I apologize beforehand if this question already exists. I did search for it but based on what I queried, nothing came up.

    Read the article

  • I'm trying to run some PHP scripts as CLI instead of over HTTP. How do I make them play nice?

    - by gnfti
    Hi everyone. I'm using some PHP scripts from FeedForAll to join together RSS feeds (RSSmesh) and display them as HTML (RSS2HTML). Because I intend to run these scripts fairly intensively and don't want the resulting HTTP requests and bandwidth to count towards my hosting quota, I am in the process of moving to running them on the web host's server in an umbrella PHP "batch" script, and call this script via cron (this is a Linux server, by the way). Here's a (working) sample request over HTTP: http://www.mydomain.com/a/rss2htmlcore/rss2html2.php?XMLFILE=http://www.mydomain.com/a/myapp/xmlcache/feed.xml&TEMPLATE=template.html This will produce the desired HTML output. An example of how I want this to work on the command line: /srv/customers/mycustomer#/mydomain.com/www/a/rss2htmlcore/rss2html2-cli.php /srv/customers/mycustomer#/mydomain.com/www/a/myapp/xmlcache/feed.xml /srv/customers/mycustomer#/mydomain.com/www/a/template.html This is with the correct shebang line added to "rss2html2-cli.php". I could just as well specify the executable ("/usr/local/bin/php") in the request, I doubt it makes a difference because I am able to run another script (that I wrote myself) either way without problems. Now, RSS2HTML and RSSmesh are different in that, for starters, they include secondary files -- for example, both include an XML parser script -- and I suspect that this is where I am getting a bit in over my head. Right now I'm calling exec() from the "umbrella" batch script, like so: exec("/srv/customers/mycustomer#/mydomain.com/www/a/rss2htmlcore/rss2html2-cli.php /srv/customers/mycustomer#/mydomain.com/www/a/myapp/xmlcache/feed.xml /srv/customers/mycustomer#/mydomain.com/www/a/template.html", $output) But no output is being produced. What's the best way to go about this and what "gotchas" should I keep in mind? Is exec() the right way to approach this? It works fine for the other (simple) script but that writes its own output. For this I want to get the output and write it to a file from within the umbrella script if possible. I've also tried output buffering but to no avail. Do I need to pay attention to anything specific with regard to the includes? Right now they're specified in the scripts as include_once("FeedForAll_XMLParser.inc.php"); and the specified files are indeed in the same folder. Further info: -This is a Linux server. -I have no direct access to the shell, so I can't test things directly on a command line, everything is via crontab. -I will admit that support for the FeedForAll scripts leaves a lot to be desired, but I'd like to keep using their scripts if at all possible, if only because I know them and have been using them for a while. I have looked into Simplepie, but the FFA scripts do some things that I've seen no obvious solutions for with Simplepie, like limiting the number of items per individual feed (RSSmesh) or limiting the description length (RSS2HTML). -Yahoo! Pipes is out, they cache their data for too long for my application. Should you want to take a look at the code, here are the scripts as txt files. RSS2HTML2 and RSSmesh are the FeedForAll scripts, FeedForAll_XMLParser... is the included parser. Note that I have not yet amended these to handle $argv etc. I have however in "scraper-universal-rss-cli", which works fine with CLI. If anyone has any thoughts to share on this it would be very much appreciated. Thank you in advance.

    Read the article

< Previous Page | 22 23 24 25 26 27 28 29 30 31 32 33  | Next Page >