Programatticaly grabing text from a web page that is dynamically generated.

Posted by bstullkid on Stack Overflow See other posts from Stack Overflow or by bstullkid
Published on 2010-04-16T17:38:32Z Indexed on 2010/04/16 17:43 UTC
Read the original article Hit count: 220

Filed under:
|
|
|
|

There is a website I am trying to pull information from in perl, however the section of the page I need is being generated using javascript so all you see in the source is

<div id="results"></div>

I need to somehow pull out the contents of that div and save it to a file using perl/proxies/whatever. e.g. the information I want to save would be

document.getElementById('results').innerHTML;

I am not sure if this is possible or if anyone had any ideas or a way to do this. I was using a lynx source dump for other pages but since I cant straight forward screen scrape this page I came here to ask about it!

© Stack Overflow or respective owner

Related posts about perl

Related posts about html