Your skill will accomplish what the force of many cannot |
|
PerlMonks |
comment on |
( [id://3333]=superdoc: print w/replies, xml ) | Need Help?? |
Scrappy Comments: Looks interesting, but the docs are a bit scattered and felt confusing, and the development seems to have stagnated. scrappy is too much pee as I say in Re: Scrappy user_agent error ... spidering ... scraping ... You're confusing yourself there a little, spidering (anything goes) is a completely different ballgame than scraping (this one particular site) WWW::Mechanize makes LWP bearable for scraping WWW::Mechanize::Firefox adds JS support that's it for scraping, the bare essentials and the state of the art, all the others add a little sugar and some baggage WWW::Scripter adds "JS support" WWW::Scripter::Plugin::JavaScript/WWW::Scripter::Plugin::Ajax... very impressive/neat, but it ain't no WWW::Mechanize::Firefox Web::Scraper adds a little sugar, like Web::Magic adds maximum magic with maximum dependencies Web::Query adds maximum sugar with minimal dependencies, like jQuery, but has odd bits -Mojo is slick but Mojo::DOM is still barren (pita) XML::LibXML -- has work to do, it can load html/xml documents from urls but the headers/post/... HTTP stuff isn't there yet
Mozilla::Mechanize -- good luck building that :) it ain't easy, no it aint easy Gtk3::WebKit -- its a browser , you might scrape with it somehow (probably not) Gtk2::WebKit -- its a browser , you might scrape with it somehow (probably not) Wx::Htmlwindow -- its a (old/weak/limited) browser , you can scrape with it somehow , its clumsy and limited, not good for scraping Wx::WebView-- its a (new/moderner/css+js) browser , even more useless for scraping than wx::htmlwindow ... looks nice but like all these browsers, not designed for scraping, although it could be In reply to Re: The State of Web spidering in Perl
by Anonymous Monk
|
|