Beefy Boxes and Bandwidth Generously Provided by pair Networks
laziness, impatience, and hubris
 
PerlMonks  

Re: Checking Google PageRank

by biohisham (Priest)
on Oct 11, 2011 at 17:12 UTC ( #930861=note: print w/replies, xml ) Need Help??


in reply to Checking Google PageRank

Can you provide more specifics on the 'is not working for me anymore' part? What complaints do you get? and what do you make of these complaints?

From CPAN the other PageRank moduels are more or less wrappers to WWW::Google::PageRank or are simple algorithms that may perform on a limited size of data



David R. Gergen said "We know that second terms have historically been marred by hubris and by scandal." and I am a two y.o. monk today :D, June,12th, 2011...

Replies are listed 'Best First'.
Re^2: Checking Google PageRank
by vit (Pilgrim) on Oct 11, 2011 at 17:52 UTC
    Method get returns an empty string.
    use strict; use Data::Dumper; use WWW::Google::PageRank; my $pr = WWW::Google::PageRank->new; print Dumper($pr); print scalar($pr->get('http://www.yahoo.com/')), "\n";
      Well, what appears is that an HTTP request gets executed and is actually successful but no content is returned that we can decode via HTTP::Response::decoded_content or HTTP::Response::content, it is very hard for me to find out but the behavior seems universal regardless of the website I am querying, so maybe Google has stopped entertaining automated access via code for security probably, just a guess ..
      use strict; use warnings; use HTTP::Response; use HTTP::Status; use Data::Dumper; use WWW::Google::PageRank; my $pr = WWW::Google::PageRank->new; my @context = $pr->get("http://www.perlmonks.org"); print $context[1]->status_line, "<<<<\n"; #The request is carried out! print $context[1]->content,"\n"; #No result is returned print $context[1]->decoded_content,"\n"; #No result either



      David R. Gergen said "We know that second terms have historically been marred by hubris and by scandal." and I am a two y.o. monk today :D, June,12th, 2011...

        This is most likely. Google's PageRank endpoints are very trigger-happy with the banhammer, and will begin demanding CAPTCHAs (which I'm sure WWW::Google::PageRank will not like) very quickly.

Log In?
Username:
Password:

What's my password?
Create A New User
Node Status?
node history
Node Type: note [id://930861]
help
Chatterbox?
[RonW]: james28909 Why not write a Perl program to do the task?
[choroba]: RonW Yes, but then, one day, they needed to switch from FTP to SFTP, and... but I can't give the whole talk away here :)
[RonW]: Tunnel FTP through stunnel?
[james28909]: i guess i should have done it in perl. i bet it woulnt have taken me 5 hours to figure out haha
[choroba]: In the end, I was able to upgrade Perl from 5.8.3 to 5.22 and install Net::SFTP::Foreign , at least for the task involved

How do I use this? | Other CB clients
Other Users?
Others examining the Monastery: (11)
As of 2017-05-22 21:45 GMT
Sections?
Information?
Find Nodes?
Leftovers?
    Voting Booth?