Beefy Boxes and Bandwidth Generously Provided by pair Networks
"be consistent"
 
PerlMonks  

Reading a remote Filesystem

by CodeJunkie (Monk)
on May 12, 2003 at 10:52 UTC ( #257387=perlquestion: print w/replies, xml ) Need Help??

CodeJunkie has asked for the wisdom of the Perl Monks concerning the following question:

Hi,
Can anyone tell me how I can read a remote file system and do I listing via my CGI script?

I am trying to write a web based content management system and need to get file listing from the remote machine. We are doing the transfering of files using a program called rsync (i'm sure some of you are aware of this program).

Can anyone tell me if there are any perl modules out there for reading a remote file system, I think maybe rsync has some support for this, so that's my next place to check, but if anyone can offer i'd any advice I would appreciate it.

Cheers,
Tom

Replies are listed 'Best First'.
Re: Reading a remote Filesystem
by Jaap (Curate) on May 12, 2003 at 11:22 UTC
    This can be done using any of the following: (there's probably plenty more possibilities)
    • Mount the filesystem to your webserver
    • FTP or SFTP to the remote box
    • telnet / ssh to the remote box
    • Use rsync to obtain the file listing
    • Create your own protocol with daemon & client ;-)
    I'd start exploring what rsync has to offer.
Re: Reading a remote Filesystem
by hacker (Priest) on May 12, 2003 at 12:56 UTC
    There's always File::RsyncP::FileList, File::Rsync, URI::rsync, and File::DirSync if you want to go that route.

    You might also take a look at rcopy for some other ideas, which purports to mirror two servers together.

    I did something similar, using a remote webserver's directory listing of files, and presenting them to my local machine as "local" files, for users to click/download/etc. Here's an example of that, using kernel.org in this case.

    Basically what this does is fetch the remote "page" of files, extract the links, calculate the file sizes, "commify" them, and prints them out. Simple. Maybe this will give you a few ideas, depending on where/how these files are found.

    use strict; use LWP::UserAgent; use HTML::LinkExtor; use URI::URL qw(url); my $url = "http://kernel.org/pub/linux/kernel/v1.0/"; my $ua = LWP::UserAgent->new; $ua->agent('pps 0.1.43'); my @links = (); sub callback { my ($tag, %attr) = @_; return if $tag eq 'href'; push(@links, values %attr); return (@links); } my $p = HTML::LinkExtor->new(\&callback); my $res = $ua->request( HTTP::Request->new(GET => $url), sub {$p->parse($_[0])}); my $base = $res->base; @links = map {$_ = url($_, $base)->abs;} @links; my @krn = grep(/tar/, @links); foreach my $krn (@krn) { my @remote_files = HTTP::Request->new(HEAD=>$krn); my $resp = $ua->request(@remote_files); my $length = $resp->header('Content-Length'); my $bprecise = sprintf "%.0f", $length; my $bsize = insert_commas($bprecise); my $kprecise = sprintf "%.0f", ($length/1024); my $ksize = insert_commas($kprecise); my $archtype; # tarball? or bzip2? or zip? my $krnhref = substr($krn, 48, 70); if ($krn =~ /tar.gz/) { $archtype = "tarball"; } elsif ($krn =~ /bz2/) { $archtype = "bzip"; } else { $archtype = "zip"; } print "Path: $krn\nFile: $krnhref "; print "($archtype) - ${ksize}kb, $bsize bytes\n\n"; } sub insert_commas { my $text = reverse $_[0]; $text =~ s/(\d{3})(?=\d)(?!\d*\.)/$1,/g; return scalar reverse $text; }
Re: Reading a remote Filesystem
by Abigail-II (Bishop) on May 12, 2003 at 11:55 UTC
    You don't give much information. What the best solution is depends a lot on what is available on the remote site, or how much control you have of said remote site. Do you admin the remote site? Can you install software there? Do you have an account there?

    Personally, I'd go for NFS or samba, if that's in the realm of the possibilities.

    Abigail

Re: Reading a remote Filesystem
by CodeJunkie (Monk) on May 12, 2003 at 12:05 UTC

    yeah I have access to both machines. The reason we need to develop this system is because we are not allowing any kind of ftp access to the live server, i'm not sure about sftp. We just connect using ssh and rsync. After giving it some thought I think I am going to get the file listing from machine a, then check the file exists on machine b by sending a http request for it via LWP. This should be ok because we will only need to list top level folders, so only make a few requests to check the folders exist.

    This functionality is a bit limited, but it's enough to fulfill our requirements ;-)

    Cheers,
    Tom

      Dunno all the whys and wherefores of not allowing remote FTP access to the system, but if you put a web server on the remote system and write a secure CGI, you can easily provide listings and downloads, perhaps by only allowing HTTP access from the machine that's currently serving up CGI (i.e. the one through which the users will access the site); even that's not necessary if you just give the web server access to the directories containing the files and let it list the files in the directories (this is very easy to do with Apache, for example).

      Keeping it short and sweet, why not let HTTP be your remote file access protocol =)

      If not P, what? Q maybe?
      "Sidney Morgenbesser"

        Reasons not to use FTP:

        1. Sends passwords in plaintext
        2. Multiple-port design is the cause of many headaches for firewall administrators
        3. Generally considered insecure (though more for historical problems with server software than today's software)

        Even if we accept that #3 is due to old and buggy software, #1 is reason enough to dump it. #2 isn't a big deal, since admins running firewalls tend to notice the problem the first week on the job and always keep it in mind from then on. However, it also is a problem for the firewall software itself (the early Linux 2.4 series packet filter had a bug where if you let FTP in, an attacker could get through any port).

        Its not like there aren't better alterantives. In particular, the sftp subsystem that comes with OpenSSH is great. It's a one-line change to your sshd.config and works over the existing SSH port.

        IMHO, HTTP is a poor subsitute for FTP, except as a quick-and-dirty document retrevial system. Uploading is hacked in, and its stateless nature has been the cause of hackery elsewhere.

        ----
        I wanted to explore how Perl's closures can be manipulated, and ended up creating an object system by accident.
        -- Schemer

        Note: All code is untested, unless otherwise stated

      If you have ssh already installed, then you can get sftp with a one-line change to sshd.config. Add this:

      Subsystem sftp sftp-server

      And then restart sshd.

      ----
      I wanted to explore how Perl's closures can be manipulated, and ended up creating an object system by accident.
      -- Schemer

      Note: All code is untested, unless otherwise stated

Log In?
Username:
Password:

What's my password?
Create A New User
Node Status?
node history
Node Type: perlquestion [id://257387]
Approved by broquaint
help
Chatterbox?
and the web crawler heard nothing...

How do I use this? | Other CB clients
Other Users?
Others imbibing at the Monastery: (11)
As of 2020-05-26 13:38 GMT
Sections?
Information?
Find Nodes?
Leftovers?
    Voting Booth?
    If programming languages were movie genres, Perl would be:















    Results (150 votes). Check out past polls.

    Notices?