There's more than one way to do things | |
PerlMonks |
Re: The Eternal "filter.pl"by moritz (Cardinal) |
on Aug 25, 2011 at 16:40 UTC ( [id://922406]=note: print w/replies, xml ) | Need Help?? |
I've done similar stuff many times over, though from your description it seems that you've done it much more often than me :-). I can certainly relate to that feeling that the repetition is bothersome, but often not quite enough to attack the problem properly. It seems to me that the only truly common code is "parse this datasource into a stream of records, where 'record' is a list of consistently sequenced fields corresponding to a table definition. To you that's not much, but for others that's enough to start a new hype around "map/reduce". The parsing step is basically a "map", and the filtering and aggregation is a "reduce". As for your actual problem: Or if you have a database to put it in. ... what if all you have is a pair of files about 3gig each Can't you get a developer machine with a few hundred gig of free disc space, and set up your own private database into which you can import such files? I mean, come on, 2x 3gig ain't that much. The import will take some time, but you said yourself that time isn't the problem. Or maybe you want something like an SQL engine that works on in-memory objects? If yes, DBI::DBD::SqlEngine looks promising, though I've never used it before.
In Section
Meditations
|
|