Beefy Boxes and Bandwidth Generously Provided by pair Networks
No such thing as a small change
 
PerlMonks  

Re: 15 billion row text file and row deletes - Best Practice?

by serf (Chaplain)
on Dec 02, 2006 at 15:34 UTC ( #587415=note: print w/replies, xml ) Need Help??


in reply to 15 billion row text file and row deletes - Best Practice?

Wow! what an eye catching question, good one!

I would be wary of thinking of using grep, as bsdz and sgt have mentioned.

If you have a look at

grep -vf exclude_file to_thin_file in perl

you will see that Perl can do this much faster and with less memory than grep if the script is written efficiently.

My workmate swears by DBD::CSV - but I haven't used it.

Personally I think I'd feel safer writing to a new file in case anything went wrong while it was writing back - if it's running for a week that's a long time to risk it crashing!

  • Comment on Re: 15 billion row text file and row deletes - Best Practice?

Log In?
Username:
Password:

What's my password?
Create A New User
Node Status?
node history
Node Type: note [id://587415]
help
Chatterbox?
[marto]: so if someone just patched something, there's a problem
[marto]: or perhaps just a temporary thing

How do I use this? | Other CB clients
Other Users?
Others meditating upon the Monastery: (11)
As of 2017-12-14 21:27 GMT
Sections?
Information?
Find Nodes?
Leftovers?
    Voting Booth?
    What programming language do you hate the most?




















    Results (410 votes). Check out past polls.

    Notices?