Beefy Boxes and Bandwidth Generously Provided by pair Networks
Just another Perl shrine
 
PerlMonks  

Re^9: Serializing a large object

by BrowserUk (Pope)
on Sep 28, 2010 at 06:30 UTC ( #862359=note: print w/ replies, xml ) Need Help??


in reply to Re^8: Serializing a large object
in thread Serializing a large object

Does it makes sense to compress the store files?

Yes & no. :(

  • Yes.

    I generated a random set of 3,000 overlaps--positive & negative--with a maximum range of 10,000.

    The nstore'd file on disk was: 26/09/2010  15:26        60,783,878 fred.bin.

    gzipping that resulted in:     26/09/2010  15:26           423,984 fred.bin.gz.

    It'll certainly save you large amounts of disk space. But that's not your aim.

  • No.

    The problem is that whilst you save time reading from disk. You send time decompressing. And in the end, much of the time spent retrieve()ing the data, is the time required to allocate the memory and reconstruct it.

It would certainly be worth you investigating the idea with your real-world datasets; and it will absolutely save huge amounts of disk space. But whether it will actually load faster will depend on many factors; you'll have to try for yourself with real data.


Examine what is said, not who speaks -- Silence betokens consent -- Love the truth but pardon error.
"Science is about questioning the status quo. Questioning authority".
In the absence of evidence, opinion is indistinguishable from prejudice.


Comment on Re^9: Serializing a large object
Select or Download Code

Log In?
Username:
Password:

What's my password?
Create A New User
Node Status?
node history
Node Type: note [id://862359]
help
Chatterbox?
and the web crawler heard nothing...

How do I use this? | Other CB clients
Other Users?
Others perusing the Monastery: (6)
As of 2014-08-22 06:47 GMT
Sections?
Information?
Find Nodes?
Leftovers?
    Voting Booth?

    The best computer themed movie is:











    Results (148 votes), past polls