in reply to
Re^2: how to merge many files of sorted hashes?

in thread how to merge many files of sorted hashes?

so let's see if I'm reading this right:

- you have 100 points
- for every possible triple (1,000,000 combinations) you're computing a matrix
- each matrix gets pushed 100 times (onto 100 different keys)
- you start a new file every 1000 matrices (so there'll be 1000 files when you're done, and each file will have 100,000 matrices.

but you don't say how many distinct keys there are (i.e., how finely

`binning` is grouping things).

But it doesn't matter. For **every single key** you are doing a **linear search** through **every file**. So that's (number of keys) * 100,000,000 matrices to wade through; no wonder this is slow.

You need to be writing each `%small_hash` in sorted order. Sorting 1000 small_hashes will be faster than sorting the big hash
`(1000*(n/1000)*log(n/1000) = n log(n/1000) < n log(n)`
but the really important thing is this will enable the merge I was talking about before, which only reads each file **once** rather than once for every key value.