BrowserUk> I dummied up two days worth of data files:
Could you please share your code to create the dummy data? One, I would see what your data look like, and two, I would learn a few tricks meself.
BrowserUk> With compression, that could be as little as 1.3 TB.
BrowserUk> Though you'd have a pay the price for unpacking--
BrowserUk> ~30 seconds per file.
RIght. Good to keep in mind, but, not interested in paying the 30 seconds price for optimizing for space. Space is not a constraining factor, for now.
BrowserUk> But the main point of partitioning your dataset
BrowserUk> this way is that you reduce the search space to
BrowserUk> 1/8th of 1% as soon as you specify the year/day.
BrowserUk> And there is no searching of indexes involved in
BrowserUk> the rest of the query. Just a simple calculation
BrowserUk> and a direct seek.
Of course, I have to add the cost of looking up the spatial extent, which I have to do via the Pg database first. But, that can give me the area I want to pluck out of my image, and then work with it.
BrowserUk> Anyway, t'is your data and your employers money :)
True. This is research, so trying out different ways is a worthwhile exercise in itself. I don't get paid too much, so it is not a lot of money on my employer's part. ;-). Nevertheless, thanks much. This is fantastic. Always a pleasure to ask and learn different approaches to solving a problem.
--
when small people start casting long shadows, it is time to go to bed