|Pathologically Eclectic Rubbish Lister|
Re^5: dynamic number of threads based on CPU utilizationby BrowserUk (Pope)
|on Sep 27, 2012 at 17:27 UTC||Need Help??|
Okay. This is what I think I would try given your situation:
The comments in this code -- which simulates your XML files using simple flat text files of sha256_hex keys and a number from which the sha256 was derived -- are sparse; and the questions probably many. Easier to answer your actual questions than try and guess what they might be and answer them in comments.
There are two tunable parameters -- the size of the queues; which should be at least double the size of the pools -- and the size of the pools (threads). I've used the same numbers for both halves of the program, but you might want to try using different values and tuning them separately.
Take a look -- ask whatever questions arise :)
It uses my own threads::Q implementation of a self-limiting (size) queue:
That is pretty well exercised -- except the q_nb() which I've never had occasion to use for real -- but is undocumented, hence not on CPAN; new() takes a single argument that is the maximum number of items the Q can hold at any given time. dq() is dequeue(); nq() is enqueue(); cq() (see-queue) is pending().
Its benefit is that it controls the unrestricted growth that can occur when the producer runs faster than the consumer. To its possible debit is that it doesn't accept anything other than scalars or references to pre-shared structures. I consider this a positive as in my experiments it is faster and far less memory hungry to pass compound data joined into a scalar and split it in the consumer, than to copy data in to a shared structure and copy it back again.
With the rise and rise of 'Social' network sites: 'Computers are making people easier to use everyday'
Examine what is said, not who speaks -- Silence betokens consent -- Love the truth but pardon error.
"Science is about questioning the status quo. Questioning authority".
In the absence of evidence, opinion is indistinguishable from prejudice.