tomotopy
tomotopy copied to clipboard
Make threadpool queue size configurable
Analysis here suggests that the queue size is too small to be effective for parallel processing. It's a bit hard to follow the code path, but it seems to be capped at 8*num_threads
in the case of LDA. While it isn't obvious what a good default is, this being hard coded seems to be causing issues. It should be a kwarg from infer
so users can tune inference.