USAGE: %(program)s SIZE_OF_JOBS_QUEUE
Dispatcher process which orchestrates distributed LDA computations. Run this script only once, on any node in your cluster.
Example: python -m gensim.models.lda_dispatcher
Dispatcher object that communicates and coordinates individual workers.
There should never be more than one dispatcher running at any one time.
Note that the constructor does not fully initialize the dispatcher; use the initialize() function to populate it with workers etc.
Terminate all registered workers and then the dispatcher.
Merge states from across all workers and return the result.
Return pyro URIs of all registered workers.
model_params are parameters used to initialize individual workers (gets handed all the way down to worker.initialize()).
A worker has finished its job. Log this event and then asynchronously transfer control back to the worker.
In this way, control flow basically oscillates between dispatcher.jobdone() and worker.requestjob().
Wrap self._jobsdone, needed for remote access through Pyro proxies
Initialize all workers for a new EM iterations.