A multiprocess bulk-uploading helper for CouchDB
This is not yet finished, but can already be used.
- Updated readme
- Introduced the jobsbuffersizemax argument to limit the amount of upload-processed buffered in working memory. Now it is possible to work on data of infinite size.
- only spawn one concurrent process for performing uploads
- working version, still spans unlimited amount of processes
Create a new mpcouchPusher object
myCouchPusher = mpcouch.mpcouchPusher( "http://localhost:5984/myDatabase", 30000 )
If the data is created faster than the upload happens, it might be necessary to let the generation of new documents pause befor allowing another batch-upload process to be buffered. The amount of bufferd upload-processes is specified by the optional jobsbuffersizemax argument which defaults to 10. This means that when there are 10 batch-upload processes already waiting in line to be executed, the module holds the main python thread until one process has finished. By adjusing this value it is possible to find an equilibrium between upload-buffer (working memory) and document generation speed.
Use this object every time you have one single document ready to be stored in the database:
The module will collect all documents until the threshold is reached (in our example this would be the 30000 specified above) and upload them as a batch to the CouchDB also specified at creation time of the object (myCouchDbDatabase).
Since every bulk-upload is performed by a single process, the original program continues while the upload happens in the background.
To wait for all running uploads to finish and to make sure the very last batch of documents gets pushed to the server, run
after your final document was sent to pushData. The module now waits for all the uploads to finish and uploads the final bulk of collected documents.