Little process-based workers to do your bidding.
Deliberately minimalist, you provide the number of workers
to use & a list of commands (to be executed at the shell) & littleworkers
will eat through the list as fast as it can.
littleworkers
shines when you just want to parallelize something without a lot of fuss & when you care more about the data/commands to be run.
- Tiny source
- Easy to queue a set of actions
- Works with any runnable commands
- Uses processes
- Non-blocking
Seriously, it's not a replacement for threading or multiprocessing if your application needs to share a ton of data with the children.
Usage is trivial:
from littleworkers import Pool
# Define your commands.
commands = [
'ls -al',
'cd /tmp && mkdir foo',
'date',
'echo "Hello There."',
'sleep 2 && echo "Done."'
]
# Setup a pool. Since I have two cores, I'll use two workers.
lil = Pool(workers=2)
# Run!
lil.run(commands)
For more advanced uses, please see the API documentation.
- Python 2.6+
- author
Daniel Lindsley
- updated
2011/11/10
- version
0.3.1
- license
BSD