質問

Do you know/use any distributed job queue for python? Can you share links or tools

役に立ちましたか?

解決

In addition to multiprocessing there's also the Celery project, if you're using Django.

他のヒント

Pyres is a resque clone built in python. Resque is used by Github as their message queue. Both use Redis as the queue backend and provide a web-based monitoring application.

http://binarydud.github.com/pyres/intro.html

There's also "bucker" by Sylvain Hellegouarch which you can find here:

It describes itself like this:

  • bucker is a queue system that supports multiple storage for the queue (memcached, Amazon SQS for now) and is driven by XML messages sent over a TCP connections between a client and the queue server.

Look at beanstalkd

redqueue? It's implemented in python+tornado framework, speaks memcached protocol and is optionally persistent into log files. Currently it is also able to behave like beanstalkd, the reserve/delete way in memcache protocol as well.

REDQUEUE

If you think that Celery is too heavy for your needs then you might want to look at the simple distributed task queue:

It's a year late or whatever, but this is something I've hacked together to make a queue of Processes executing them only X number at a time. http://github.com/goosemo/job_queue

You probably want to look at multiprocessing's Queue. Included in Python 2.6, get it on PyPI for earlier versions of Python.

Standard library documentation: http://docs.python.org/library/multiprocessing.html On PyPI: http://pypi.python.org/pypi/multiprocessing

Also there is Unix 'at'

For more info: man at

ライセンス: CC-BY-SA帰属
所属していません StackOverflow
scroll top