Skip to main content

Distributed Task framework, fast than any other taskqueue

Project description

Distributed Task framework, fast than any other taskqueue

Functional description

  • A distributed crawler framework that is more flexible than scrapy and easier to use than Celery. Do the most with the least code and the simplest way

  • Can skillfully use the framework to crawl data within 1 minute, without learning complex documents. Easily expand various middleware
    Feature description:

    Support middleware:
       Support reids kafka sqlite memory  four kinds of middleware (redis is the first choice, support batch publishing tasks, distributed consumption is lightning fast)
    
    Concurrent support:
       Support process threading gevent three concurrent consumption modes (can be mixed)
    
    Frequency control and current limit:
       Precisely control how many functions are run in 1 second
    
    Task deduplication:
       If you repeatedly push the successfully consumed task, the task will be automatically filtered out
    
    Consumption confirmation:
      When consumption confirmation is enabled, and the consumption task is stopped manually, the task will not be lost
    
    Number of retries:
       When the function fails, it will retry the specified number of times immediately, and the consumption will be confirmed when the maximum number of retries is reached
    
    Task visualization:
       You can view the current task consumption in real time through the redis web version management tool
    

version description

  • Supported version: Python 3.0+

Pip installation

pip install leek

DEMO description

0. Release tasks and consumer tasks (decorator version)
from leek import task_deco

@task_deco('test0') # Added task queue decorator on consumption function
def f0(a, b):
    print(f"t_demo0,a:{a},b:{b}")

# Post task
for i in range(1, 51):
    f0.pub(i, i)

# Consumer task
f0.start()
1. Publish consumer tasks (use extra parameters)
from leek import task_deco

@task_deco('test1', qps=30, threads_num=30, max_retry_times=3, ack=True)
def f1(a, b):
    print(f"t_demo1,a:{a},b:{b}")

# Post task
for i in range(1, 31):
    f1.pub(i, i + 1) # or f1.publish_redispy(i,i+1)

# Consumer task
f1.start()
2. Publish consumer tasks (non-decorator version)
from leek import RedisPublish, RedisCustomer


for zz in range(1, 501):
    param = {"a": zz, "b": zz, "c": zz}
    RedisPublish(queue_name='test2').publish_redispy(param)


def print_msg_dict(a, b, c):
    print(f"msg_dict:{a},{b},{c}")


# Consumption of multi-parameter type tasks queue_name Consumption queue name qps The number of consumption tasks per second (there is no limit by default)
RedisCustomer(queue_name='test2', consuming_function=print_msg_dict,
              qps=50).start_consuming_message()
3. Batch submit task consumption(consumption using coroutine)
from gevent import monkey
monkey.patch_all()
from leek import task_deco

# #### 3. Submit tasks in batches
result = [{'a': i,'b': i,'c': i} for i in range(1, 51)]

# customer_type Consumer type (default thread), max_push_size the number of records submitted in batches each time (default value 50)
# If you use gevent, please add at the beginning of the code: from gevent import monkey monkey.patch_all()
@task_deco('test3', qps=50, customer_type='gevent', max_push_size=100) # Added task queue decorator on the consumption function
def f3(a, b, c):
    print(f"t_demo3:{a},{b},{c}")

# Post task
f3.pub_list(result)

# Consumer task
f3.start()
4. Switch task queue middleware to sqlite (default is redis)
from leek import task_deco, MiddlewareEum

@task_deco('test4', middleware=MiddlewareEum.SQLITE, qps=10)
def f4(a, b, c):
    print(f"t_demo4:{a},{b},{c}")

for zz in range(1, 51):
    f4.pub(zz, zz, zz)

f4.start()

Reids install

reids install

reids docker install

docker run  -d -p 6379:6379 redis

redis web manage tool redisweb avatar

Usage scenarios and features

1. Highly concurrent distributed crawler (verified by tens of millions of online data crawling verification)

2. Distributed data cleaning (automatic deduplication of cleaning, support to continue cleaning after interruption at any time)

3. Short video processing (video download and upload, bandwidth is sufficient without waiting)

4. Asynchronous real-time online query interface (speed reaches millisecond level)

5. Other usage scenarios are being expanded

Release Notes

2020-06-11 Version 4.1.5 Added support for gevent coroutine consumption parameter customer_type='gevent'

2020-05-20 Added consumption function timeout time parameter

2020-05-10 Added sqlite middleware support

2020-04-13 The consumption function adds automatic control of the number of threads

2020-04-10 New frequency limiting parameter for consumption function

2020-01-08 Consumer function supports multiple parameter types

2019-12-06 Simplified multi-threaded consumer queue class

2019-10-14 Added consumption function error retry mechanism, retry by default 3 times

2019-10-12 Task deduplication ignores parameter order

2019-09-27 Fix the bug of submitting list task

2019-05-25 Added dynamic parameter transfer when adding tasks

2019-04-06 Added automatic deduplication function for crawling tasks

2019-03-23 ​​Added single-thread asynchronous batch submission function

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

leek-1.0.3.tar.gz (5.7 kB view hashes)

Uploaded Source

Built Distribution

leek-1.0.3-py3-none-any.whl (4.4 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page