A lightweight network request framework
Project description
async-request
A lightweight network request framework based on requests & asyncio
install
pip install async_request
usage
Just like scrapy:
from async_request import AsyncSpider, Request
class MySpider(AsyncSpider):
start_urls = ['https://cn.bing.com/']
async def parse(self, response):
print(response.xpath('//a/@href').get())
yield Request('https://github.com/financialfly/async-request', callback=self.parse_github)
def parse_github(self, response):
yield {'hello': 'github'}
async def process_result(self, result):
# Process result at here.
print(result)
if __name__ == '__main__':
# Run spider
MySpider().run()
For more detailed control (like: handle cookies, download delay, concurrent requests, max retries, logs settings etc.): (refer to the constructor of the Crawler
class):
from async_request import AsyncSpider
class MySpider(AsyncSpider):
...
if __name__ == '__main__':
MySpider(
handle_cookies=True,
download_delay=0,
concurrent_requests=10,
max_retries=3,
log_file='spider.log'
).run()
test
Use fetch
function to get a response immediately:
from async_request import fetch
def parse():
response = fetch('https://www.bing.com')
print(response)
if __name__ == '__main__':
parse()
the output will like this:
<Response 200 https://cn.bing.com/>
Use the test
decorator is also a method to test spider:
import async_request as ar
@ar.test('https://www.baidu.com')
def parse(response):
print(response.url, response.status_code)
if __name__ == '__main__':
parse()
then run the script, you will see the result:
https://www.baidu.com/ 200
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
async-request-0.1887.tar.gz
(7.0 kB
view details)
File details
Details for the file async-request-0.1887.tar.gz
.
File metadata
- Download URL: async-request-0.1887.tar.gz
- Upload date:
- Size: 7.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/1.13.0 pkginfo/1.5.0.1 requests/2.14.2 setuptools/39.0.1 requests-toolbelt/0.9.1 tqdm/4.28.1 CPython/3.7.1
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 6073ac0a3aec0452a50e29a8a1480f0cfa34fca4ce7f817fe4b3bed1a0573c0e |
|
MD5 | 561460ee3c0f918ad1ac1c2436094a8e |
|
BLAKE2b-256 | 27dfeba5ae5e7a046ec0ecfdd17efa0ad74350b495835e80f01bbbe4524fee72 |