this is a small spider,you can easy running. When you often need to crawl a single site, you can reduce some repeated code every time, using this small framework you can quickly crawl data into a file or database.
Project description
lrabbit_scrapy
this is a small spider,you can easy running. When you often need to crawl a single site, you don't have to redo some repeated code every time, using this small framework you can quickly crawl data into a file or database.
Installing
$ pip install lrabbit_scrapy
quick start
- create blog_spider.py
- python3 -m lrabbit_scrapy new_scrapy blog
- this is file content
from lrabbit_scrapy.spider import LrabbitSpider
from lrabbit_scrapy.common_utils.network_helper import RequestSession
from lrabbit_scrapy.common_utils.print_log_helper import LogUtils
from lrabbit_scrapy.common_utils.all_in_one import FileStore
import os
from lrabbit_scrapy.common_utils.mysql_helper import MysqlClient
from parsel import Selector
class Spider(LrabbitSpider):
"""
spider_name : lrabbit blog spider
"""
# unique spider name
spider_name = "lrabbit_blog"
# max thread worker numbers
max_thread_num = 2
# is open for every thread a mysql connection,if your max_thread_num overpass 10 and in code need mysql query ,you need open this config
thread_mysql_open = True
# reset all task_list,every restart program will init task list
reset_task_config = False
# open loop init_task_list ,when your task is all finish,and you want again ,you can open it
loop_task_config = False
# remove config option,if open it,then confirm option when you init task
remove_confirm_config = False
# config_path_name, this is env name ,is this code ,you need in linux to execute: export config_path="crawl.ini"
config_env_name = "config_path"
# redis db_num
redis_db_config = 0
# debug log ,open tracback log
debug_config = False
def __init__(self):
super().__init__()
self.session = RequestSession()
self.proxy_session = RequestSession(proxies=None)
csv_path = os.path.join(os.path.abspath(os.getcwd()), f"{self.spider_name}.csv")
self.field_names = ['id', 'title', 'datetime']
self.blog_file = FileStore(file_path=csv_path, filed_name=self.field_names)
def worker(self, *args):
task = args[0]
mysql_client: MysqlClient
if len(args) == 2:
mysql_client = args[1]
# mysql_client.execute("")
res = self.session.send_request(method='GET', url=f'http://www.lrabbit.life/post_detail/?id={task}')
selector = Selector(res.text)
title = selector.css(".detail-title h1::text").get()
datetime = selector.css(".detail-info span::text").get()
if title:
post_data = {"id": task, "title": title, 'datetime': datetime}
self.blog_file.write(post_data)
# when you succes get content update redis stat
self.update_stat_redis()
LogUtils.log_finish(task)
def init_task_list(self):
# you can get init task from mysql
# res = self.mysql_client.query("select id from rookie limit 100 ")
# return [task['id'] for task in res]
return [i for i in range(100)]
if __name__ == '__main__':
spider = Spider()
spider.run()
-
set config.ini and config env variable
- create crawl.ini,for example this file path is /root/crawl.ini
[server] mysql_user = root mysql_password = 123456 mysql_database = test mysql_host = 192.168.1.1 redis_user = lrabbit redis_host = 192.168.1.1 redis_port = 6379 redis_password = 123456 [test] mysql_user = root mysql_password = 123456 mysql_database = test mysql_host = 192.168.1.1 redis_user = lrabbit redis_host = 192.168.1.1 redis_port = 6379 redis_password = 123456
- set config env
- windows power shell
- $env:config_path = "/root/crawl.ini"
- linux
- export config_path="/root/crawl.ini"
-
python3 blog_spider.py
other function
- python3 blog_spider.py stat
- show task stat
- python3 -m lrabbit-scrapy sslpass
- pass android ssl Links
- author: https://www.lrabbit.life/
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file lrabbit_scrapy-2.0.8.tar.gz.
File metadata
- Download URL: lrabbit_scrapy-2.0.8.tar.gz
- Upload date:
- Size: 22.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.7.0 importlib_metadata/4.8.2 pkginfo/1.8.2 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.62.3 CPython/3.6.8
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
59cf4356e05dbd65514dab215ca3d65c385d3114aa668752f835c86843a287dd
|
|
| MD5 |
b5ff1d64b2abc0e7a7ad37e5f9b33036
|
|
| BLAKE2b-256 |
8074477cb807d286fccb3a7b3a20a97f715d48637d4ed5893ecfb4bab25c77b3
|
File details
Details for the file lrabbit_scrapy-2.0.8-py3-none-any.whl.
File metadata
- Download URL: lrabbit_scrapy-2.0.8-py3-none-any.whl
- Upload date:
- Size: 27.0 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.7.0 importlib_metadata/4.8.2 pkginfo/1.8.2 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.62.3 CPython/3.6.8
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
c4738cc146ea66c7bca6b027ce256753031c6f010873ed32317d0c6aaebde35f
|
|
| MD5 |
93bb0f6edc623994d8cc2e96c2cd8e08
|
|
| BLAKE2b-256 |
7482d14529a080def6d41c2ebf50b4ee60c09a43577866c340a83d45229aaf19
|