A small package that contains commonly used codes while scraping
Project description
Scrap Utils
This is small package that contains some code regularly repeated when scraping.
To install
pip install scrap-utils
Sample code
import scrap_utils as su
response = su.requests_get("https://python.org")
len(response.text)
It has the following functions:
load_json(filepath, encoding=None, errors=None, parse_float=None,
parse_int=None, parse_constant=None)
dump_json(data, filepath, encoding=None, errors=None, indent=4, skipkeys=False,
ensure_ascii=False, separators=None, sort_keys=False)
to_csv(dataset, filepath, mode="a", encoding=None, errors=None, newline='',
header=True, dialect='excel', **fmtparams)
requests_get(url, trials=0, sleep_time=30, max_try=math.inf, **requests_kwargs)
requests_post(url, trials=0, sleep_time=30, max_try=math.inf, **requests_kwargs)
To-do list I'm considering:
- remove print statements
- add unittest
- soup_get()
- driver_get()
- start_firefox()
- read_csv()
Feel free to add your contribution here
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
scrap-utils-0.0.2.tar.gz
(3.5 kB
view hashes)
Built Distribution
Close
Hashes for scrap_utils-0.0.2-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | df812d0576a405229b60f10e21d9d9c54c0d732ce676dd9896388e68e79ad907 |
|
MD5 | a41c6f68a7134cba2371f6d4199c43cb |
|
BLAKE2b-256 | 606411f9ac926aa43ae783fd5de3e0fa4e79d76c97d9ec1c956b7e8a64a167c6 |