Eek, a [web] spider.
Project description
Eek is a web crawler that outputs metadata about a website in CSV format.
Installation
$ pip install eek
Usage
usage: eek [-h] [–graph] [–delay SECONDS] [–grep PATTERN] [-i] URL
eek recursively crawls a website, outputing metadata about each page in CSV format.
positional arguments:
URL The base URL to start the crawl
optional arguments:
-h, --help show this help message and exit
--graph output a graphviz digraph of links instead of CSV
metadata
--delay SECONDS Time, in seconds, to wait in between fetches. Defaults to
0.
--grep PATTERN Print urls containing PATTERN (a python regular
expression).
-i, --ignore-case Ignore case. Only valid with --grep
Example:
eek http://example.com/
To save output to a file, use redirection
eek http://example.com/ > ~/some_file.csv
To slow down crawling, use --delay=[seconds]
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
eek-1.0.2.tar.gz
(7.4 kB
view details)
File details
Details for the file eek-1.0.2.tar.gz.
File metadata
- Download URL: eek-1.0.2.tar.gz
- Upload date:
- Size: 7.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
6e8532e0fbbdbf555c4ac7589d10bad78351d9ea9f8e421004997c3cb7dce198
|
|
| MD5 |
4d0a795c358e3269b5add8613bb8e020
|
|
| BLAKE2b-256 |
77a5fb62b6a3d85972018ebef8525016634a8559a7af5094a3c2186c27926ee1
|