scrape proxies from more than 5 different sources and check which ones are still alive
Project description
Proxy Scraper and Checker
Scrape more than 1K HTTP - HTTPS - SOCKS4 - SOCKS5 proxies in less than 2 seconds.
Scraping fresh public proxies from different sources:
- sslproxies.org (HTTP, HTTPS)
- free-proxy-list.net (HTTP, HTTPS)
- us-proxy.org (HTTP, HTTPS)
- socks-proxy.net (Socks4, Socks5)
- proxyscrape.com (HTTP, Socks4, Socks5)
- proxy-list.download (HTTP, HTTPS, Socks4, Socks5)
- geonode.com (HTTP, HTTPS, Socks4, Socks5)
Installation
You can install the package directly from PyPI using pip
:
pip install proxyz
Alternatively, you can install dependencies manually if you're working from the source code:
pip3 install -r requirements.txt
Usage
Using the Command-Line Interface
Once installed via pip
, you can use the command-line tools proxy_scraper
and proxy_checker
directly.
For Scraping Proxies:
proxy_scraper -p http
- With
-p
or--proxy
, you can choose your proxy type. Supported proxy types are: HTTP - HTTPS - Socks (Both 4 and 5) - Socks4 - Socks5. - With
-o
or--output
, specify the output file name where the proxies will be saved. (Default is output.txt). - With
-v
or--verbose
, increase output verbosity. - With
-h
or--help
, show the help message.
For Checking Proxies:
proxy_checker -p http -t 20 -s https://google.com -l output.txt
- With
-t
or--timeout
, set the timeout in seconds after which the proxy is considered dead. (Default is 20). - With
-p
or--proxy
, check HTTPS, HTTP, SOCKS4, or SOCKS5 proxies. (Default is HTTP). - With
-l
or--list
, specify the path to your proxy list file. (Default is output.txt). - With
-s
or--site
, check proxies against a specific website like google.com. (Default is https://google.com). - With
-r
or--random_agent
, use a random user agent per proxy. - With
-v
or--verbose
, increase output verbosity. - With
-h
or--help
, show the help message.
Running Directly from Source
If you prefer running the scripts directly from the source code, you can use the following commands:
For Scraping:
python3 proxyScraper.py -p http
For Checking:
python3 proxyChecker.py -p http -t 20 -s https://google.com -l output.txt
Good to Know
- Dead proxies will be removed, and only alive proxies will remain in the output file.
- This script is capable of scraping SOCKS proxies, but
proxyChecker
currently only checks HTTP(S) proxies.
Star History
Contributing
Pull requests are welcome. For major changes, please open an issue first to discuss what you would like to change.
Issues
Feel free to submit issues and enhancement requests or contact me via vida.page/nima.
License
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file proxyz-0.2.0.tar.gz
.
File metadata
- Download URL: proxyz-0.2.0.tar.gz
- Upload date:
- Size: 6.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.1 CPython/3.12.4
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | c63f54048fb16931e82ed2ce89629469c070277036b31c8625f0b8457a28d90d |
|
MD5 | 58ede78c157df54c85667b877228c630 |
|
BLAKE2b-256 | 32964663c083fbf3c96499b2c5fd81278858afbf2ab612a65112fa96ba5da795 |
File details
Details for the file proxyz-0.2.0-py3-none-any.whl
.
File metadata
- Download URL: proxyz-0.2.0-py3-none-any.whl
- Upload date:
- Size: 7.8 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.1 CPython/3.12.4
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | dc4bc9d6c961fcbe419cb250a80ab6b3860dbf3566ac4baab4ed4f812ea856ff |
|
MD5 | 6d5291becbaf777c1910381737d7ae25 |
|
BLAKE2b-256 | d54750264df747c72957486fdd42a4a24ad2e21de05de8a86be49d76524dc57e |