Skip to main content

No project description provided

Project description

FreeProxyScraper

This is a plugin driven web scraper meant to retrieve and test free proxies for use. Note that this package may be unstable and should not be used in a production environment.

Installation

Run the following to install:

pip install FreeProxyScraper

Usage

import FreeProxyScraper

pq = FreeProxyScraper.ProxyQuery()

# Returns any proxy's found
for proxy in pq.find_proxies(limit=20):
    print(proxy)

# Returns only proxies that are anonymous or "elite"
for proxy in pq.find_filter(limit=20, min_anon_level=1):
    print(proxy)

There are 3 anonymity levels, indicated as integers between 0-2.

  • Level 0: Transparent. The end server can see your real ip even though it's being routed through a proxy
  • Level 1: Anonymous. The end server knows you are using a proxy, but does not know your real ip
  • Level 2: High Anonymity, also sometimes called "elite". The end server does not know you are using a proxy or know your real ip. The end server may have a database known proxies, so they still may know that you are using a proxy by matching your ip against such a database.

List of sites implemented for scraping:

FAQ

  • Why implement so many websites for scraping?

Websites are always changing, or going down, or banning ip's very quickly. In order to make sure this package stays reliable it is essential that it implements many websites

  • I want to make sure that I am truly not using transparent proxies, how do I know the websites being scraped from aren't lying abut the anonymity of the proxies?

By default, all proxies will be checked if they are transparent or not before ever giving them to you if you specified a higher anon_level then 0. There's no need to worry, your ip should be safe.

Development

to install FreeProxyScraper, along with the tools you need to develop, run the following in the directory containing this repo:

pip install -e .[dev]

If you'd like to contribute to development, right now the most needed thing is writing more plugins. In order to help, you need basic knowledge of BeautifulSoup4 and a little of patience with websites purposely making it hard for you to scrape information. Check out src/plugins/examplePlugin.py to see an example layout of a plugin file.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

FreeProxyScraper-0.1.17.tar.gz (12.6 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

FreeProxyScraper-0.1.17-py3-none-any.whl (13.7 kB view details)

Uploaded Python 3

File details

Details for the file FreeProxyScraper-0.1.17.tar.gz.

File metadata

  • Download URL: FreeProxyScraper-0.1.17.tar.gz
  • Upload date:
  • Size: 12.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.2 importlib_metadata/4.6.4 pkginfo/1.7.1 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.48.1 CPython/3.8.1

File hashes

Hashes for FreeProxyScraper-0.1.17.tar.gz
Algorithm Hash digest
SHA256 96119cc4e5e4a4f7eb360d102830c228cde3ec7c14a97419cdd3609839f99a06
MD5 5c828e8b938d7cf18e2def9433d943dd
BLAKE2b-256 d058cd036bc135080c8da969bbf352369f69f7285c0ecd1fac14477ddc1ba619

See more details on using hashes here.

File details

Details for the file FreeProxyScraper-0.1.17-py3-none-any.whl.

File metadata

  • Download URL: FreeProxyScraper-0.1.17-py3-none-any.whl
  • Upload date:
  • Size: 13.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.2 importlib_metadata/4.6.4 pkginfo/1.7.1 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.48.1 CPython/3.8.1

File hashes

Hashes for FreeProxyScraper-0.1.17-py3-none-any.whl
Algorithm Hash digest
SHA256 d86b501edb13f945fdd22db4bc8c86d439f362384ed384af8e8ad509d7da404a
MD5 3d4c3bfffb2417ee1e3d1f078112cf6c
BLAKE2b-256 bed69a33395e6e6dba298a563a7b534be3c847a73e32076ca63ad1f55b6b1542

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page