Skip to main content

Essential Python library that scrapes HTTP(S) and UDP trackers for torrent information.

Project description

Scrapeer-py

A tiny Python library that lets you scrape HTTP(S) and UDP trackers for torrent information.

Scrapeer-py is a Python port of the original PHP Scrapeer library by TorrentPier.

Overview

Scrapeer-py allows you to retrieve peer information from BitTorrent trackers using both HTTP(S) and UDP protocols. It can fetch seeders, leechers, and completed download counts for multiple torrents from multiple trackers simultaneously.

Features

  • Support for both HTTP(S) and UDP tracker protocols
  • Batch scraping of multiple infohashes at once (up to 64)
  • Support for trackers with passkeys
  • Optional announce mode for trackers that don't support scrape
  • Configurable timeout settings
  • Detailed error reporting
  • Well-organized modular codebase

Installation

pip install scrapeer

Usage

Scrapeer-py can be used both as a Python library and as a command-line tool.

Python Library Usage

from scrapeer import Scraper

# Initialize the scraper
scraper = Scraper()

# Define your infohashes and trackers
infohashes = [
    "0123456789abcdef0123456789abcdef01234567",
    "fedcba9876543210fedcba9876543210fedcba98"
]

trackers = [
    "udp://tracker.example.com:80",
    "http://tracker.example.org:6969/announce",
    "https://private-tracker.example.net:443/YOUR_PASSKEY/announce"
]

# Get the results (timeout of 3 seconds per tracker)
results = scraper.scrape(
    hashes=infohashes,
    trackers=trackers,
    timeout=3
)

# Print the results
for infohash, data in results.items():
    print(f"Results for {infohash}:")
    print(f"  Seeders: {data['seeders']}")
    print(f"  Leechers: {data['leechers']}")
    print(f"  Completed: {data['completed']}")

# Check if there were any errors
if scraper.has_errors():
    print("\nErrors:")
    for error in scraper.get_errors():
        print(f"  {error}")

Command-Line Usage

After installation, you can use the scrapeer command directly:

# Basic usage
scrapeer INFOHASH1 INFOHASH2 -t TRACKER1 TRACKER2

# Example with real values
scrapeer abc123def456...890 fedcba987654...321 \
  -t udp://tracker.example.com:80 \
  -t http://tracker.example.org:6969/announce

# With options
scrapeer INFOHASH -t TRACKER --timeout 5 --announce --json

# Get help
scrapeer --help

CLI Options

  • -t, --trackers: One or more tracker URLs (required)
  • --timeout: Timeout in seconds for each tracker (default: 2)
  • --announce: Use announce instead of scrape
  • --max-trackers: Maximum number of trackers to scrape
  • --json: Output results in JSON format
  • --quiet, -q: Suppress error messages
  • --version: Show version information

CLI Examples

Basic scraping:

scrapeer d4344b390d7bc7b7d332c6d89ef1ff5d6f78ca48 \
  -t udp://tracker.opentrackr.org:1337/announce

Multiple hashes and trackers:

scrapeer hash1 hash2 hash3 \
  -t udp://tracker1.com:80 \
  -t http://tracker2.org:8080/announce \
  --timeout 10

JSON output for scripting:

scrapeer INFOHASH -t TRACKER --json > results.json

Private tracker with passkey:

scrapeer INFOHASH \
  -t https://private-tracker.net:443/YOUR_PASSKEY/announce \
  --announce

Package Structure

Scrapeer-py is organized into the following modules:

  • scrapeer/ - Main package directory
    • __init__.py - Package initialization that exports the Scraper class
    • scraper.py - Main Scraper class implementation
    • http.py - HTTP(S) protocol scraping functionality
    • udp.py - UDP protocol scraping functionality
    • utils.py - Utility functions used across the package

API Reference

Scraper class

scrape(hashes, trackers, max_trackers=None, timeout=2, announce=False)

Scrape trackers for torrent information.

  • Parameters:

    • hashes: List (>1) or string of infohash(es)
    • trackers: List (>1) or string of tracker(s)
    • max_trackers: (Optional) Maximum number of trackers to be scraped, Default all
    • timeout: (Optional) Maximum time for each tracker scrape in seconds, Default 2
    • announce: (Optional) Use announce instead of scrape, Default False
  • Returns:

    • Dictionary of results with infohashes as keys and stats as values

has_errors()

Checks if there are any errors.

  • Returns:
    • bool: True if errors are present, False otherwise

get_errors()

Returns all the errors that were logged.

  • Returns:
    • list: All the logged errors

Limitations

  • Maximum of 64 infohashes per request
  • Minimum of 1 infohash per request
  • Only supports BitTorrent trackers (HTTP(S) and UDP)

License

This project is licensed under the MIT License - see the LICENSE.txt file for details.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

scrapeer-1.0.3.tar.gz (29.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

scrapeer-1.0.3-py3-none-any.whl (16.7 kB view details)

Uploaded Python 3

File details

Details for the file scrapeer-1.0.3.tar.gz.

File metadata

  • Download URL: scrapeer-1.0.3.tar.gz
  • Upload date:
  • Size: 29.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.11

File hashes

Hashes for scrapeer-1.0.3.tar.gz
Algorithm Hash digest
SHA256 a7f39b66ef566de196433a6f6dfe94201dee0d341d345593e090b14a97b8f338
MD5 ac536f05e2a6811ac807ad77c0c1c480
BLAKE2b-256 4992fd994d3bfff7d21c7ce086c077ff43c0105cbbf2c58d8298836cd39e61e8

See more details on using hashes here.

File details

Details for the file scrapeer-1.0.3-py3-none-any.whl.

File metadata

  • Download URL: scrapeer-1.0.3-py3-none-any.whl
  • Upload date:
  • Size: 16.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.11

File hashes

Hashes for scrapeer-1.0.3-py3-none-any.whl
Algorithm Hash digest
SHA256 8c0adc64cf969e77e4884d5688d1627f34b5c1195285e5f522ca4d1001d383e8
MD5 8064d674f22e87281b5b417413cae271
BLAKE2b-256 0191f4b29b7a8f1578712d3b7a805f44e22c20956fb6eeefce3dcc7573a04e0e

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page