Skip to main content

Python SDK for the Scraping Pros API — web scraping with browser rendering, proxy rotation, and structured data extraction.

Project description

scrapingpros

Python SDK for the Scraping Pros API — web scraping with browser rendering, proxy rotation, and structured data extraction.

Installation

pip install scrapingpros

On Windows, if you get "Permission denied", use:

python -m pip install scrapingpros

Quick Start

No signup needed — use the demo token to start immediately:

from scrapingpros import ScrapingPros

client = ScrapingPros("demo_6x595maoA6GdOdVb")
result = client.scrape("https://example.com")
print(result.html)

The demo token includes 5,000 credits/month and 30 req/min (1 simple request = 1 credit, 1 browser request = 5 credits). Credits are NOT consumed for requests that fail due to infrastructure errors. For higher limits, contact the Scraping Pros team.

Usage Examples

Markdown Output (for AI/LLM)

result = client.scrape("https://example.com", format="markdown")
print(result.markdown)

Browser Rendering

result = client.scrape(
    "https://spa-site.com",
    browser=True,
    use_proxy="any",
)

Structured Data Extraction

result = client.scrape(
    "https://quotes.toscrape.com/",
    extract={
        "quotes": {"selector": "css:.text", "multiple": True},
        "authors": {"selector": "css:.author", "multiple": True},
    },
)
print(result.extracted_data["quotes"])

Async Batch Processing

collection = client.create_collection("my-batch", [
    {"url": "https://example.com/1"},
    {"url": "https://example.com/2"},
])
run = client.run_and_wait(collection.id)
print(f"Done: {run.success_requests}/{run.total_requests}")

Async Client

from scrapingpros import AsyncScrapingPros

async with AsyncScrapingPros("demo_6x595maoA6GdOdVb") as client:
    result = await client.scrape("https://example.com", format="markdown")

API Methods

Method Description
client.scrape(url, ...) Scrape a URL (HTML or markdown)
client.download(url, ...) Download a file as base64
client.create_collection(name, requests) Create a batch collection
client.run_and_wait(collection_id) Run a batch and wait for completion
client.create_viability_test(urls) Analyze sites before scraping
client.list_proxy_countries() List available proxy countries
client.billing() Check usage and billing
client.health() API health check

Error Handling

from scrapingpros import ScrapingPros, AuthenticationError, RateLimitError, QuotaExceededError

try:
    result = client.scrape("https://example.com")
except AuthenticationError:
    print("Invalid token — use demo_6x595maoA6GdOdVb for testing")
except RateLimitError as e:
    print(f"Rate limited. Retry after {e.retry_after}s")
except QuotaExceededError:
    print("Monthly quota exceeded — upgrade your plan for more requests")

All exceptions inherit from ScrapingProsError.

Usage & Quota Tracking

# Check remaining quota after any API call
client.scrape("https://example.com")
print(f"Requests remaining: {client.quota_remaining}")
print(f"Rate limit remaining: {client.rate_limit_remaining}")

# Get detailed billing info
billing = client.billing()

Plans

Plan Price Credits/mo Rate Concurrent
Demo (public) Free 5,000 30/min 5
Free $0 1,000 30/min 5
Starter $29 25,000 30/min 10
Growth $69 100,000 60/min 20
Pro $199 500,000 120/min 50
Scale $499 2,500,000 200/min 100
Enterprise Custom Unlimited 2,000/min Custom

1 simple request = 1 credit, 1 browser request = 5 credits. See all features: client.plans() or visit scrapingpros.com.

Configuration

client = ScrapingPros(
    "demo_6x595maoA6GdOdVb",  # or your dedicated token / SP_TOKEN env var
    base_url="https://api.scrapingpros.com",  # default
    timeout=120.0,         # request timeout in seconds
    max_retries=3,         # auto-retry on 429
)

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

scrapingpros-0.2.4.tar.gz (68.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

scrapingpros-0.2.4-py3-none-any.whl (29.0 kB view details)

Uploaded Python 3

File details

Details for the file scrapingpros-0.2.4.tar.gz.

File metadata

  • Download URL: scrapingpros-0.2.4.tar.gz
  • Upload date:
  • Size: 68.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.10

File hashes

Hashes for scrapingpros-0.2.4.tar.gz
Algorithm Hash digest
SHA256 b1155747ab207892a0bbf4741d6a4732074bc46ed085a183b6ae694c20126d9c
MD5 e4c400d59aa4cb13ca13482f88ded306
BLAKE2b-256 f8a28cece27860d9e441580904a23ba746fc2318c3bab9fd54eaa03988d5e52d

See more details on using hashes here.

File details

Details for the file scrapingpros-0.2.4-py3-none-any.whl.

File metadata

  • Download URL: scrapingpros-0.2.4-py3-none-any.whl
  • Upload date:
  • Size: 29.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.10

File hashes

Hashes for scrapingpros-0.2.4-py3-none-any.whl
Algorithm Hash digest
SHA256 a4624ee52353c1a2684b7d3bb3ee9102c3da7d4378cc612fad007b53eb3b5078
MD5 8e151dd942fbde9fd0fbd30e499a6cc5
BLAKE2b-256 80a3ca4ad0afc172b5db9de693feb016c06d7b3d4c80ea14788e00c98e16c035

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page