Skip to main content

Scrapy with selenium

Project description

Scrapy with selenium

PyPI Build Status Test Coverage Maintainability

Scrapy middleware to handle javascript pages using selenium.

Installation

$ pip install scrapy-selenium

You should use python>=3.6. You will also need one of the Selenium compatible browsers.

Configuration

  1. Add the browser to use, the path to the driver executable, and the arguments to pass to the executable to the scrapy settings:
from shutil import which
SELENIUM_DRIVER_NAME = 'firefox'
SELENIUM_DRIVER_EXECUTABLE_PATH = which('geckodriver')
SELENIUM_DRIVER_ARGUMENTS=['-headless']  # '--headless' if using chrome instead of firefox

Optionally, set the path to the browser executable:

SELENIUM_BROWSER_EXECUTABLE_PATH = which('firefox')

In order to use a remote Selenium driver, specify SELENIUM_COMMAND_EXECUTOR instead of SELENIUM_DRIVER_EXECUTABLE_PATH:

SELENIUM_COMMAND_EXECUTOR = 'http://localhost:4444/wd/hub'
  1. Add the SeleniumMiddleware to the downloader middlewares:
DOWNLOADER_MIDDLEWARES = {
    'scrapy_selenium.SeleniumMiddleware': 800
}

Usage

Use the scrapy_selenium.SeleniumRequest instead of the scrapy built-in Request like below:

from scrapy_selenium import SeleniumRequest

yield SeleniumRequest(url=url, callback=self.parse_result)

The request will be handled by selenium, and the request will have an additional meta key, named driver containing the selenium driver with the request processed.

def parse_result(self, response):
    print(response.request.meta['driver'].title)

For more information about the available driver methods and attributes, refer to the selenium python documentation

The selector response attribute work as usual (but contains the html processed by the selenium driver).

def parse_result(self, response):
    print(response.selector.xpath('//title/@text'))

Additional arguments

The scrapy_selenium.SeleniumRequest accept 4 additional arguments:

wait_time / wait_until

When used, selenium will perform an Explicit wait before returning the response to the spider.

from selenium.webdriver.common.by import By
from selenium.webdriver.support import expected_conditions as EC

yield SeleniumRequest(
    url=url,
    callback=self.parse_result,
    wait_time=10,
    wait_until=EC.element_to_be_clickable((By.ID, 'someid'))
)

screenshot

When used, selenium will take a screenshot of the page and the binary data of the .png captured will be added to the response meta:

yield SeleniumRequest(
    url=url,
    callback=self.parse_result,
    screenshot=True
)

def parse_result(self, response):
    with open('image.png', 'wb') as image_file:
        image_file.write(response.meta['screenshot'])

script

When used, selenium will execute custom JavaScript code.

yield SeleniumRequest(
    url=url,
    callback=self.parse_result,
    script='window.scrollTo(0, document.body.scrollHeight);',
)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

scrapy-selenium-mm-0.1.1.tar.gz (7.5 kB view details)

Uploaded Source

Built Distribution

scrapy_selenium_mm-0.1.1-py3-none-any.whl (9.1 kB view details)

Uploaded Python 3

File details

Details for the file scrapy-selenium-mm-0.1.1.tar.gz.

File metadata

  • Download URL: scrapy-selenium-mm-0.1.1.tar.gz
  • Upload date:
  • Size: 7.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.11.6

File hashes

Hashes for scrapy-selenium-mm-0.1.1.tar.gz
Algorithm Hash digest
SHA256 6a2e72bc7613c12ca92f9b0add5dc8eec6ff3df7bee7b4af0340dafe83bada1d
MD5 2db309555d5fdb7638126b363339c2f8
BLAKE2b-256 1a2f9cbc992af0588e5fbe6d5d7090ba4ae0ae28ab3a1400728cefa2172f22de

See more details on using hashes here.

File details

Details for the file scrapy_selenium_mm-0.1.1-py3-none-any.whl.

File metadata

File hashes

Hashes for scrapy_selenium_mm-0.1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 f1e9ba2b0d38af6f377ddf9450915d6b86ea27fd6a3babfad6d0e1bb45966510
MD5 de8c6115815216d62c044226b871bb68
BLAKE2b-256 87b2df4ed27ceb017670f06cea9b00ca0c9e6268f54f0afb8f800fbdcf017d89

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page