Skip to main content

Scrapy with selenium==4.3

Project description

Scrapy with selenium

Scrapy middleware to handle javascript pages using selenium==4.3.0.

This was originally a fork of scrapy-selenium but I couldn't use the below command on it. Therefore this is actually a rip-off of said repo.

Installation

$ pip install git+https://github.com/mrafee113/selenium_scrapy.git@v0.1.0

You should use python>=3.9. You will also need one of the Selenium compatible browsers.

Configuration

  1. Add the browser to use, the path to the driver executable, and the arguments to pass to the executable to the scrapy settings:
    from shutil import which
    
    SELENIUM_DRIVER_NAME = 'firefox'
    SELENIUM_DRIVER_EXECUTABLE_PATH = which('geckodriver')
    SELENIUM_DRIVER_ARGUMENTS=['-headless']  # '--headless' if using chrome instead of firefox
    SELENIUM_DRIVER_PREFERENCES = {
        # (for firefox) to allow browser data to be cached among sessions.
        "browser.cahce.disk.parent_directory": '/tmp/firefox-data-directory',
    
        # template firefox preferences to customize downloading files.
        "browser.download.folderList": 2,
        "browser.download.manager.showWhenStarting": False,
        "browser.download.dir": '/tmp/firefox-download-directory',
        "browser.helperApps.neverAsk.saveToDisk": "application/x-gzip"
    }
    # WARNING: To support localstorage the github project pyselenium_localstorage was used.
    #  This project has its obvious limitations (read its code).
    #  So be careful for type conversion when setting and getting values.
    SELENIUM_DRIVER_LOCALSTORAGE_DATA = [
        ("https://python.org", {"key1": "value1"}),
        ("https://google.com", {"data": "information"})
    ]
    # or alternatively
    SELENIUM_DRIVER_LOCALSTORAGE_DATA = ("https://github.ccm", {"username": "code"})
    

Optionally, set the path to the browser executable: python SELENIUM_BROWSER_EXECUTABLE_PATH = which('firefox')

In order to use a remote Selenium driver, specify SELENIUM_COMMAND_EXECUTOR instead of SELENIUM_DRIVER_EXECUTABLE_PATH: python SELENIUM_COMMAND_EXECUTOR = 'http://localhost:4444/wd/hub'

  1. Add the SeleniumMiddleware to the downloader middlewares:
    DOWNLOADER_MIDDLEWARES = {
        'scrapy_selenium.SeleniumMiddleware': 800
    }
    

Usage

Use the scrapy_selenium.SeleniumRequest instead of the scrapy built-in Request like below:

from scrapy_selenium import SeleniumRequest

yield SeleniumRequest(url=url, callback=self.parse_result)

The request will be handled by selenium, and the request will have an additional meta key, named driver containing the selenium driver with the request processed.

def parse_result(self, response):
    print(response.request.meta['driver'].title)

For more information about the available driver methods and attributes, refer to the selenium python documentation

The selector response attribute work as usual (but contains the html processed by the selenium driver).

def parse_result(self, response):
    print(response.selector.xpath('//title/@text'))

Additional arguments

The scrapy_selenium.SeleniumRequest accept 4 additional arguments:

wait_time / wait_until

When used, selenium will perform an Explicit wait before returning the response to the spider.

from selenium.webdriver.common.by import By
from selenium.webdriver.support import expected_conditions as EC

yield SeleniumRequest(
    url=url,
    callback=self.parse_result,
    wait_time=10,
    wait_until=EC.element_to_be_clickable((By.ID, 'someid'))
)

screenshot

When used, selenium will take a screenshot of the page and the binary data of the .png captured will be added to the response meta:

yield SeleniumRequest(
    url=url,
    callback=self.parse_result,
    screenshot=True
)

def parse_result(self, response):
    with open('image.png', 'wb') as image_file:
        image_file.write(response.meta['screenshot'])

script

When used, selenium will execute custom JavaScript code.

yield SeleniumRequest(
    url=url,
    callback=self.parse_result,
    script='window.scrollTo(0, document.body.scrollHeight);',
)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

selenium-scrapy-0.1.4.tar.gz (5.0 kB view details)

Uploaded Source

Built Distribution

selenium_scrapy-0.1.4-py3-none-any.whl (5.9 kB view details)

Uploaded Python 3

File details

Details for the file selenium-scrapy-0.1.4.tar.gz.

File metadata

  • Download URL: selenium-scrapy-0.1.4.tar.gz
  • Upload date:
  • Size: 5.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.9.13

File hashes

Hashes for selenium-scrapy-0.1.4.tar.gz
Algorithm Hash digest
SHA256 9077f357357b821c99c04b2b474a0944b2c01d7ddfb2d8e449c603661967e6ec
MD5 c5bc7bc2106a3615dc3e27b1cb37a38f
BLAKE2b-256 de9043a83293047b9719cd91ae94a99a391c1aed3b96b245ff7e47d2ea5a3d45

See more details on using hashes here.

File details

Details for the file selenium_scrapy-0.1.4-py3-none-any.whl.

File metadata

File hashes

Hashes for selenium_scrapy-0.1.4-py3-none-any.whl
Algorithm Hash digest
SHA256 d67bbda2797dc1a6129b60c1db57f563b087e85b850ecce8459a3737644ad495
MD5 d6c41268d585541390f2a1e3530346d0
BLAKE2b-256 90dc864b7df8c41c1b26558cc6807b7a65cd7721f926a3a826d6f31f2d914d75

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page