Scrapy with selenium
Project description
Scrapy with selenium
Scrapy middleware to handle javascript pages using selenium.
Installation
$ pip install scrapy-selenium
You should use python>=3.6. You will also need one of the Selenium compatible browsers.
Configuration
- Add the browser to use, the path to the driver executable, and the arguments to pass to the executable to the scrapy settings:
from shutil import which
SELENIUM_DRIVER_NAME = 'firefox'
SELENIUM_DRIVER_EXECUTABLE_PATH = which('geckodriver')
SELENIUM_DRIVER_ARGUMENTS=['-headless'] # '--headless' if using chrome instead of firefox
Optionally, set the path to the browser executable:
SELENIUM_BROWSER_EXECUTABLE_PATH = which('firefox')
In order to use a remote Selenium driver, specify SELENIUM_COMMAND_EXECUTOR
instead of SELENIUM_DRIVER_EXECUTABLE_PATH
:
SELENIUM_COMMAND_EXECUTOR = 'http://localhost:4444/wd/hub'
- Add the
SeleniumMiddleware
to the downloader middlewares:
DOWNLOADER_MIDDLEWARES = {
'scrapy_selenium.SeleniumMiddleware': 800
}
Usage
Use the scrapy_selenium.SeleniumRequest
instead of the scrapy built-in Request
like below:
from scrapy_selenium import SeleniumRequest
yield SeleniumRequest(url=url, callback=self.parse_result)
The request will be handled by selenium, and the request will have an additional meta
key, named driver
containing the selenium driver with the request processed.
def parse_result(self, response):
print(response.request.meta['driver'].title)
For more information about the available driver methods and attributes, refer to the selenium python documentation
The selector
response attribute work as usual (but contains the html processed by the selenium driver).
def parse_result(self, response):
print(response.selector.xpath('//title/@text'))
Additional arguments
The scrapy_selenium.SeleniumRequest
accept 4 additional arguments:
wait_time
/ wait_until
When used, selenium will perform an Explicit wait before returning the response to the spider.
from selenium.webdriver.common.by import By
from selenium.webdriver.support import expected_conditions as EC
yield SeleniumRequest(
url=url,
callback=self.parse_result,
wait_time=10,
wait_until=EC.element_to_be_clickable((By.ID, 'someid'))
)
screenshot
When used, selenium will take a screenshot of the page and the binary data of the .png captured will be added to the response meta
:
yield SeleniumRequest(
url=url,
callback=self.parse_result,
screenshot=True
)
def parse_result(self, response):
with open('image.png', 'wb') as image_file:
image_file.write(response.meta['screenshot'])
script
When used, selenium will execute custom JavaScript code.
yield SeleniumRequest(
url=url,
callback=self.parse_result,
script='window.scrollTo(0, document.body.scrollHeight);',
)
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file scrapy-selenium-mm-0.1.1.tar.gz
.
File metadata
- Download URL: scrapy-selenium-mm-0.1.1.tar.gz
- Upload date:
- Size: 7.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.11.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 6a2e72bc7613c12ca92f9b0add5dc8eec6ff3df7bee7b4af0340dafe83bada1d |
|
MD5 | 2db309555d5fdb7638126b363339c2f8 |
|
BLAKE2b-256 | 1a2f9cbc992af0588e5fbe6d5d7090ba4ae0ae28ab3a1400728cefa2172f22de |
File details
Details for the file scrapy_selenium_mm-0.1.1-py3-none-any.whl
.
File metadata
- Download URL: scrapy_selenium_mm-0.1.1-py3-none-any.whl
- Upload date:
- Size: 9.1 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.11.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | f1e9ba2b0d38af6f377ddf9450915d6b86ea27fd6a3babfad6d0e1bb45966510 |
|
MD5 | de8c6115815216d62c044226b871bb68 |
|
BLAKE2b-256 | 87b2df4ed27ceb017670f06cea9b00ca0c9e6268f54f0afb8f800fbdcf017d89 |