Export scraped items of different types to multiple feeds.
Project description
This Scrapy extension exports scraped items of different types to multiple feeds. By default each item gets its own feed.
Installation
$ pip install scrapy-multifeedexporter
Configuration
You’ll have to switch the default FeedExporter with MultiFeedExporter by adding the following lines to the settings.py file of your spider:
from multifeedexporter import MultiFeedExporter
EXTENSIONS = {
'scrapy.contrib.feedexport.FeedExporter': None,
'multifeedexporter.MultiFeedExporter': 500,
}
# Automatically configure available item names from your module
MULTIFEEDEXPORTER_ITEMS = MultiFeedExporter.get_bot_items(BOT_NAME)
Usage
When calling scrapy crawl you need to use the %(item_name)s placeholder in the output file/URI name. The following calls to scrapy crawl demonstrate the placeholder:
$ scrapy crawl -o "spider_name_%(item_name)s.csv" -t csv spider_name
$ scrapy crawl -o "ftp://foo:bar@example.com/spider_name_%(item_name)s.csv" -t csv spider_name
If you omit the placeholder, all items will be placed in one file.
License
scrapy-multifeedexporter is published under MIT license
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
File details
Details for the file scrapy-multifeedexporter-0.1.1.tar.gz
.
File metadata
- Download URL: scrapy-multifeedexporter-0.1.1.tar.gz
- Upload date:
- Size: 2.7 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 6bf3fc7d192b68dbe1b33d823da89298a234af95f7b346cbb2cf3edf9c6378d3 |
|
MD5 | 235617822829e0ae92424e76633aceff |
|
BLAKE2b-256 | 49959956776a86f0c2318a0ab33fa7b3897bb571e0932bad39f79db809b6bdd7 |