Skip to main content

A collection of scrapy item pipelines

Project description

Overview

CII Best Practices pylint Score Travis branch Coverage Report codebeat badge https://api.codacy.com/project/badge/Grade/aeda92e058434a9eb2e8b0512a02235f Updates Known Vulnerabilities Code style: black License: AGPL v3

These pipelines enable Scrapy to save items into various backends, including:

  • MongoDB

And also these pipelines provide multiple ways to save or update the items, and return id created by backends

Requirements

Python 3
  • Python 3.6+

  • Works on Linux, Windows, Mac OSX

Installation

PyPI PyPI - Python Version PyPI - Wheel

The quick way:

pip install scrapy-pipelines

For more details see the installation section in the documentation: https://scrapy-pipelines.readthedocs.io/en/latest/intro/installation.html

Documentation

Documentation is available online at https://scrapy-pipelines.readthedocs.io/en/latest/ and in the docs directory.

Community (blog, twitter, mail list, IRC)

Keeping this section same as Scrapy is intending to benefit back to Scrapy.

See https://scrapy.org/community/

Contributing

Keeping this section same as Scrapy is intending to be easier when this repo merge back to Scrapy.

See https://doc.scrapy.org/en/master/contributing.html

Code of Conduct

Please note that this project is released with a Contributor Code of Conduct (see https://github.com/scrapy/scrapy/blob/master/CODE_OF_CONDUCT.md).

By participating in this project you agree to abide by its terms. Please report unacceptable behavior to opensource@scrapinghub.com.

Companies using Scrapy

Keeping this section same as Scrapy is intending to benefit back to Scrapy.

See https://scrapy.org/companies/

Commercial Support

Keeping this section same as Scrapy is intending to benefit back to Scrapy.

See https://scrapy.org/support/

TODO

  • [X] Add indexes creation in open_spider()

  • [X] Add item_completed method

  • [X] Add signals for MongoDB document’s id return

  • [ ] Add MongoDB document update

  • [ ] Add Percona Server for MongoDB docker support

  • [ ] Add Redis support

  • [ ] Add InfluxDB support

  • [ ] Add LevelDB support

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

Scrapy-Pipelines-0.2.tar.gz (22.1 kB view details)

Uploaded Source

Built Distribution

Scrapy_Pipelines-0.2-py3-none-any.whl (25.0 kB view details)

Uploaded Python 3

File details

Details for the file Scrapy-Pipelines-0.2.tar.gz.

File metadata

  • Download URL: Scrapy-Pipelines-0.2.tar.gz
  • Upload date:
  • Size: 22.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.13.0 pkginfo/1.5.0.1 requests/2.22.0 setuptools/41.0.1 requests-toolbelt/0.9.1 tqdm/4.32.1 CPython/3.6.7

File hashes

Hashes for Scrapy-Pipelines-0.2.tar.gz
Algorithm Hash digest
SHA256 6812b3438a6c1cdfbb024f3448101637fc40db03731a16b5533f93a09617b392
MD5 709d033b87674d040029ae36952a0d69
BLAKE2b-256 13d66ca908c0584ad4f61546d407efaad855094f7003d4a67a6a1b40692f1c29

See more details on using hashes here.

File details

Details for the file Scrapy_Pipelines-0.2-py3-none-any.whl.

File metadata

  • Download URL: Scrapy_Pipelines-0.2-py3-none-any.whl
  • Upload date:
  • Size: 25.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.13.0 pkginfo/1.5.0.1 requests/2.22.0 setuptools/41.0.1 requests-toolbelt/0.9.1 tqdm/4.32.1 CPython/3.6.7

File hashes

Hashes for Scrapy_Pipelines-0.2-py3-none-any.whl
Algorithm Hash digest
SHA256 e5f336e5b69b79e6653dcc5d49a34184580107c7a84941f973ff5e69b64710fa
MD5 c67e197b3eba77f03d0ebfa581134572
BLAKE2b-256 74104fa4c50099386727714960d96822d51908d30ab2ce582da134032586ab01

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page