Skip to main content

spaCy pipelines for pre-trained BERT and other transformers

Project description

spacy-transformers: Use pretrained transformers like BERT, XLNet and GPT-2 in spaCy

This package provides spaCy components and architectures to use transformer models via Hugging Face's transformers in spaCy. The result is convenient access to state-of-the-art transformer architectures, such as BERT, GPT-2, XLNet, etc.

This release requires spaCy v3. For the previous version of this library, see the v0.6.x branch.

Azure Pipelines PyPi GitHub Code style: black

Features

  • Use pretrained transformer models like BERT, RoBERTa and XLNet to power your spaCy pipeline.
  • Easy multi-task learning: backprop to one transformer model from several pipeline components.
  • Train using spaCy v3's powerful and extensible config system.
  • Automatic alignment of transformer output to spaCy's tokenization.
  • Easily customize what transformer data is saved in the Doc object.
  • Easily customize how long documents are processed.
  • Out-of-the-box serialization and model packaging.

🚀 Installation

Installing the package from pip will automatically install all dependencies, including PyTorch and spaCy. Make sure you install this package before you install the models. Also note that this package requires Python 3.6+, PyTorch v1.5+ and spaCy v3.0+.

pip install spacy[transformers]

For GPU installation, find your CUDA version using nvcc --version and add the version in brackets, e.g. spacy[transformers,cuda92] for CUDA9.2 or spacy[transformers,cuda100] for CUDA10.0.

If you are having trouble installing PyTorch, follow the instructions on the official website for your specific operating system and requirements, or try the following:

pip install spacy-transformers -f https://download.pytorch.org/whl/torch_stable.html

📖 Documentation

⚠️ Important note: This package has been extensively refactored to take advantage of spaCy v3.0. Previous versions that were built for spaCy v2.x worked considerably differently. Please see previous tagged versions of this README for documentation on prior versions.

Project details


Release history Release notifications | RSS feed

This version

1.1.7

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

spacy-transformers-1.1.7.tar.gz (42.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

spacy_transformers-1.1.7-py2.py3-none-any.whl (53.5 kB view details)

Uploaded Python 2Python 3

File details

Details for the file spacy-transformers-1.1.7.tar.gz.

File metadata

  • Download URL: spacy-transformers-1.1.7.tar.gz
  • Upload date:
  • Size: 42.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.7.9

File hashes

Hashes for spacy-transformers-1.1.7.tar.gz
Algorithm Hash digest
SHA256 96b0b524442edff617e28e5acece140d3b2bb5f8a6dc82720faa6509bc791b71
MD5 b21c1eff1cf80a8ffdd945507144d2bd
BLAKE2b-256 1233ad65b2abf14d3bda727d294d1a806a4df45301b9eea44038d6906957e56b

See more details on using hashes here.

File details

Details for the file spacy_transformers-1.1.7-py2.py3-none-any.whl.

File metadata

File hashes

Hashes for spacy_transformers-1.1.7-py2.py3-none-any.whl
Algorithm Hash digest
SHA256 0708c8e0f100e22b67c23578fd61f975055729f2dff4ab423fc526513b62ec7f
MD5 e313c872126863990cea60b0b4751e0a
BLAKE2b-256 d89cfde2bdcd45caaf36bb0415092d5e74954b122d2d49b1e2e8929f06f58f70

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page