Skip to main content

Facebook AI Research Sequence-to-Sequence Toolkit

Project description



Support Ukraine MIT License Latest Release Build Status Documentation Status CicleCI Status


Fairseq(-py) is a sequence modeling toolkit that allows researchers and developers to train custom models for translation, summarization, language modeling and other text generation tasks.

We provide reference implementations of various sequence modeling papers:

List of implemented papers

What's New:

Previous updates

Features:

We also provide pre-trained models for translation and language modeling with a convenient torch.hub interface:

en2de = torch.hub.load('pytorch/fairseq', 'transformer.wmt19.en-de.single_model')
en2de.translate('Hello world', beam=5)
# 'Hallo Welt'

See the PyTorch Hub tutorials for translation and RoBERTa for more examples.

Requirements and Installation

  • PyTorch version >= 1.10.0
  • Python version >= 3.8
  • For training new models, you'll also need an NVIDIA GPU and NCCL
  • To install fairseq and develop locally:
git clone https://github.com/pytorch/fairseq
cd fairseq
pip install --editable ./

# on MacOS:
# CFLAGS="-stdlib=libc++" pip install --editable ./

# to install the latest stable release (0.10.x)
# pip install fairseq
  • For faster training install NVIDIA's apex library:
git clone https://github.com/NVIDIA/apex
cd apex
pip install -v --no-cache-dir --global-option="--cpp_ext" --global-option="--cuda_ext" \
  --global-option="--deprecated_fused_adam" --global-option="--xentropy" \
  --global-option="--fast_multihead_attn" ./
  • For large datasets install PyArrow: pip install pyarrow
  • If you use Docker make sure to increase the shared memory size either with --ipc=host or --shm-size as command line options to nvidia-docker run .

Getting Started

The full documentation contains instructions for getting started, training new models and extending fairseq with new model types and tasks.

Pre-trained models and examples

We provide pre-trained models and pre-processed, binarized test sets for several tasks listed below, as well as example training and evaluation commands.

We also have more detailed READMEs to reproduce results from specific papers:

Join the fairseq community

License

fairseq(-py) is MIT-licensed. The license applies to the pre-trained models as well.

Citation

Please cite as:

@inproceedings{ott2019fairseq,
  title = {fairseq: A Fast, Extensible Toolkit for Sequence Modeling},
  author = {Myle Ott and Sergey Edunov and Alexei Baevski and Angela Fan and Sam Gross and Nathan Ng and David Grangier and Michael Auli},
  booktitle = {Proceedings of NAACL-HLT 2019: Demonstrations},
  year = {2019},
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distributions

If you're not sure about the file name format, learn more about wheel file names.

fairseq_built-0.12.3-cp312-cp312-win_amd64.whl (11.7 MB view details)

Uploaded CPython 3.12Windows x86-64

fairseq_built-0.12.3-cp311-cp311-win_amd64.whl (11.7 MB view details)

Uploaded CPython 3.11Windows x86-64

fairseq_built-0.12.3-cp310-cp310-win_amd64.whl (11.7 MB view details)

Uploaded CPython 3.10Windows x86-64

File details

Details for the file fairseq_built-0.12.3-cp312-cp312-win_amd64.whl.

File metadata

File hashes

Hashes for fairseq_built-0.12.3-cp312-cp312-win_amd64.whl
Algorithm Hash digest
SHA256 2e847169a8f0ed1c416a77dd8c57108e74458a29efd622917d882210c0d998ef
MD5 f1607f8e8b00770e572018c03735eba0
BLAKE2b-256 2e29958c6db18c6533229e6026675912cdab007a1087f52cfde1b7001861a9cd

See more details on using hashes here.

File details

Details for the file fairseq_built-0.12.3-cp311-cp311-win_amd64.whl.

File metadata

File hashes

Hashes for fairseq_built-0.12.3-cp311-cp311-win_amd64.whl
Algorithm Hash digest
SHA256 e4903c5a9ed28599da6cfef5fa32ed95365b701d79e6acca8284563c5203912e
MD5 89721e7b0e7fb8f1907ec9194c26ee85
BLAKE2b-256 6941951e1fd83fad4d7bc920d2c3f64ec56e6ffecab958794c85f22ae2a20265

See more details on using hashes here.

File details

Details for the file fairseq_built-0.12.3-cp310-cp310-win_amd64.whl.

File metadata

File hashes

Hashes for fairseq_built-0.12.3-cp310-cp310-win_amd64.whl
Algorithm Hash digest
SHA256 76f99fb09d11be5877e39d842cab3287daf72e597c63e78714ad450a45a203c2
MD5 e5a3088857e760006cd553aa03553816
BLAKE2b-256 670424099666e43ae49bb03528953f01df8153d337f47c90a59728b2231742fd

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page