Skip to main content

An open-source NLP research library, built on PyTorch.

Project description

Build Status codecov docker

An Apache 2.0 NLP research library, built on PyTorch, for developing state-of-the-art deep learning models on a wide variety of linguistic tasks.

Installation

The preferred way to install AllenNLP is via pip. Just run pip install allennlp in your Python 3.6 environment and you're good to go!

If you need pointers on setting up a Python 3.6 environment or would like to install AllenNLP using a different method, see below.

Installing via pip

Setting up a virtual environment

Conda can be used set up a virtual environment with the version of Python required for AllenNLP. If you already have a Python 3.6 environment you want to use, you can skip to the 'installing via pip' section.

  1. Download and install Conda.

  2. Create a Conda environment with Python 3.6

    conda create -n allennlp python=3.6
    
  3. Activate the Conda environment. You will need to activate the Conda environment in each terminal in which you want to use AllenNLP.

    source activate allennlp
    

Installing the library and dependencies

Installing the library and dependencies is simple using pip.

pip install allennlp

That's it! You're now ready to build and train AllenNLP models. AllenNLP installs a script when you install the python package, meaning you can run allennlp commands just by typing allennlp into a terminal.

pip currently installs Pytorch for CUDA 8 only (or no GPU). If you require a newer version, please visit http://pytorch.org/ and install the relevant pytorch binary.

Installing using Docker

Docker provides a virtual machine with everything set up to run AllenNLP-- whether you will leverage a GPU or just run on a CPU. Docker provides more isolation and consistency, and also makes it easy to distribute your environment to a compute cluster.

Once you have installed Docker just run docker run -it -p 8000:8000 --rm allennlp/allennlp:v0.5.1 to get an environment that will run on either the cpu or gpu.

You can now test your installation with ./scripts/verify.py.

Our Docker image contains the AllenNLP source rather than a pip installation. Consequently, the allennlp commandline tool is not installed and you will have to use ./bin/allennlp instead.

Installing from source

You can also install AllenNLP by cloning our git repository:

git clone https://github.com/allenai/allennlp.git

Create a Python 3.6 virtual environment, and install the necessary requirements by running:

INSTALL_TEST_REQUIREMENTS=true scripts/install_requirements.sh

Changing the flag to false if you don't want to be able to run tests.

Note that if you use the source installation, you won't be able to use the allennlp command but rather you'll need to run ./bin/allennlp.

You can test your installation with ./scripts/verify.py.

Running AllenNLP

Once you've installed AllenNLP, you can run the command-line interface either with the allennlp command (if you installed via pip) or python -m allennlp.run (if you installed via source).

$ allennlp
Run AllenNLP

optional arguments:
  -h, --help    show this help message and exit

Commands:

    train       Train a model
    evaluate    Evaluate the specified model + dataset
    predict     Use a trained model to make predictions.
    make-vocab  Create a vocabulary
    elmo        Use a trained model to make predictions.
    fine-tune   Continue training a model on a new dataset
    dry-run     Create a vocabulary, compute dataset statistics and other
                training utilities.
    test-install
                Run the unit tests.

What is AllenNLP?

Built on PyTorch, AllenNLP makes it easy to design and evaluate new deep learning models for nearly any NLP problem, along with the infrastructure to easily run them in the cloud or on your laptop. AllenNLP was designed with the following principles:

  • Hyper-modular and lightweight. Use the parts which you like seamlessly with PyTorch.
  • Extensively tested and easy to extend. Test coverage is above 90% and the example models provide a template for contributions.
  • Take padding and masking seriously, making it easy to implement correct models without the pain.
  • Experiment friendly. Run reproducible experiments from a json specification with comprehensive logging.

AllenNLP includes reference implementations of high quality models for Semantic Role Labelling, Question and Answering (BiDAF), Entailment (decomposable attention), and more.

AllenNLP is built and maintained by the Allen Institute for Artificial Intelligence, in close collaboration with researchers at the University of Washington and elsewhere. With a dedicated team of best-in-field researchers and software engineers, the AllenNLP project is uniquely positioned to provide state of the art models with high quality engineering.

allennlp an open-source NLP research library, built on PyTorch
allennlp.commands functionality for a CLI and web service
allennlp.data a data processing module for loading datasets and encoding strings as integers for representation in matrices
allennlp.models a collection of state-of-the-art models
allennlp.modules a collection of PyTorch modules for use with text
allennlp.nn tensor utility functions, such as initializers and activation functions
allennlp.service a web server to serve our demo and API
allennlp.training functionality for training models

Docker images

AllenNLP releases Docker images to Docker Cloud for each release. For information on how to run these releases, see Installing using Docker.

Building a Docker image

For various reasons you may need to create your own AllenNLP Docker image. The same image can be used either with a CPU or a GPU.

First, follow the instructions above for setting up a development environment. Then run the following command (it will take some time, as it completely builds the environment needed to run AllenNLP.)

docker build --tag allennlp/allennlp .

You should now be able to see this image listed by running docker images allennlp.

REPOSITORY          TAG                 IMAGE ID            CREATED             SIZE
allennlp/allennlp            latest              b66aee6cb593        5 minutes ago       2.38GB

Running the Docker image

You can run the image with docker run --rm -it allennlp/allennlp. The --rm flag cleans up the image on exit and the -it flags make the session interactive so you can use the bash shell the Docker image starts.

You can test your installation by running ./scripts/verify.py.

Citing

If you use AllenNLP in your research, please cite AllenNLP: A Deep Semantic Natural Language Processing Platform.

@inproceedings{Gardner2017AllenNLP,
  title={AllenNLP: A Deep Semantic Natural Language Processing Platform},
  author={Matt Gardner and Joel Grus and Mark Neumann and Oyvind Tafjord
    and Pradeep Dasigi and Nelson F. Liu and Matthew Peters and
    Michael Schmitz and Luke S. Zettlemoyer},
  year={2017},
  Eprint = {arXiv:1803.07640},
}

Team

AllenNLP is an open-source project backed by the Allen Institute for Artificial Intelligence (AI2). AI2 is a non-profit institute with the mission to contribute to humanity through high-impact AI research and engineering. To learn more about who specifically contributed to this codebase, see our contributors page.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

allennlp-0.5.1.tar.gz (5.6 MB view details)

Uploaded Source

Built Distribution

allennlp-0.5.1-py3-none-any.whl (3.9 MB view details)

Uploaded Python 3

File details

Details for the file allennlp-0.5.1.tar.gz.

File metadata

  • Download URL: allennlp-0.5.1.tar.gz
  • Upload date:
  • Size: 5.6 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No

File hashes

Hashes for allennlp-0.5.1.tar.gz
Algorithm Hash digest
SHA256 c6c3d47fabf8b1a4ba93c09e833b56fadcf706c343f807a6e7b41ba93637b18e
MD5 580764afed213bbfb77500ae4c2a5482
BLAKE2b-256 24f03e3df9f06d534270b5da0de31ba9cf742f7e2a5fc0d2d70afb069e44308d

See more details on using hashes here.

File details

Details for the file allennlp-0.5.1-py3-none-any.whl.

File metadata

File hashes

Hashes for allennlp-0.5.1-py3-none-any.whl
Algorithm Hash digest
SHA256 bbf3ac1f97d882e32c55057b3b37852ff44d5a87842673c2421c449e5887a724
MD5 327ff42af9680f4dc819c74f9d871875
BLAKE2b-256 8d341439e851e8ebd913f958b40c48bc75451d37f337105703fa04dd855ae3f5

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page