Skip to main content

Neural Question Answering at Scale. Use modern transformer based models like BERT to find answers in large document collections

None None

Project description

Haystack Logo Build Checked with mypy Release License Last Commit

Introduction

The performance of modern Question Answering Models (BERT, ALBERT …) has seen drastic improvements within the last year enabling many new opportunities for accessing information more efficiently. However, those models are designed to find answers within rather small text passages. Haystack lets you scale QA models to large collections of documents! While QA is the focussed use case for Haystack, we will address further options around neural search in the future (re-ranking, most-similar search …).

Haystack is designed in a modular way and lets you use any models trained with FARM or Transformers.

Core Features

  • Powerful ML models: Utilize all latest transformer based models (BERT, ALBERT, RoBERTa …)

  • Modular & future-proof: Easily switch to newer models once they get published.

  • Developer friendly: Easy to debug, extend and modify.

  • Scalable: Production-ready deployments via Elasticsearch backend & REST API

  • Customizable: Fine-tune models to your own domain & improve them continuously via user feedback

Components

https://raw.githubusercontent.com/deepset-ai/haystack/master/docs/_src/img/sketched_concepts_white.png
  1. DocumentStore: Database storing the documents for our search. We recommend Elasticsearch, but have also more light-weight options for fast prototyping (SQL or In-Memory).

  2. Retriever: Fast, simple algorithm that identifies candidate passages from a large collection of documents. Algorithms include TF-IDF or BM25, custom Elasticsearch queries, and embedding-based approaches. The Retriever helps to narrow down the scope for Reader to smaller units of text where a given question could be answered.

  3. Reader: Powerful neural model that reads through texts in detail to find an answer. Use diverse models like BERT, RoBERTa or XLNet trained via FARM or Transformers on SQuAD like tasks. The Reader takes multiple passages of text as input and returns top-n answers with corresponding confidence scores. You can just load a pretrained model from Hugging Face’s model hub or fine-tune it to your own domain data.

  4. Finder: Glues together a Reader and a Retriever as a pipeline to provide an easy-to-use question answering interface.

  5. REST API: Exposes a simple API for running QA search, collecting feedback and monitoring requests

  6. Haystack Annotate: Create custom QA labels, Hosted version (Beta), Docker images (coming soon)

Resources

Documentation: https://haystack.deepset.ai

Tutorials

Quick Start

Installation

PyPi:

pip install farm-haystack

Master branch (if you wanna try the latest features):

git clone https://github.com/deepset-ai/haystack.git
cd haystack
pip install --editable .

To update your installation, just do a git pull. The –editable flag will update changes immediately.

Note: On Windows you might need pip install farm-haystack -f https://download.pytorch.org/whl/torch_stable.html to install PyTorch correctly

Usage

https://raw.githubusercontent.com/deepset-ai/haystack/master/docs/_src/img/code_snippet_usage.png

Quick Tour

1) DocumentStores

Haystack offers different options for storing your documents for search. We recommend Elasticsearch, but have also light-weight options for fast prototyping and will soon add DocumentStores that are optimized for embeddings (FAISS & Co).

SQL / InMemory (Alternative)

haystack.database.sql.SQLDocumentStore & haystack.database.memory.InMemoryDocumentStore

These DocumentStores are mainly intended to simplify the first development steps or test a prototype on an existing SQL Database containing your texts. The SQLDocumentStore initializes by default a local file-based SQLite database. However, you can easily configure it for PostgreSQL or MySQL since our implementation is based on SQLAlchemy. Limitations: Retrieval (e.g. via TfidfRetriever) happens in-memory here and will therefore only work efficiently on small datasets

2) Retrievers

DensePassageRetriever

Using dense embeddings (i.e. vector representations) of texts is a powerful alternative to score similarity of texts. This retriever uses two BERT models - one to embed your query, one to embed your passage. It’s based on the work of Karpukhin et al and is especially an powerful alternative if there’s no direct overlap between tokens in your queries and your texts.

Example

retriever = DensePassageRetriever(document_store=document_store,
                                  embedding_model="dpr-bert-base-nq",
                                  do_lower_case=True, use_gpu=True)
retriever.retrieve(query="Why did the revenue increase?")
# returns: [Document, Document]

ElasticsearchRetriever

Scoring text similarity via sparse Bag-of-words representations are strong and well-established baselines in Information Retrieval. The default ElasticsearchRetriever uses Elasticsearch’s native scoring (BM25), but can be extended easily with custom queries or filtering.

Example

retriever = ElasticsearchRetriever(document_store=document_store, custom_query=None)
retriever.retrieve(query="Why did the revenue increase?", filters={"years": ["2019"], "company": ["Q1", "Q2"]})
# returns: [Document, Document]

EmbeddingRetriever

This retriever uses a single model to embed your query and passage (e.g. Sentence-BERT) and finds similar texts by using cosine similarity. This works well if your query and passage are a similar type of text, e.g. you want to find the most similar question in your FAQ given a user question.

Example

retriever = EmbeddingRetriever(document_store=document_store,
                               embedding_model="deepset/sentence_bert",
                               model_format="farm")
retriever.retrieve(query="Why did the revenue increase?", filters={"years": ["2019"], "company": ["Q1", "Q2"]})
# returns: [Document, Document]

TfidfRetriever

Basic in-memory retriever getting texts from the DocumentStore, creating TF-IDF representations in-memory and allowing to query them. Simple baseline for quick prototypes. Not recommended for production.

3) Readers

Neural networks (i.e. mostly Transformer-based) that read through texts in detail to find an answer. Use diverse models like BERT, RoBERTa or XLNet trained via FARM or on SQuAD-like datasets. The Reader takes multiple passages of text as input and returns top-n answers with corresponding confidence scores. Both readers can load either a local model or any public model from Hugging Face’s model hub

FARMReader

Implementing various QA models via the FARM Framework.

Example

reader = FARMReader(model_name_or_path="deepset/roberta-base-squad2",
                use_gpu=False, no_ans_boost=-10, context_window_size=500,
                top_k_per_candidate=3, top_k_per_sample=1,
                num_processes=8, max_seq_len=256, doc_stride=128)

# Optional: Training & eval
reader.train(...)
reader.eval(...)

# Predict
reader.predict(question="Who is the father of Arya Starck?", documents=documents, top_k=3)

This Reader comes with:

  • extensive configuration options (no answer boost, aggregation options …)

  • multiprocessing to speed-up preprocessing

  • option to train

  • option to evaluate

  • option to load all QA models directly from HuggingFace’s model hub

TransformersReader

Implementing various QA models via the pipeline class of Transformers Framework.

Example

reader = TransformersReader(model="distilbert-base-uncased-distilled-squad",
                            tokenizer="distilbert-base-uncased",
                            context_window_size=500,
                            use_gpu=-1)

reader.predict(question="Who is the father of Arya Starck?", documents=documents, top_k=3)

5. REST API

A simple REST API based on FastAPI is provided to:

  • search answers in texts (extractive QA)

  • search answers by comparing user question to existing questions (FAQ-style QA)

  • collect & export user feedback on answers to gain domain-specific training data (feedback)

  • allow basic monitoring of requests (currently via APM in Kibana)

To serve the API, adjust the values in rest_api/config.py and run:

gunicorn rest_api.application:app -b 0.0.0.0:8000 -k uvicorn.workers.UvicornWorker -t 300

You will find the Swagger API documentation at http://127.0.0.1:8000/docs

6. Labeling Tool

  • Use the hosted version (Beta) or deploy it yourself via Docker images (coming soon)

  • Create labels with different techniques: Come up with questions (+ answers) while reading passages (SQuAD style) or have a set of predefined questions and look for answers in the document (~ Natural Questions).

  • Structure your work via organizations, projects, users

  • Upload your documents or import labels from an existing SQuAD-style dataset

  • Coming soon: more file formats for document upload, metrics for label quality …

https://raw.githubusercontent.com/deepset-ai/haystack/master/docs/_src/img/annotation_tool.png

7. Indexing PDF / Docx files

Haystack has basic converters to extract text from PDF and Docx files. While it’s almost impossible to cover all types, layouts and special cases in PDFs, the implementation covers the most common formats and provides basic cleaning functions to remove header, footers, and tables. Multi-Column text layouts are also supported. The converters are easily extendable, so that you can customize them for your files if needed.

Example:

#PDF
from haystack.indexing.file_converters.pdf import PDFToTextConverter
converter = PDFToTextConverter(remove_header_footer=True, remove_numeric_tables=True, valid_languages=["de","en"])
pages = converter.extract_pages(file_path=file)
# => list of str, one per page
#DOCX
from haystack.indexing.file_converters.docx import DocxToTextConverter
converter = DocxToTextConverter()
paragraphs = converter.extract_pages(file_path=file)
#  => list of str, one per paragraph (as docx has no direct notion of pages)

Advanced document convertion is enabled by leveraging mature text extraction library Apache Tika, which is mostly written in Java. Although it’s possible to call Tika API from Python, the current TikaConverter only supports RESTful call to a Tika server running at localhost. One may either run Tika as a REST service at port 9998 (default), or to start a docker container for Tika. The latter is recommended, as it’s easily scalable, and does not require setting up any Java runtime environment. What’s more, future update is also taken care of by docker. Either way, TikaConverter makes RESTful calls to convert any document format supported by Tika. Example code can be found at indexing/file_converters/utils.py’s tika_convert)_files_to_dicts function:

TikaConverter supports 341 file formats, including

  • most common text file formats, e.g. HTML, XML, Microsoft Office OLE2/XML/OOXML, OpenOffice ODF, iWorks, PDF, ePub, RTF, TXT, RSS, CHM…

  • text embedded in media files, e.g. WAV, MP3, Vorbis, Flac, PNG, GIF, JPG, BMP, TIF, PSD, WebP, WMF, EMF, MP4, Quicktime, 3GPP, Ogg, FLV…

  • mail and database files, e.g. Unitx mailboxes, Outlook PST/MSG/TNEF, SQLite3, Microsoft Access, dBase…

  • and many more other formats…

  • and all those file formats in archive files, e.g. TAR, ZIP, BZip2, GZip 7Zip, RAR!

Check out complete list of files supported by the most recent Apache Tika 1.24.1. If you feel adventurous, Tika even supports some image OCR with Tesseract, or object recognition for image and video files. (not implemented yet)

TikaConverter also makes a document’s metadata available, including typical fields like file name, file dates and a lot more (e.g. Author and keywords for PDF if they’re available in the files), which may save you some time in data labeling or other downstream tasks.

converter = TikaConverter(remove_header_footer=True)
pages = converter.extract_pages(file_path=path)
pages, meta = converter.extract_pages(file_path=path, return_meta=True)

Contributing

We are very open to contributions from the community - be it the fix of a small typo or a completely new feature! You don’t need to be an Haystack expert for providing meaningful improvements. To avoid any extra work on either side, please check our Contributor Guidelines first.

Tests will automatically run for every commit you push to your PR. You can also run them locally by executing pytest in your terminal from the root folder of this repository:

pytest test/

Project details

None None

Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

farm-haystack-0.4.0.tar.gz (71.4 kB view hashes)

Uploaded Source

Built Distribution

farm_haystack-0.4.0-py3-none-any.whl (81.0 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page