Skip to main content

RankGen is a suite of encoder models (100M-1.2B parameters) which map prefixes and generations from any pretrained English language model to a shared vector space. RankGen can be used to rerank multiple full-length samples from an LM, and it can also be incorporated as a scoring function into beam search to significantly improve generation quality (0.85 vs 0.77 MAUVE, 75% preference according to humans annotators who are English writers).

Project description

RankGen - Improving Text Generation with Large Ranking Models

made-with-python arxiv PyPI version rankgen License: Apache 2.0

This is the official repository for our preprint, RankGen - Improving Text Generation with Large Ranking Models. RankGen is a 1.2 billion encoder model which maps prefixes and generations from any pretrained English language model to a shared vector space. RankGen can be used to rerank multiple full-length samples from an LM, and it can also be incorporated as a scoring function into beam search to significantly improve generation quality (0.85 vs 0.77 MAUVE, 75% preference according to humans annotators who are English writers). RankGen can also be used like a dense retriever, and achieves state-of-the-art performance on literary retrieval.

This repository contains human evaluation data, links to HuggingFace-compatible model checkpoints, and code to integrate RankGen in beam search on HuggingFace models. RankGen is trained by fine-tuning the T5-XL encoder using the T5X library.

Updates

  • (July 2022) RankGen is now a PyPI package, just run pip install rankgen to use it!
  • (July 2022) RankGen checkpoints are now available on the HuggingFace Model Hub (link)!

Model checkpoints

All RankGen checkpoints are available on the HuggingFace Model Hub - link

We recommend using RankGen-XL-all.

Checkpoint Size Hub Model Name HF Hub Link
RankGen-base-all 0.1B kalpeshk2011/rankgen-t5-base-all link
RankGen-large-all 0.3B kalpeshk2011/rankgen-t5-large-all link
RankGen-XL-all 1.2B kalpeshk2011/rankgen-t5-xl-all link
RankGen-XL-PG19 1.2B kalpeshk2011/rankgen-t5-xl-pg19 link

Older versions of the checkpoints:

RankGen XL checkpoints compatible with T5XEmbeddingGeneratorLegacy - here

T5X JAX checkpoints (base, large, XL) - here

Setup

Requirements

Python 3.7+, torch (CUDA recommended), transformers

Installation

(from PyPI)

python3.7 -m virtualenv rankgen-venv
source rankgen-venv/bin/activate
pip install rankgen

(from source)

python3.7 -m virtualenv rankgen-venv
source rankgen-venv/bin/activate
git clone https://github.com/martiansideofthemoon/rankgen
cd rankgen
pip install --editable .

Data Download / Test

Get the data here and place folder in root directory. Alternatively, use gdown as shown below,

gdown --folder https://drive.google.com/drive/folders/1DRG2ess7fK3apfB-6KoHb_azMuHbsIv4

Run the test script to make sure the RankGen checkpoint has loaded correctly,

python -m rankgen.test_rankgen_encoder --model_path kalpeshk2011/rankgen-t5-base-all

### Expected output
0.0009239262409127233
0.0011521980725477804

Using RankGen

Loading RankGen is simple using the HuggingFace APIs, but we suggest using RankGenEncoder, which is a small wrapper around the HuggingFace APIs for correctly preprocessing data and doing tokenization automatically. Please see rankgen/test_rankgen_encoder.py for an example of the usage or see below.

from rankgen import RankGenEncoder, RankGenGenerator

rankgen_encoder = RankGenEncoder("kalpeshk2011/rankgen-t5-xl-all")

Encoding text to prefix/suffix vectors

prefix_vectors = rankgen_encoder.encode(["This is a prefix sentence."], vectors_type="prefix")
suffix_vectors = rankgen_encoder.encode(["This is a suffix sentence."], vectors_type="suffix")

Generating text

# use a HuggingFace compatible language model
generator = RankGenGenerator(rankgen_encoder=rankgen_encoder, language_model="gpt2-medium")

inputs = ["Whatever might be the nature of the tragedy it would be over with long before this, and those moving black spots away yonder to the west, that he had discerned from the bluff, were undoubtedly the departing raiders. There was nothing left for Keith to do except determine the fate of the unfortunates, and give their bodies decent burial. That any had escaped, or yet lived, was altogether unlikely, unless, perchance, women had been in the party, in which case they would have been borne away prisoners."]

# Baseline nucleus sampling
print(generator.generate_single(inputs, top_p=0.9)[0][0])
# Over-generate and re-rank
print(generator.overgenerate_rerank(inputs, top_p=0.9, num_samples=10)[0][0])
# Beam search
print(generator.beam_search(inputs, top_p=0.9, num_samples=10, beam_size=2)[0][0])

Running beam search with RankGen (reproducing experiments in the paper)

The main file is rankgen/rankgen_beam_search.py. To execute it,

python rankgen/rankgen_beam_search.py \
    --dataset rankgen_data/wiki.jsonl \
    --rankgen_encoder kalpeshk2011/rankgen-t5-xl-all \
    --num_tokens 20 --num_samples 10 --beam_size 2 \
    --output_file outputs_beam/wiki_t5_xl_beam_2_tokens_20_samples_10.jsonl

Evaluating using MAUVE (make sure JSONL file has several thousand generations for intuitive MAUVE scores, 7713 in our experiments),

python rankgen/score_multi_beam.py --dataset outputs_beam/wiki_t5_xl_beam_2_tokens_10_samples_10.jsonl

Human evaluation data

We conducted our human evaluation on Upwork, hiring English teachers and writers. We performed blind A/B testing between RankGen and nucleus sampling. We also asked our annotators to provide a 1-3 sentence explanation. You can find all the 600 annotations across two files in human-eval-data. To compute the evaluation scores run,

python rankgen/score_ab_text.py

Citation Information

If you use RankGen, please cite it as follows:

@article{krishna2022rankgen,
  title={RankGen: Improving Text Generation with Large Ranking Models},
  author={Kalpesh Krishna and Yapei Chang and John Wieting and Mohit Iyyer},
  journal={arXiv preprint arXiv:2205.09726},
  year={2022}
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

rankgen-0.1.1.tar.gz (37.0 kB view details)

Uploaded Source

Built Distribution

rankgen-0.1.1-py3-none-any.whl (50.6 kB view details)

Uploaded Python 3

File details

Details for the file rankgen-0.1.1.tar.gz.

File metadata

  • Download URL: rankgen-0.1.1.tar.gz
  • Upload date:
  • Size: 37.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.1.14 CPython/3.7.5 Linux/5.4.0-80-generic

File hashes

Hashes for rankgen-0.1.1.tar.gz
Algorithm Hash digest
SHA256 4a8a3c3c5ff545b9b17169167a30cbcdca8c5e15b172f167ec3da31057f95dcb
MD5 5746ce033e3f6435a74e0fc436a5c204
BLAKE2b-256 0551e317fa3ff34940275c58f36c0f19a6d759b83565a5153a6a361d640440c0

See more details on using hashes here.

File details

Details for the file rankgen-0.1.1-py3-none-any.whl.

File metadata

  • Download URL: rankgen-0.1.1-py3-none-any.whl
  • Upload date:
  • Size: 50.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.1.14 CPython/3.7.5 Linux/5.4.0-80-generic

File hashes

Hashes for rankgen-0.1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 5e038a2d2245c05f4cf52e3533dd930a1b629f12f19d1f144f6edbcfbb98683a
MD5 7e5f799a6986f49ec48e6c30efeaf7c1
BLAKE2b-256 424ed881a0415e7ede9c6ab607879859099b30150ee9d6a792bc164d84fa8ca8

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page