Skip to main content

This is a modified version based on the original openicl 0.1.8 for certain research usages.

Project description


OverviewInstallationPaperExamplesDocsCitation

version

Overview

OpenICL provides an easy interface for in-context learning, with many state-of-the-art retrieval and inference methods built in to facilitate systematic comparison of LMs and fast research prototyping. Users can easily incorporate different retrieval and inference methods, as well as different prompt instructions into their workflow.

What's News

  • v0.1.8 Support LLaMA and self-consistency

Installation

Note: OpenICL requires Python 3.8+

Using Pip

pip install openicl

Installation for local development:

git clone https://github.com/Shark-NLP/OpenICL
cd OpenICL
pip install -e .

Quick Start

Following example shows you how to perform ICL on sentiment classification dataset. More examples and tutorials can be found at examples

Step 1: Load and prepare data

from datasets import load_dataset
from openicl import DatasetReader

# Loading dataset from huggingface
dataset = load_dataset('gpt3mix/sst2')

# Define a DatasetReader, with specified column names where input and output are stored.
data = DatasetReader(dataset, input_columns=['text'], output_column='label')

Step 2: Define the prompt template (Optional)

from openicl import PromptTemplate
tp_dict = {
    0: "</E>Positive Movie Review: </text>",
    1: "</E>Negative Movie Review: </text>" 
}

template = PromptTemplate(tp_dict, {'text': '</text>'}, ice_token='</E>')

The placeholder </E> and </text> will be replaced by in-context examples and testing input, respectively. For more detailed information about PromptTemplate (such as string-type template) , please see tutorial1.

Step 3: Initialize the Retriever

from openicl import TopkRetriever
# Define a retriever using the previous `DataLoader`.
# `ice_num` stands for the number of data in in-context examples.
retriever = TopkRetriever(data, ice_num=8)

Here we use the popular TopK method to build the retriever.

Step 4: Initialize the Inferencer

from openicl import PPLInferencer
inferencer = PPLInferencer(model_name='distilgpt2')

Step 5: Inference and scoring

from openicl import AccEvaluator
# the inferencer requires retriever to collect in-context examples, as well as a template to wrap up these examples.
predictions = inferencer.inference(retriever, ice_template=template)
# compute accuracy for the prediction
score = AccEvaluator().score(predictions=predictions, references=data.references)
print(score)

Docs

(updating...)

OpenICL Documentation

Citation

If you find this repository helpful, feel free to cite our paper:

@article{wu2023openicl,
  title={OpenICL: An Open-Source Framework for In-context Learning},
  author={Zhenyu Wu, Yaoxiang Wang, Jiacheng Ye, Jiangtao Feng, Jingjing Xu, Yu Qiao, Zhiyong Wu},
  journal={arXiv preprint arXiv:2303.02913},
  year={2023}
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

openicl_nolabel-0.0.2.tar.gz (26.8 kB view details)

Uploaded Source

Built Distribution

openicl_nolabel-0.0.2-py3-none-any.whl (43.1 kB view details)

Uploaded Python 3

File details

Details for the file openicl_nolabel-0.0.2.tar.gz.

File metadata

  • Download URL: openicl_nolabel-0.0.2.tar.gz
  • Upload date:
  • Size: 26.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.9.18

File hashes

Hashes for openicl_nolabel-0.0.2.tar.gz
Algorithm Hash digest
SHA256 12a6793c2f2232f3339275d2ad9dc3af882eafb6c8f2c63d3a41f423ee222398
MD5 dc146662898d97bb4fbb448c54a7e436
BLAKE2b-256 18f5e7295be82104a5c091a9f07876d45ccef4ebccc4da883207799d825a9898

See more details on using hashes here.

File details

Details for the file openicl_nolabel-0.0.2-py3-none-any.whl.

File metadata

File hashes

Hashes for openicl_nolabel-0.0.2-py3-none-any.whl
Algorithm Hash digest
SHA256 ce9ca36960dfeeb02128e98b0573656097552b4f13e3f9f7d3c0a04e5065cf85
MD5 d71c4a92b19c5c37530e99d9fee37a5b
BLAKE2b-256 bd47d772b600653ace882b48dfd878682b8369689ec294134d2bf8ada1e523cc

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page