Skip to main content

convmodel provides a conversation model based on decoders models.

Project description


convmodel provides a conversation model based on transformers GPT-2 model :wink:

:sparkles: Features :sparkles:

  • Utilizes GPT2 model to generate response
  • Handles multi-turn conversation
  • Provides useuful interfaces to fine-tune model and generate a response from a given context

A simple example of fine-tune GPT-2 model and generate a response:

from convmodel import ConversationModel
from convmodel import ConversationExample

# Load model on GPU
model = ConversationModel.from_pretrained("gpt2")

# Define training/validation examples
train_iterator = [
        "Hi, how are you?",
        "Good, thank you, how about you?",
        "Good, thanks!"
        "I am hungry",
        "How about eating pizza?"
valid_iterator = [
        "Let's have a break!",
        "Nice idea!"

# Fine-tune model, valid_iterator=valid_iterator)

# Generate response
model.generate(context=["Hello", "How are you"], do_sample=True, top_p=0.95, top_k=50)
# Output could be like below if sufficient examples were given.
# => ConversationModelOutput(responses=['Good thank you'], context=['Hello', 'How are you'])

Please refer to document for more details of installation, model architecture and usage.

Enjoy talking with your conversational AI :wink:

Project details

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

convmodel-0.3.0.tar.gz (9.9 kB view hashes)

Uploaded Source

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page