Yandex GPT Support for LangChain
Project description
yandex-chain - LangChain-compatible integrations with YandexGPT and YandexGPT Embeddings
This library is community-maintained Python package that provides support for Yandex GPT LLM and Embeddings for LangChain Framework.
Currently, Yandex GPT is in preview stage, so this library may occasionally break. Please use it at your own risk!
What's Included
The library includes the following main classes:
- YandexLLM is a class representing YandexGPT Text Generation.
- ChatYandexGPT exposes the same model in chat interface that expects messages as input.
- YandexEmbeddings represents YandexGPT Embeddings service.
- YandexGPTClassifier that supports zero-shot and few-shot classification.
Usage
You can use YandexLLM
in the following manner:
from yandex_chain import YandexLLM
LLM = YandexLLM(folder_id="...", api_key="...")
print(LLM("How are you today?"))
You can use YandexEmbeddings
to compute embedding vectors:
from yandex_chain import YandexEmbeddings
embeddings = YandexEmbeddings(...)
print(embeddings("How are you today?"))
Use ChatYandexGPT
to execute a dialog with the model:
from yandex_chain import YandexLLM
gpt = ChatYandexGPT(...)
print(gpt(
[
HumanMessage(content='Привет! Придумай 10 новых слов для приветствия.')
]))
Authentication
In order to use Yandex GPT, you need to provide one of the following authentication methods, which you can specify as parameters to YandexLLM
, ChatYandexGPT
and YandexEmbeddings
classes:
- A pair of
folder_id
andapi_key
- A pair of
folder_id
andiam_token
- A path to
config.json
file, which may in turn contain parameters listed above in a convenient JSON format.
Complete Example
A pair of LLM and Embeddings are a good combination to create problem-oriented chatbots using Retrieval-Augmented Generation (RAG). Here is a short example of this approach, inspired by this LangChain tutorial.
To begin with, we have a set of documents docs
(for simplicity, let's assume it is just a list of strings), which we store in vector storage. We can use YandexEmbeddings
to compute embedding vectors:
from yandex_chain import YandexLLM, YandexEmbeddings
from langchain.vectorstores import FAISS
embeddings = YandexEmbeddings(config="config.json")
vectorstore = FAISS.from_texts(docs, embedding=embeddings)
retriever = vectorstore.as_retriever()
We can now retrieve a set of documents relevant to a query:
query = "Which library can be used to work with Yandex GPT?"
res = retriever.get_relevant_documents(query)
Now, to provide a full-text answer to the query, we can use LLM. We will prompt the LLM, giving it retrieved documents as a context, and the input query, and ask it to answer the question. This can be done using LangChain chains:
from operator import itemgetter
from langchain.prompts import ChatPromptTemplate
from langchain.schema.output_parser import StrOutputParser
from langchain.schema.runnable import RunnablePassthrough
template = """Answer the question based only on the following context:
{context}
Question: {question}
"""
prompt = ChatPromptTemplate.from_template(template)
model = YandexLLM(config="config.json")
chain = (
{"context": retriever, "question": RunnablePassthrough()}
| prompt
| model
| StrOutputParser()
)
This chain can now answer our questions:
chain.invoke(query)
Lite vs. Full Models
YandexGPT model comes in several flavours - YandexGPT Lite (current and RC), YandexGPT Pro and Summarization model. By default, YandexGPT Lite is used. If you want to use different model, please specify it in the constructor of YandexLLM
or ChatYandexGPT
language model classes:
- Pro (based on Yandex GPT 3):
model=YandexGPTModel.Pro
- Lite (based on Yandex GPT 3):
model=YandexGPTModel.Lite
- Pro RC (based on Yandex GPT 4):
model=YandexGPTModel.ProRC
- Lite RC (based on Yandex GPT 4):
model=YandexGPTModel.LiteRC
- Pro 32k (based on Yandex GPT 4):
model=YandexGPTModel.Pro32k
- Summarization (based on Yandex GPT 2):
model=YandexGPTModel.Summarization
In previous versions, we were using
use_lite
flag to switch between Lite and Pro models. This behavior is still supported, but is deprecated, and will be removed in the next version.
Async Operations
The library supports explicit async mode of Yandex GPT API. Provided model
is YandexLLM
model,
you can call model.invokeAsync(...)
to obtain the id
of the async operation. You can then call model.checkAsyncResult(id)
to check if the result if ready. checkAsyncResult
returns None
when the result is not ready, otherwise it returns the result of the operation (string, or Message if return_message
argument is True
).
Testing
This repository contains some basic unit tests. To run them, you need to place a configuration file config.json
with your credentials into tests
folder. Use config_sample.json
as a reference. After that, please run the following at the repository root directory:
python -m unittest discover -s tests
Credits
- This library has originally been developed by Dmitri Soshnikov.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
File details
Details for the file yandex-chain-0.0.10.tar.gz
.
File metadata
- Download URL: yandex-chain-0.0.10.tar.gz
- Upload date:
- Size: 10.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.10.13
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 2215738d2d255379d65337c0f21f60d5522c2608d037c23f949b7ffb4bbcd630 |
|
MD5 | 0b48ff2bd45c398cfc9acaa38f4a2015 |
|
BLAKE2b-256 | 01e25217740f8f3f055675f025008f2ad6da8aa59b19ab203f432bb6652178b4 |