aBLT Python API
Project description
aBLT Python API wrapper
About
This is Python aBLT API wrapper. It is used to communicate with aBLT API. You may use it to create your own aBLT client using asynchronous or synchronous Python. You can find more information about aBLT API here.
At first, you need obtain aBLT API Token. You can do it here by get in touch and contact with support for paid plans. Next, you may use any existing ready-made bots or templates or create your own bot via UI interface.
Installation
API wrapper is available on PyPI. You can install it with pip (recommended to use Python 3.9+):
pip install ablt-python-api
Usage
Then you can import it and use it:
# for asynchronous API wrapper use ABLTApi_async
from ablt_python_api import ABLTApi # this is synchronous API wrapper
# Init with explicit token
api = ABLTApi(bearer_token=YOUR_ABLT_API_TOKEN)
# Init with environment variable, use ABLT_BEARER_TOKEN
api = ABLTApi()
For some reason you may want to use your own logger, then you can initialize API wrapper with logger:
# logger it's pre-configured instance of logging.logger
api = ABLTApi(logger=your_logger)
API methods
Bots
List of bots
You may get list of bots:
# will return list of bots, see schema docs: https://docs.ablt.ai/api_docs/bots/get#Response-body
bots = api.api.get_bots()
Or you may use actual schema:
from ablt_python_api.schemas import BotsSchema
# Use await for asynchronous API wrapper
bots = [BotsSchema.model_validate(bot_dict) for bot_dict in api.get_bots()]
Single bot
You may get bot by UID:
bot = api.find_bot_by_uid(bot_uid='F0b98A09-c5ed-1197-90F3-BBfF1DBb28ee')
Alternatively, you may get bot by slug:
bot = api.find_bot_by_slug(bot_slug='omni')
Or by name (but it's not recommended, because name may be not unique):
bot = api.find_bot_by_name(bot_name='Miles Hiker')
In case if no bot found, then None
will be returned.
Chat
To chat with bot you may use `chat' method:
# Specify bot UID, will return generator with bot responses.
response = api.chat(bot_uid=BOT_UID, prompt='Hello, bot!')
# Or specify bot slug, return generator with bot responses.
response = api.chat(bot_slug=BOT_SLUG, prompt='Hello, bot!')
# To get response as string, you may use loop or extract just first response from generator
str_response = response.__next__()
# Or use __anext__ for async mode
Most probably, you will use messages
list instead of prompt
to save context. In this case, you may call method like
following:
messages = [
{"content": "I like to eat pizza", "role": "user"},
{"content": "Hello, I like pizza too!", "role": "assistant"},
{"content": "What do I like to eat?", "role": "user"},
]
# Will return generator with bot response
response = api.chat(bot_uid=BOT_UID, messages=messages)
You need ensure, that your prompt is last message in messages
list.
More options
Additionally, you may extend \ override system context with passing system
instruction to bot:
# Use with caution, it may conflict or replace all settings stored in UI
# and may lead to unexpected results
messages = [
{"content": "You are a bibliophile bot, you know everything about literature", "role": "system"},
{"content": "Who is author of 'The Sirens of Titan'?", "role": "user"},
],
You may call chat
method with params to override system settings, if you want:
response = api.chat(bot_uid=BOT_UID,
prompt='Hello, bot!',
language='Arabic', # may be: "Arabic", "French", "English", "Spanish", "Russian"
max_words=100, # any integer, values less than 100 and greater than 2000 not recommended
user_id=42, # unique user ID, used to split up usage statistics per user
use_search=False) # use search mode, if True, then bot will try to find answer in internet
Notes: In general, you may to try to use unusual values for:
language
, like "Serbian", but it's not guaranteed that it will work as you expected.max_words
- you may try to use values less than 100 words to save tokens, but you may be experienced with cut-offs. For values greater than 2000 words you may be experienced with timeouts or errors for regular, not 32k or 128k models.user_id
- you may use any integer value, but it's recommended to use your own unique user ID, because it's used to split up usage statistics per user.use_search
- it's special feature for premium plans, you may try to manage it from API, and not from UI, but it's highly not recommended to use with smallermax_words
values, so, while using search, please use values at least 100 or more formax words
.
Streaming mode
By default, bots are working in streaming mode (as in UI), so, you may use chat
method to chat with bot in streaming mode, but you may to switch it off by:
response = api.chat(bot_uid=BOT_UID, prompt='Hello, bot!', stream=False)
In case if you prefer to use streaming mode, you need to get response from generator:
import sys
from ablt_python_api import DoneException
try:
for response in api.chat(bot_uid=BOT_UID, prompt='Hello, bot!', stream=True):
# I use direct stdout output to make output be printed on-the-fly
sys.stdout.write(response)
# To get typewriter effect I forcefully flush output each time
sys.stdout.flush()
except DoneException:
pass # DoneException is raised when bot finished conversation
Statistics
Statistics may be used to obtain data for words and tokens usage for period of time.
Full statistics
# Will return statistics for current date for default user = -1
statistics_for_today = api.get_usage_statistics()
# Will return statistics for date range
statistics_for_range = api.get_usage_statistics(start_date='2022-02-24',
end_date='2023-11-17')
# Will return statistics for user with ID 42
statistics_for_user = api.get_usage_statistics(user_id=42)
You may use schema to validate statistics:
from ablt_python_api.schemas import StatisticsSchema
statistics = StatisticsSchema.model_validate(api.get_usage_statistics())
Statistic for a day
# Will return statistics for current date for user = -1
statistics_for_today = api.get_usage_statistics_for_day()
# Will return statistics for specified date
statistics_for_specific_day = api.get_usage_statistics_for_day(date='2022-02-24')
# Will return statistics for user with ID 42
statistics_for_user = api.get_usage_statistics_for_day(user_id=42)
With schema:
from ablt_python_api.schemas import StatisticItemSchema
statistics = StatisticItemSchema.model_validate(api.get_usage_statistics_for_day())
Total statistics
# Will return statistics for current date for user = -1
statistics_for_today = api.get_total_usage_statistics()
# Well return statistics for specified date
statistics_for_specific_day = api.get_total_usage_statistics(date='2022-02-24')
# Will return statistics for user with ID 42
statistics_for_user = api.get_total_usage_statistics(user_id=42)
With schema:
from ablt_python_api.schemas import StatisticTotalSchema
statistics = StatisticTotalSchema.model_validate(api.get_usage_statistics_for_day())
Troubleshooting:
You can always contact support or contact us in Discord channel.
SSL errors
In some cases, you may be experienced with SSL errors, then you may disable certificate verification:
import ssl
sslcontext = ssl.create_default_context()
sslcontext.check_hostname = False
sslcontext.verify_mode = ssl.CERT_NONE
ABLTApi_async(ssl_context=sslcontext) # for async
ABLTApl(ssl_verify=False) # for sync
Rate limit errors
Never try to flood API with requests, because you may be experienced with rate limit errors. In this case, you need to wait for some time and retry your request. Especially, never try to flood API with simultaneous requests for more users than allowed in your plan.
Timeout errors
In some cases, you may be experienced with timeout errors, then you may decrease max_words
value or use stream = True
to get response on-the-fly.
Other errors
In vary rare cases, you may be experienced with other errors, like rebooting of API itself. To check-up API health, you may use health_check
method:
# will return True if API is healthy, otherwise False
api.health_check()
Best practices
Always check-up and follow Guides, References and Examples from ABLT documentation.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file ablt_python_api-0.0.3.tar.gz
.
File metadata
- Download URL: ablt_python_api-0.0.3.tar.gz
- Upload date:
- Size: 41.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.11.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | f37517f6e1a309c5d26ca674c5cfbcd92a79fa9dc2f83ef3029e58d587ec9e88 |
|
MD5 | 22946ac5c4ee7e983a4a9c253570c0e3 |
|
BLAKE2b-256 | dcc239b621a63d88c09ea3e672afdf6f92e4237bcd87d8122b6ceb5f2b50c768 |
File details
Details for the file ablt_python_api-0.0.3-py3-none-any.whl
.
File metadata
- Download URL: ablt_python_api-0.0.3-py3-none-any.whl
- Upload date:
- Size: 21.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.11.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | fd8b5ed5492bb602fe3930faba86e289b4a7ca6ea792716359e6df8fb43aac0d |
|
MD5 | ffb143e0fae43a030a16c1c0a042af16 |
|
BLAKE2b-256 | cb7cb654cc09653fa56eba6b3313b8258d64a34be5979953c481696c6b45b75d |