Unofficial Python client for Gradient Chat (supports gpt-oss-120b and qwen3-235b)
Project description
gradient-chat-python
Unofficial Python client for Gradient Chat which utilizes the decentralized inference network called Parallax. When using Gradient Chat (i.e Parallax), the inference load is distributed among multiple P2P devices.
Note: Currently Parallax is in testing phase and has limited number of participating devices.
Features
- Maintain conversation context between requests.
- Optionally choose model, cluster mode and context size per request.
- GPT OSS 120B
- Qwen3 235B
- Support for reasoning output (
enableThinking). - Logging of all requests and responses (JSON + plain text).
Installation
python3 -m venv venv
source venv/bin/activate # Linux/macOS
venv\Scripts\activate # Windows
pip install gradient-chat-client
Or if you want to install the latest development version:
pip install git+https://github.com/abswn/gradient-chat-python.git
Usage
from gradient_chat import GradientChatClient, GradientChatError
# Create client
client = GradientChatClient()
# Show available models
print("Available Models:", client.available_models)
# Send a message
try:
response = client.generate(
user_message="Hi, Good morning!",
enableThinking=True
)
print("Model:", response["model"])
print("Reasoning:", response["reasoning"])
print("Reply:", response["reply"])
except GradientChatError as e:
print("Request failed:", e)
API Reference
GradientChatClient
GradientChatClient(
model="GPT OSS 120B", # GPT OSS 120B (default) or Qwen3 235B
cluster_mode="nvidia", # nvidia (default) or hybrid, Qwen3 supports only hyrbid
log_dir="logs",
timeout=None # default is 60 seconds
)
These parameters can also be set per request in the generate method.
client.generate()
response = gradient_client.generate(
user_message, # required
context_size=5, # default is 15 and capped at a max of 50
model="GPT OSS 120B",
cluster_mode="nvidia",
enableThinking=True, # enables reasoning, False by default
timeout=100, # default timeout is 60 seconds
)
OUTPUT:
{
"reply": str, # response to the user message
"reasoning": str, # reasoning used by the model
"model": str # model name
}
All parameters except user_message are optional. There is also a parameter called conversation of type GradientConversation which can be used to send custom conversation history as context.
from gradient_chat import GradientConversation
custom_convo = GradientConversation(max_history=500)
custom_convo.add_user_message("Hi")
custom_convo.add_assistant_message("Hello!") # can also add reasoning text
custom_convo.add_user_message("How are you?")
custom_convo.add_assistant_message("I'm fine.")
client.generate("Hello", conversation=custom_convo)
Custom convo object can also be created directly as follows:
custom_convo = [
{"role": "user", "content": "How are you?"},
{"role": "assistant", "content": "I'm fine.", "reasoningContent": "Responded with a polite, conventional reply to a common greeting to keep the conversation natural."},
# add more
]
client.generate(user_message, custom_convo)
client.get_model_info()
(can also use client.available_models)
['GPT OSS 120B', 'Qwen3 235B']
client.get_conversation()
[
{"role": "user", "content": "Hello"},
{"role": "assistant", "content": "Hi there! How can I help you?", "reasoningContent": "Greeted the user."},
{"role": "user", "content": "Can you tell me a joke?"},
{"role": "assistant", "content": "Why don’t scientists trust atoms? Because they make up everything!"},
{"role": "user", "content": "Thanks! What's the weather like today?"},
{"role": "assistant", "content": "I cannot access real-time weather, but I recommend checking a local weather site.", "reasoningContent": "Explained limitations."}
]
Error Handling
All fatal errors raise GradientChatError.
Catch this single exception to handle any request failure.
from gradient_chat import GradientChatClient, GradientChatError
client = GradientChatClient()
try:
response = client.generate("Status update?")
print(response)
except GradientChatError as e:
msg = str(e)
if "Timeout" in msg:
print("Retrying with a higher timeout...")
client.generate("Status update?", timeout=120)
elif "Job Failed" in msg:
print("Switching to another model...")
client.generate("Status update?", model="Qwen3 235B")
else:
raise # Let other errors (such as HTTP error, Network error) propagate
| Failure Type | Cause | Suggested Action |
|---|---|---|
| Request Timeout | API took longer than timeout seconds. | Retry or increase timeout. |
| HTTP Error | API returned non-2xx status. | Retry or investigate payload. |
| Network Error | Connection issues, DNS failure, SSL errors, etc. | Check connection / proxy. |
| Job Failed | API responded but never sent status == "completed". |
Retry or switch to another model/cluster. |
Non fatal errors are logged via warnings.warn() and do not stop execution.
Disclaimer
This project is a personal undertaking and is not an official Gradient product. It is not affiliated with Gradient in any way and should not be mistaken as such.
License
MIT License
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file gradient_chat_client-0.1.0.tar.gz.
File metadata
- Download URL: gradient_chat_client-0.1.0.tar.gz
- Upload date:
- Size: 14.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
74d45afcb263b1d298e45f2e835810414a6e5da151087cd7843258e610d196bf
|
|
| MD5 |
58769ed6dfa0c8a5322d07c88239884e
|
|
| BLAKE2b-256 |
fe3782fea3b1556f48aa2894223bbca0acee3fe7934b1c4e3296927ddc744584
|
File details
Details for the file gradient_chat_client-0.1.0-py3-none-any.whl.
File metadata
- Download URL: gradient_chat_client-0.1.0-py3-none-any.whl
- Upload date:
- Size: 9.9 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
7b0ce21d1f6691139d9737d24f7cf14fefe595c9697ed0658afc0d7fb5cfd638
|
|
| MD5 |
f1234fbe6894d0c3db505e6521f52483
|
|
| BLAKE2b-256 |
01b157daa96324c0abecaf78b97608a831af65097c18581d50ea8c7709dc37f2
|