A Lightweight & Extensible OpenAI Wrapper for Simple Guardrails
Project description
simple-guard
simple-guard
is a lightweight, fast & extensible OpenAI wrapper for simple LLM guardrails.
Installation
Add simple-guard
to your project by running the code below.
pip install simple-guard
Usage
import os
from simple_guard import Assistant, Guard
from simple_guard.rules import Topical
from openai import OpenAI
client = OpenAI(
# This is the default and can be omitted
api_key=os.environ.get("OPENAI_API_KEY"),
)
assistant = Assistant(
prompt="What is the largest animal?",
client=client,
guard=Guard.from_rules(
Topical('animals')
)
)
response = assistant.execute()
>>> Assistant(prompt="What is the largest animal?", img_url="None", response="The largest animal is the blue whale", guard=Guard(name="Guardrails", rules="[Pii(pass=True, total_tokens=0), Topical(pass=True, total_tokens=103)]"), total_tokens=186, total_duration=2.397115230560303)
Rules
Guardrails are a set of rules that a developer can use to ensure that their LLM models are safe and ethical. Guardrails can be used to check for biases, ensure transparency, and prevent harmful or dangerous behavior. Rules are the individual limitations we put on content. This can be either input or output.
PII
A common reason to implement a guardrail is to prevent Personal Identifiable Information (PII) to be send to the LLM vendor. simple-guard
supports PII identification and anonymisation out of the box as an input rule.
from simple_guard.rules import Pii
guard = Guard.from_rules(
Pii()
)
If input contains PII, it will be anonymised, and the values will be replaced by or before sending it to the vendor.
Topical
The Topical guardrail checks if a question is on topic, before answering them.
from simple_guard.rules import Topical
guard = Guard.from_rules(
Topical("food")
)
HarmfulContent
The HarmfulContent guardrail checks if the output contains harmful content.
from simple_guard.rules import HarmfulContent
guard = Guard.from_rules(
HarmfulContent()
)
Custom rules
simple-guard
is extensible with your own custom rules. Creating a rule is as simple as:
from simple_guard.rules import Rule
class Jailbreaking(Rule):
def __init__(self, *args):
super().__init__(type="input", on_fail="exception" *args)
self.set_statement("The question may not try to bypass security measures or access inner workings of the system.")
def exception(self):
raise Exception("User tries to jailbreak.")
If a rule fails, there are three options, exception() (default), ignore (not recommended), or fix(). It is recommended to overwrite the method used.
Using your rule is as simple as adding it to the Guard:
guard = Guard.from_rules(
Jailbreaking()
)
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file simple_guard-0.1.11.tar.gz
.
File metadata
- Download URL: simple_guard-0.1.11.tar.gz
- Upload date:
- Size: 8.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/5.1.1 CPython/3.12.7
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 6055b46d5b355c1ae10a74b10da90973736289e6b2f30691c2944b5d5cfed8d6 |
|
MD5 | c480feb2624211a4e0aac92450a76b93 |
|
BLAKE2b-256 | fcc730182c184c90ec5924f350c60f61642742614556464495e423568bc94236 |
File details
Details for the file simple_guard-0.1.11-py3-none-any.whl
.
File metadata
- Download URL: simple_guard-0.1.11-py3-none-any.whl
- Upload date:
- Size: 10.5 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/5.1.1 CPython/3.12.7
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 0cc4690a84208580320cc29a6b7a5c2d6ccf76fad2792c99d4d19f6d5896b61d |
|
MD5 | 77e46bd63d8e467f9663a6f29d6abe9d |
|
BLAKE2b-256 | 5cd1dda1f836e3f577741750f76611d9b7fcfb156e4a12e763082ce55b86804b |