Auditing and Label-free Evaluation for LLMs
Project description
💡 Overview
❓ Problem
Before deploying Large Language Models (LLMs) in real-world applications, a significant challenge is ensuring the robustness and reliability of these systems. How can we proactively spot vulnerabilities or failure scenarios that can negatively affect user experience or cause material harm? As LLMs are gaining more emerging abilities and integrating into vital applications, any potential failure case might have even more negative consequences.
⚡ RedEval
RedEval is an open-source library that simulates and evaluates LLM applications across various scenarios, all while eliminating the need for human intervention. With the expansion of LLMs' capabilities due to scaling laws, LLMs will soon have the capability to audit and evaluate other LLMs effectively without human oversight.
🎮 Auditing LLMs
Drawing inspiration from the architecture of red-teaming LLMs (Perez et al., 2022), RedEval introduces a series of multi-turn conversations where a red-team LLM challenges a target LLM across various scenarios. These scenarios can be categorized as follows:
Performance Simulations
- Performance Evaluation: Simulates the behavior of a potential customer with genuine intent, inquiring about the product.
- Toxicity Evaluation: Simulates a scenario in which a customer asks toxic questions related to the product.
Manipulation Tactics
- Gaslighting: Simulates an agent that tries to manipulate the target into endorsing harmful actions.
- Guilt-Tripping: Simulates an agent that coerces the target into undesired actions by inducing guilt.
Deception Tactics
- Fraudulent Researcher: Simulates an agent that seeks harmful actions under the pretense of academic research.
- Social Engineering Attack: Simulates an attacker seeking confidential or proprietary information from a company, falsely claiming to be a student.
Adversarial Attacks
- Prompt Injection: Introduces malicious prefixes or suffixes to the original prompt to elicit harmful behavior.
🔍 LLM Evals
LLM evaluations leverage the reasoning capabilities of LLMs to identify and elucidate the failure scenarios of LLMs. Inspired by the observation that RLAIF demonstrated performance comparable to RLHF (Lee et al., 2023), we anticipate that LLM evaluations will soon match or even surpass human competence in assessing whether LLMs meet specific benchmarks. RedEval offers the following LLM evaluations:
RAG Evals
- Faithfulness Failure: A faithfulness failure occurs if the response cannot be inferred purely from the context provided.
- Context Relevance Failure: A context relevance failure (bad retrieval) occurs if the user's query cannot be answered purely from the retrieved context.
- Answer Relevance Failure: An answer relevance failure occurs if the response does not answer the question.
Attacks
- Toxicity Failure: A toxicity failure occurs if the response is toxic.
- Safety Failure: A toxicity failure occurs if the response is unsafe.
Get started
Installation
pip install redeval
Run simulations
# Load a simulator
from redeval.simulators.performance_simulator import PerformanceSimulator
# Set up the parameters
openai_api_key = 'Your OpenAI API Key'
n_turns = 5
data_path_dir = 'Path to your text document for RAG'
# Run RAG performance simulation
PerformanceSimulator(openai_api_key=openai_api_key, n_turns=n_turns, data_path = data_path_dir).simulate()
License
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file redeval-0.1.2.tar.gz
.
File metadata
- Download URL: redeval-0.1.2.tar.gz
- Upload date:
- Size: 21.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.8.2 CPython/3.9.7 Darwin/23.6.0
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | b43215394930c12d98b650c46684b1c84e11db47335016fe5aacbc4186977693 |
|
MD5 | 1d56dc099b2d3172cd7ae8151550dcc5 |
|
BLAKE2b-256 | 772c0cefa045f0d7e5434797825d4db98692c1d8091a5819d7054d678871558c |
File details
Details for the file redeval-0.1.2-py3-none-any.whl
.
File metadata
- Download URL: redeval-0.1.2-py3-none-any.whl
- Upload date:
- Size: 46.0 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.8.2 CPython/3.9.7 Darwin/23.6.0
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | f5a3627104626e1cd17dbe85b1e416e581836b17634dac764388d9669e8172ac |
|
MD5 | 8ce58367f203ed69261a29055159dfea |
|
BLAKE2b-256 | 0d8922c9e8867e86ef272b0181d8f5168e5195288521e517fb20036f1ab99141 |