AI Prompt Engineer Agent — prompt preprocessing, security screening, and context enrichment for LLM systems
Project description
AIPEA — AI Prompt Engineer Agent
A standalone Python library for prompt preprocessing, security screening, query analysis, and context enrichment for LLM systems. Extracted from Agora IV production (v4.1.49).
Security: Report vulnerabilities privately via GitHub Security Advisories or email
security@undercurrentholdings.com. SeeSECURITY.mdfor scope, response SLAs, and honest framing of what AIPEA's compliance modes do and do not enforce.
Architecture
AIPEA processes prompts through a multi-stage pipeline:
User Query → SecurityScanner → QueryAnalyzer → SearchOrchestrator → PromptEngine → Enhanced Prompt
│ │ │ │
PII/PHI scan Tier routing Context fetch Model-specific
Classification Domain detect Knowledge base prompt formatting
Injection guard Complexity score MCP providers Tier processing
Core Modules
| Module | Purpose |
|---|---|
security |
PII/PHI detection, classification markers, injection prevention, compliance modes |
analyzer |
Query complexity scoring, domain detection, temporal awareness, tier routing |
search |
Multi-provider search orchestration (Exa, Firecrawl, Context7) |
knowledge |
Offline knowledge base with SQLite storage and domain-aware retrieval |
engine |
Model-specific prompt formatting, tier-based processing (Offline/Tactical/Strategic) |
learning |
Adaptive strategy learning — records user feedback, tracks per-strategy performance, suggests better strategies over time |
enhancer |
High-level facade coordinating the full pipeline |
Processing Tiers
| Tier | Latency | Use Case |
|---|---|---|
| Offline | <2s | Air-gapped, classified, simple queries |
| Tactical | 2-5s | Standard queries with search context |
| Strategic | 5-15s | Complex research, multi-source synthesis |
Installation
# Library only (no CLI)
pip install aipea
# With CLI tools (adds Typer + Rich)
pip install aipea[cli]
# From source (development)
pip install -e ".[dev]"
Getting Started
AIPEA works out of the box with zero configuration. API keys and Ollama are entirely optional — they unlock richer enhancement when available.
Path 1: Minimal (no setup needed)
pip install aipea
from aipea import enhance_prompt
result = await enhance_prompt("What is quantum computing?", model_id="gpt-4")
# Works immediately with template-based enhancement — no API keys required
Path 2: With Search Providers (real-time web context)
Exa provides AI-powered web search; Firecrawl provides structured web content retrieval. Both offer free tiers.
pip install aipea[cli]
aipea configure # interactive wizard — press Enter to skip any key
Path 3: With Ollama (local LLM enhancement)
Ollama runs open-source LLMs locally for richer offline enhancement. AIPEA auto-detects it when available.
# macOS
brew install ollama
# Linux
curl -fsSL https://ollama.ai/install.sh | sh
# Pull a lightweight model
ollama pull gemma3:1b
# Populate the offline knowledge base
aipea seed-kb
Run aipea doctor at any time to see what capabilities are active and what you can add.
Configuration
All API keys are optional. AIPEA can be configured via environment variables, a .env file, or ~/.aipea/config.toml. Priority: env vars > .env > global TOML > defaults.
# Interactive setup wizard (requires [cli] extra)
aipea configure
# Save to global config instead of project .env
aipea configure --global
# Check current configuration
aipea check
# Full diagnostic report
aipea doctor
Or configure manually with environment variables:
export EXA_API_KEY="your-exa-key"
export FIRECRAWL_API_KEY="your-firecrawl-key"
export AIPEA_HTTP_TIMEOUT=30 # seconds (optional)
Or create a .env file in your project root:
EXA_API_KEY="your-exa-key"
FIRECRAWL_API_KEY="your-firecrawl-key"
Usage
Quick Start — enhance_prompt
The simplest way to use AIPEA is through the enhance_prompt facade:
import asyncio
from aipea import enhance_prompt
async def main():
result = await enhance_prompt(
"What are the latest advances in transformer architectures?",
model_id="gpt-5.2",
)
print(result.enhanced_prompt)
print(f"Processing tier: {result.processing_tier}")
print(f"Security context: {result.security_context.security_level}")
asyncio.run(main())
Security Scanning
from aipea import SecurityScanner, SecurityContext, SecurityLevel, ComplianceMode
scanner = SecurityScanner()
context = SecurityContext(
security_level=SecurityLevel.UNCLASSIFIED,
compliance_mode=ComplianceMode.HIPAA,
)
# Scan for PII, PHI, classification markers, and injection attempts
result = scanner.scan("Patient John Doe, SSN 123-45-6789, diagnosed with...", context)
print(result.has_pii()) # True
print(result.has_phi()) # True
print(result.is_blocked) # False (PII is flagged, not blocked)
print(result.flags) # ["pii_detected:ssn", "phi_detected:diagnosis", ...]
Query Analysis
from aipea import QueryAnalyzer
analyzer = QueryAnalyzer()
analysis = analyzer.analyze("Compare CRISPR-Cas9 efficiency across cell types in 2026 studies")
print(analysis.query_type) # QueryType.RESEARCH
print(analysis.suggested_tier) # ProcessingTier.STRATEGIC
print(analysis.complexity) # 0.85
print(analysis.needs_current_info) # True (detected temporal reference)
print(analysis.domain_indicators) # ["biology", "genetics"]
Offline Knowledge Base
import asyncio
from aipea import OfflineKnowledgeBase, StorageTier, KnowledgeDomain
async def main():
kb = OfflineKnowledgeBase("/path/to/knowledge.db", StorageTier.STANDARD)
# Add domain knowledge
await kb.add_knowledge(
"Transformer attention mechanism computes Q*K^T/sqrt(d_k) for scaled dot-product attention.",
domain=KnowledgeDomain.TECHNICAL,
)
# Search knowledge base
results = await kb.search("attention mechanism", limit=5)
for node in results.nodes:
print(f"[{node.relevance_score:.2f}] {node.content[:100]}")
kb.close()
asyncio.run(main())
Search Orchestration
import asyncio
from aipea import SearchOrchestrator
async def main():
orchestrator = SearchOrchestrator()
# Multi-provider search with strategy selection
results = await orchestrator.search(
"quantum error correction 2026",
strategy="multi_source",
num_results=10,
)
for result in results.results:
print(f"[{results.source}] {result.title}: {result.url}")
asyncio.run(main())
Adaptive Learning (opt-in)
AIPEA can learn from user feedback to improve strategy selection over time. This is opt-in and uses a local SQLite database — no data leaves your machine.
import asyncio
from aipea import AIPEAEnhancer
async def main():
enhancer = AIPEAEnhancer(enable_learning=True)
# Enhance a query — strategy_used is tracked on the result
result = await enhancer.enhance(
"How do transformer attention mechanisms work?",
model_id="claude-opus-4-6",
)
print(result.strategy_used) # e.g. "technical"
# Record user feedback (positive = good enhancement)
await enhancer.record_feedback(result, score=0.9)
# After enough feedback (3+ per strategy), AIPEA uses the
# historically best-performing strategy for each query type.
# The next enhance() call may use a different strategy if
# the learning engine has accumulated better data.
enhancer.close()
asyncio.run(main())
The learning engine stores data in aipea_learning.db (configurable via AIPEA_LEARNING_DB_PATH). If the database is unavailable, AIPEA falls back to default strategy selection with no error.
Integration
AIPEA is designed as a standalone preprocessing layer for LLM systems. It integrates with:
- AEGIS Governance — engineering standards & compliance SDK (
pip install aegis-governance[aipea]) - Agora IV — multi-model orchestration platform (uses AIPEA for prompt preprocessing)
Enterprise & Governance
AIPEA is free and open-source. For organizations that need full AI governance — risk registers, model cards, compliance auditing, and policy enforcement — see AEGIS, Undercurrent AI's governance platform.
What AIPEA's Compliance Modes Do — and Do Not Do
AIPEA exposes a ComplianceMode enum with three supported modes (GENERAL, HIPAA, TACTICAL). These modes are input-inspection and model-allowlist controls, not enforcement of any regulatory regime. Read this before shipping AIPEA into a regulated environment:
| Mode | What AIPEA does | What AIPEA does not do |
|---|---|---|
GENERAL |
Default. PII scanning, injection detection, and homoglyph normalization run for every request. | — |
HIPAA |
Enables PHI regex patterns (MRN, patient identifier, diagnosis-term detection); restricts the LLM allowlist to BAA-capable models; emits phi_detected:* flags in the scan result; logs a runtime warning on match. |
Does not redact. Does not block the prompt. Does not persist an audit trail. Does not satisfy the HIPAA Security Rule, Privacy Rule, or BAA requirement on behalf of the integrator. |
TACTICAL |
Forces the processing tier to Offline; restricts the LLM allowlist to locally-runnable models; enables classified-marker regex patterns (CONFIDENTIAL / SECRET / TOP SECRET and common compartment markings). | Does not validate an air-gap. Does not enforce classification handling beyond detection. Does not substitute for an accredited tactical enclave. |
Responsibility stays with the integrator. AIPEA is an input-inspection layer: it tells you what it saw. Your application is responsible for the enforcement decision (block, redact, route to a compliant backend, log to an immutable audit store, obtain BAAs, encrypt at rest and in transit, and satisfy whatever regulatory regime applies).
Deprecated:
ComplianceMode.FEDRAMP. AIPEA previously exposed a FedRAMP enum value. That mode was always a config-only stub with no behavioral enforcement. As of v1.3.4 it is formally deprecated — constructing aComplianceHandlerwithFEDRAMPemits aDeprecationWarning— and scheduled for removal in v2.0.0. AIPEA does not implement FedRAMP controls; integrators needing FedRAMP should migrate toComplianceMode.GENERALand implement FedRAMP controls in their own application layer. Decision rationale:docs/adr/ADR-002-fedramp-removal.md.
For the full scope statement, supported versions, and vulnerability disclosure policy, see SECURITY.md.
AEGIS adds the organizational layer that AIPEA deliberately does not: audit trails, human oversight workflows, policy enforcement, and regulatory reporting. If you need those, pair AIPEA with AEGIS rather than treating AIPEA's compliance modes as substitutes for them.
Learn more at undercurrentholdings.com.
Development
# Install with dev dependencies
pip install -e ".[dev]"
# Run all checks
make all # format + lint + type check + tests
# Individual commands
make fmt # Ruff format + auto-fix
make lint # Ruff check + format check
make type # mypy strict mode
make test # pytest with coverage (75% minimum)
make sec # Security-focused lint rules
make ci # CI parity (lint + type + test, no autofix)
Testing
# Full test suite with coverage
pytest tests/ -v --cov=src/aipea --cov-report=term-missing
# Run specific module tests
pytest tests/test_security.py -v
pytest tests/test_analyzer.py -v
pytest tests/test_engine.py -v
License
MIT
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file aipea-1.4.0.tar.gz.
File metadata
- Download URL: aipea-1.4.0.tar.gz
- Upload date:
- Size: 466.7 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
df3243c2c54119485e3d9ce0e1174eb8d9623d2480538b979d8f7dc4a1ae55f5
|
|
| MD5 |
0e238bb1ebc5ac05750d880a7a27d41a
|
|
| BLAKE2b-256 |
c41a63aed615fee819c0df93604544ecb1d130f07e106dbf8fce423ea52ef6e6
|
Provenance
The following attestation bundles were made for aipea-1.4.0.tar.gz:
Publisher:
publish.yml on undercurrentai/AIPEA
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
aipea-1.4.0.tar.gz -
Subject digest:
df3243c2c54119485e3d9ce0e1174eb8d9623d2480538b979d8f7dc4a1ae55f5 - Sigstore transparency entry: 1288717439
- Sigstore integration time:
-
Permalink:
undercurrentai/AIPEA@0e0f4ce477dcb38699964fc433db7722d3d28b31 -
Branch / Tag:
refs/tags/v1.4.0 - Owner: https://github.com/undercurrentai
-
Access:
private
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
self-hosted -
Publication workflow:
publish.yml@0e0f4ce477dcb38699964fc433db7722d3d28b31 -
Trigger Event:
release
-
Statement type:
File details
Details for the file aipea-1.4.0-py3-none-any.whl.
File metadata
- Download URL: aipea-1.4.0-py3-none-any.whl
- Upload date:
- Size: 106.8 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
98a5f71e0fe4a4cb192eb1eaa74c16bbc323b0eac2fcc8d16edd0a67587995e0
|
|
| MD5 |
d9b868e5b24eb02bd3e4dbf758004b6b
|
|
| BLAKE2b-256 |
33792e096b0f8e3c190d355ef3f1df2c8d644d6d4ffe023c484942ececd86782
|
Provenance
The following attestation bundles were made for aipea-1.4.0-py3-none-any.whl:
Publisher:
publish.yml on undercurrentai/AIPEA
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
aipea-1.4.0-py3-none-any.whl -
Subject digest:
98a5f71e0fe4a4cb192eb1eaa74c16bbc323b0eac2fcc8d16edd0a67587995e0 - Sigstore transparency entry: 1288717478
- Sigstore integration time:
-
Permalink:
undercurrentai/AIPEA@0e0f4ce477dcb38699964fc433db7722d3d28b31 -
Branch / Tag:
refs/tags/v1.4.0 - Owner: https://github.com/undercurrentai
-
Access:
private
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
self-hosted -
Publication workflow:
publish.yml@0e0f4ce477dcb38699964fc433db7722d3d28b31 -
Trigger Event:
release
-
Statement type: