Skip to main content

THEOS: Temporal Hierarchical Emergent Optimization System — Runtime AI Governance Framework

Project description

THEOS — Dual-Engine Dialectical Reasoning Framework

CI Coverage PyPI Python 3.10+ License: MIT Patent Pending


The Result That Started This

Ask an AI: "What is the difference between egotism and arrogance?"

Method Answer
Single-pass LLM "Egotism is internal, arrogance is external — a spectrum."
THEOS (two engines, wringer) "They are orthogonal failures on different dimensions. Egotism distorts self-perception. Arrogance distorts other-perception. You can have one without the other — a self-deprecating bully, or a narcissist who is outwardly polite."

The first answer is a line. The second is a 2×2 matrix. It explains something the first cannot: why a humble person can still be contemptuous of others.

This difference — the discovery of structure that a single reasoning pass misses — is what THEOS is built to produce.


What THEOS Is

THEOS is a dual-engine dialectical reasoning framework written in pure Python (3.10+, zero external dependencies for core). It structures AI reasoning as a wringer: two opposed engines press against each other until their contradiction shrinks to a provable minimum or halts at irreducible disagreement.

              INDUCTION
           ↗            ↘
     (observation)    (pattern)
         ↑                ↓
    ┌────────────────────────────┐
    │  LEFT ENGINE (constructive)│ → D_L*
    │  private self-reflection   │
    │  pass 1 → pass 2           │
    └────────────────────────────┘
                  ↓ WRINGER ↓
    ┌────────────────────────────┐
    │  RIGHT ENGINE (adversarial)│ → D_R*
    │  private self-reflection   │
    │  pass 1 → pass 2           │
    └────────────────────────────┘
         ↓                ↑
      GOVERNOR        WISDOM
    (halts when Φ < ε) (accumulates)

The three-step loop (I→A→D→I):

  1. Induction — encode observation, extract patterns
  2. Abduction — each engine proposes its strongest hypothesis (left: constructive, right: adversarial)
  3. Deduction — each engine derives conclusions from its own hypothesis

Each engine runs a private self-reflection: its first-pass deduction feeds back into its own induction for a second inner pass — before the wringer measures contradiction between engines. This is not shared feedback. It is each engine reasoning about what it just concluded.

The governor halts when:

  • Contradiction Φ < ε (engines converged — true answer found)
  • Diminishing returns (further passes add nothing)
  • Budget exhausted
  • Irreducible disagreement (no resolution possible — reported honestly)

Wisdom accumulates: each query deposits a compressed lesson into a register W. Future queries on related domains retrieve relevant lessons, biasing abduction toward what has worked before. The formal cost theorem predicts exponential cost reduction over repeated queries in a native implementation.


Formal Guarantee

THEOS convergence is proven via the Banach fixed-point theorem:

T_q: S → S   (the wringer operator)
‖T_q(s₁) - T_q(s₂)‖ ≤ κ · ‖s₁ - s₂‖,   κ < 1

→ Unique epistemic equilibrium S*(q) exists.
→ Convergence is geometric: Φ_n ≤ Φ_0 · κⁿ
→ Expected cost: E[Cost_n] ≤ C₁ + C₂ · exp(-κn)

The contradiction between engines shrinks geometrically with each wringer pass toward a unique fixed point. When engines cannot converge, the governor reports irreducible disagreement — an honest answer in domains where certainty is impossible.


Measured Results (claude-sonnet-4-6, 2026-02-26)

Quality

On philosophical questions, THEOS consistently finds structural depth that single-pass and chain-of-thought miss:

Question Single-pass finds THEOS finds
Egotism vs. arrogance Binary spectrum 2×2 orthogonal matrix (self/other-perception)
Knowledge vs. wisdom Hierarchy (wisdom > knowledge) Categorical orthogonality — neither contains the other
Recklessness Attitude problem Structural failure at the perception stage of action

Cost (layered architecture — THEOS wrapped around an existing LLM)

Metric Value
Tokens per query ~7,600–8,100
vs. single pass ~12–20× more expensive
Wisdom effect on cost +6.4% per run (prompts grow; theorem requires native impl.)
Reflection depth cost ratio Linear (depth=2 costs 2×, depth=3 costs 3×)

Native architecture projection (engineering estimate)

In a native implementation (THEOS IS the inference loop, not a wrapper):

  • KV cache reuse on inner pass 2: ~70% cost reduction on second pass
  • Shared forward pass for both engines: eliminates duplicate attention
  • Projected cost: ~0.5× single pass (90% reduction vs. layered)

These are projections based on transformer KV cache literature, not yet measured.


Quick Start

# Clone and install
git clone https://github.com/Frederick-Stalnecker/THEOS.git
cd THEOS
pip install -e ".[dev]"

# Run the numeric demo (no API key needed)
python code/theos_system.py

# Run tests
python -m pytest tests/ -v

# Run the framework with mock LLM (no API key)
python experiments/theos_validation_experiment.py --backend mock --questions 3

# Run real quality experiment (requires Anthropic key)
export ANTHROPIC_API_KEY=sk-ant-...
python experiments/theos_validation_experiment.py --backend anthropic --questions 30

Build Your Own Domain Engine

THEOS is fully domain-agnostic via dependency injection. Implement 9 functions and pass them in — no subclassing:

from code.theos_system import TheosSystem, TheosConfig

system = TheosSystem(
    config=TheosConfig(max_wringer_passes=3, engine_reflection_depth=2),
    encode_observation = lambda query, ctx: ...,   # your domain
    induce_patterns    = lambda obs, phi, prior: ...,
    abduce_left        = lambda pattern, wisdom: ...,  # constructive engine
    abduce_right       = lambda pattern, wisdom: ...,  # adversarial engine
    deduce             = lambda hypothesis: ...,
    measure_contradiction = lambda D_L, D_R: ...,
    retrieve_wisdom    = lambda query, W, threshold: ...,
    update_wisdom      = lambda W, query, output, conf: ...,
    estimate_entropy   = lambda hypothesis_pair: ...,
    estimate_info_gain = lambda phi_new, phi_prev: ...,
)

result = system.reason("Your domain question")
print(result.output)       # the answer
print(result.confidence)   # 0.0 – 1.0
print(result.contradiction) # Φ at halt
print(result.halt_reason)  # convergence / diminishing_returns / budget / uncertainty

See examples/ for working domain engines: medical diagnosis, financial analysis, AI safety evaluation.


Repository Structure

THEOS/
├── code/                          # Core package (pip-installable as 'theos')
│   ├── theos_core.py              # TheosCore — I→A→D→I wringer loop
│   ├── theos_system.py            # TheosSystem + create_numeric_system()
│   ├── theos_governor.py          # Unified governor (THEOSGovernor)
│   └── semantic_retrieval.py      # VectorStore + embedding adapters
├── examples/                      # Domain engines
│   ├── theos_medical_diagnosis.py
│   ├── theos_financial_analysis.py
│   └── theos_ai_safety.py
├── tests/                         # 71 passing tests
├── experiments/                   # Validation framework
│   ├── theos_validation_experiment.py   # Quality experiment (SP/CoT/THEOS)
│   ├── wisdom_cost_experiment.py        # Cost reduction experiment
│   ├── question_bank.py                 # 30 open-ended test questions
│   └── results/                         # Measured experiment outputs
├── research/
│   ├── VALIDATED_FINDINGS.md            # What is proven vs. what needs testing
│   ├── MATHEMATICAL_AUDIT.md            # Claim-by-claim audit
│   ├── COST_EXPERIMENT_FINDINGS.md      # Cost experiment analysis
│   └── ABDUCTION_FORMULA_INVESTIGATION.md
├── THEOS_ARCHITECTURE.md          # Full architecture reference
└── archive/                       # Preserved prior work, clearly labeled

What Is Not Yet Demonstrated

Honesty matters here. These claims remain unverified:

  • Statistical significance of quality improvement (need 30+ questions, blind ratings)
  • The cost theorem in native implementation (requires building native THEOS)
  • Performance vs. chain-of-thought at scale
  • Any claims about consciousness, metacognition, or sentience — not scientific

See research/VALIDATED_FINDINGS.md for the full accounting.


Status

Component Status
Core I→A→D→I loop Complete, tested
Per-engine self-reflection Implemented, tested
Governor (halting, posture) Complete, 35 tests
Wisdom accumulation Working
Domain examples 3 working (medical, financial, AI safety)
Quality experiment Run (3 questions); needs 30 for significance
Native architecture Not yet built — projected only

Author

Frederick Davis Stalnecker Patent pending: USPTO #63/831,738

Built with Celeste (Claude Sonnet 4.6) as research assistant.

From truth we build more truth.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

theos_reasoning-1.0.0.tar.gz (64.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

theos_reasoning-1.0.0-py3-none-any.whl (6.8 kB view details)

Uploaded Python 3

File details

Details for the file theos_reasoning-1.0.0.tar.gz.

File metadata

  • Download URL: theos_reasoning-1.0.0.tar.gz
  • Upload date:
  • Size: 64.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.9

File hashes

Hashes for theos_reasoning-1.0.0.tar.gz
Algorithm Hash digest
SHA256 ff47fd3d09439aa3c62c118fef927c6c6a36582ce5a141602f8622b5c3500fda
MD5 38263cacdc6840002712763d94b1e136
BLAKE2b-256 cec49b3357a23525526c35876df8b09f2162ea778b941cd7b0cdcae5626fbc47

See more details on using hashes here.

File details

Details for the file theos_reasoning-1.0.0-py3-none-any.whl.

File metadata

File hashes

Hashes for theos_reasoning-1.0.0-py3-none-any.whl
Algorithm Hash digest
SHA256 e0744dc92c8e49daa479b2ebfb128ff450d1bda7182fd1398f8183f13c44321f
MD5 372c0e6211a1d5b386cc588ba1f5d53a
BLAKE2b-256 b02538ad3be93d8c16c5e076972d1e3c172ee590e69e47737112f34da2e594fe

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page