Measure AI-to-AI influence in multi-agent systems
Project description
TraceIQ
TraceIQ is a Python library for measuring cross-agent influence in multi-agent AI systems, providing reproducible metrics for semantic drift, propagation risk, and reasoning stability.
Installation • Quick Start • Science • Docs • Cite
The Problem
Modern AI systems increasingly operate through interaction: agents collaborate, critique, retrieve, plan, and self-modify through communication with other agents.
Yet we lack scientific tools to answer fundamental questions:
- How much did one agent influence another?
- Did incorrect reasoning propagate through the system?
- When does collaboration become instability?
- How can influence be measured reproducibly and rigorously?
TraceIQ introduces a formal measurement framework for studying influence propagation in autonomous multi-agent AI systems.
Scientific Contributions
TraceIQ provides three core measurement primitives:
1. Influence Quotient (IQx)
A quantitative metric measuring semantic drift in an agent's output caused by prior agent messages.
IQx = Drift_L2 / (Baseline_Median + ε)
IQx estimates how much reasoning changed due to interaction, enabling measurement of cross-agent cognitive influence.
2. Propagation Risk (PR)
A network-level metric estimating how influence spreads across an agent graph using spectral radius analysis.
PR = max(|eigenvalues(Adjacency_Matrix)|)
PR provides early detection of unstable influence propagation and cascading reasoning errors.
3. Reproducible Influence Experiments
TraceIQ includes CI-safe experimental pipelines evaluating:
- ✓ Hint injection and misinformation propagation
- ✓ Influence stability across agent chains
- ✓ Cold-start detection behavior
- ✓ Mitigation policy effectiveness
These experiments support ongoing research into multi-agent AI safety and Contagious Intelligence.
Where TraceIQ Fits in AI Research
Modern AI tooling focuses on different layers of the stack:
| Tool | What It Measures |
|---|---|
| TensorBoard | Model training metrics |
| Weights & Biases | Experiment tracking |
| LangSmith / Prompt tools | Prompt execution traces |
| TraceIQ | Cross-agent influence and reasoning propagation |
TraceIQ introduces a missing instrumentation layer: measurement of cognitive influence between autonomous AI agents.
As multi-agent systems become standard in RAG, planning, robotics, and orchestration, understanding how agents influence each other becomes as critical as measuring accuracy.
Why TraceIQ Exists
Autonomous AI systems are evolving from isolated models into collaborative agent networks.
However, we currently lack standardized methods to measure:
| Challenge | Description |
|---|---|
| Cross-agent reasoning influence | How does one agent's output change another's behavior? |
| Error propagation | Do mistakes cascade through agent pipelines? |
| Stability | Is collaborative reasoning stable or chaotic? |
| Safety risks | What are the emergent risks in multi-agent systems? |
TraceIQ was built as a scientific instrument for studying emergent behavior in distributed intelligence.
TraceIQ is not a dashboard. It is not a monitoring SaaS.
TraceIQ is measurement infrastructure for multi-agent cognition research.
Research Vision
TraceIQ is measurement infrastructure for multi-agent AI systems.
As AI shifts from isolated models to collaborative agent networks, system behavior emerges from interactions between models—not from a single model alone. While we can measure accuracy, latency, and loss, we currently lack tools to quantify how reasoning propagates across agents.
TraceIQ provides reproducible metrics and structured tracking for cross-agent influence, reasoning drift, and propagation dynamics. The goal is not monitoring dashboards, but scientific instrumentation for studying distributed AI cognition.
If collaborative AI becomes the dominant computing paradigm, measuring influence between agents will be as fundamental as measuring model performance.
Features
| Feature | Description |
|---|---|
| 📊 Influence Tracking | Track influence between agent interactions |
| 🎯 Semantic Drift | Measure drift using embedding similarity |
| 🌐 Propagation Risk | Estimate network-level influence spread |
| ⚡ Anomaly Detection | Quantile-calibrated alerting system |
| 🧊 Cold-Start Handling | Statistical validation during warm-up |
| 🔬 Research Pipelines | CI-safe reproducible experiments |
| 🔌 Integration Ready | Templates for RAG and multi-agent orchestration |
Installation
Core library (lightweight, no heavy ML dependencies):
pip install traceiq
With real embedding models:
pip install "traceiq[embedding]"
With research plotting tools:
pip install "traceiq[research]"
Full installation:
pip install "traceiq[embedding,research]"
Real-World Use Cases
TraceIQ is designed for real multi-agent AI systems:
-
Evaluate RAG hallucination propagation Measure whether incorrect retrieval contaminates downstream reasoning.
-
Audit autonomous agent pipelines Track which agents influence critical decisions in planning systems.
-
Study collaborative reasoning stability Detect when agent feedback loops amplify errors.
-
AI governance and accountability Build audit trails showing how decisions evolved across agents.
-
Research on Contagious Intelligence Quantify cognitive transfer between AI systems in controlled experiments.
TraceIQ acts as a measurement microscope for studying distributed AI cognition.
Quick Example
from traceiq import InfluenceTracker
tracker = InfluenceTracker(use_mock_embedder=True)
result = tracker.track_event(
sender_id="agent_a",
receiver_id="agent_b",
sender_content="We should switch to renewable energy.",
receiver_content="Good point. Renewables are the future."
)
print("Drift:", result["drift_l2_state"])
print("IQx:", result["IQx"])
print("Alert:", result["alert"])
tracker.close()
Output:
Drift: 0.847
IQx: 1.23
Alert: False
What TraceIQ Outputs
Each tracked interaction returns structured metrics you can log, visualize, or audit:
drift_l2_state— semantic drift magnitudeIQx— normalized influence scorealert— anomaly signal (calibrated)valid— whether baseline is stabilizedreceiver_state— receiver baseline summary
Research Applications
TraceIQ has been evaluated on synthetic multi-agent benchmarks and integrated into experimental LLM pipelines involving chained, retrieval-augmented, and tool-using agents.
The framework enables reproducible studies of:
- Influence propagation across agent graphs
- Stability of collaborative reasoning loops
- Detection of misleading hint injection
- Mitigation policy effectiveness
- Cold-start behavior in autonomous agents
All experiments are reproducible through CI-safe pipelines
that generate structured summary.json artifacts for verification.
Integration Patterns
TraceIQ works with common agent architectures:
| Pattern | Description |
|---|---|
| LLM-only agents | Track message → response influence |
| RAG systems | Include retrieved context in receiver input |
| Tool-using agents | Track tool output influence |
| Memory agents | Track before/after memory state |
| Multi-agent orchestrators | Full conversation influence graphs |
What TraceIQ Is NOT
| Limitation | Explanation |
|---|---|
| Not causal inference | Metrics measure correlation, not proven causation |
| Not intent detection | Cannot determine manipulation intent |
| Not semantic understanding | Measures embedding-level drift |
| Not a production security system | Research measurement tool |
| Not plug-and-play safety | Thresholds require calibration per environment |
Research Context
TraceIQ supports research into:
- 🔬 AI-to-AI influence modeling
- 🧬 Contagious Intelligence hypothesis
- ⚖️ Multi-agent reasoning stability
- 🛡️ Autonomous system safety
- 🧠 Distributed cognition in AI systems
Detailed metric definitions and implementation notes are available in the project documentation.
Documentation
| Document | Description |
|---|---|
| Metrics | Metric definitions and formulas |
| Integration | Integration patterns |
| CLI Reference | Command-line interface |
| Configuration | TrackerConfig options |
| Architecture | System design |
| Theory | Mathematical foundations |
| Experiments | Research testbed |
CLI Help: traceiq --help
Reproducibility & CI
TraceIQ experiments are CI-safe:
- ✅ Quick mode never hard-fails CI
- ✅ Proof mode enforces strict statistical validation
- ✅ Artifacts upload even on failures
- ✅ Experiments produce structured outputs
This ensures reproducible research pipelines.
Contributing
Contributions welcome! See CONTRIBUTING.md
# 1. Fork the repository
# 2. Create a feature branch
git checkout -b feature/amazing-feature
# 3. Run tests and linter
pytest
ruff check src/ tests/
# 4. Submit a Pull Request
Citation
If you use TraceIQ in your research, please cite:
@software{traceiq,
title = {TraceIQ: Measuring AI-to-AI Influence in Multi-Agent Systems},
author = {Vasavada, Anarv and Contributors},
year = {2026},
url = {https://github.com/Anarv2104/TraceIQ}
}
License
TraceIQ is open-source under the MIT License, enabling academic and commercial use with minimal restrictions.
See the full license text at: https://github.com/Anarv2104/TraceIQ/blob/main/LICENSE
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file traceiq-0.5.2.tar.gz.
File metadata
- Download URL: traceiq-0.5.2.tar.gz
- Upload date:
- Size: 213.8 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
0c23a394da1dba56ec50816ba2041fb215dea30c2e3f4fa30c08082017d30c2d
|
|
| MD5 |
a3efbe9adef6408782d473df7d304bbc
|
|
| BLAKE2b-256 |
d31ebe8ce5c2169884cd7df5779749256ddabab0c60708acf1438f2fa353b740
|
Provenance
The following attestation bundles were made for traceiq-0.5.2.tar.gz:
Publisher:
publish.yml on Anarv2104/TraceIQ
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
traceiq-0.5.2.tar.gz -
Subject digest:
0c23a394da1dba56ec50816ba2041fb215dea30c2e3f4fa30c08082017d30c2d - Sigstore transparency entry: 1004939014
- Sigstore integration time:
-
Permalink:
Anarv2104/TraceIQ@d7c6df5dd9af2555bb25ff594edec72d40f6bd85 -
Branch / Tag:
refs/tags/v0.5.2 - Owner: https://github.com/Anarv2104
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@d7c6df5dd9af2555bb25ff594edec72d40f6bd85 -
Trigger Event:
push
-
Statement type:
File details
Details for the file traceiq-0.5.2-py3-none-any.whl.
File metadata
- Download URL: traceiq-0.5.2-py3-none-any.whl
- Upload date:
- Size: 65.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
b342e2a75536c9f40b1c6ab21f74d853b0b572e15a1c2acd614d9dc5f6444c2e
|
|
| MD5 |
2a1607b828122985c8d4908798da31db
|
|
| BLAKE2b-256 |
52f6e3751b89dc49fa413ca0ff12d7a135f0d77d14e0be8830d17de718f10760
|
Provenance
The following attestation bundles were made for traceiq-0.5.2-py3-none-any.whl:
Publisher:
publish.yml on Anarv2104/TraceIQ
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
traceiq-0.5.2-py3-none-any.whl -
Subject digest:
b342e2a75536c9f40b1c6ab21f74d853b0b572e15a1c2acd614d9dc5f6444c2e - Sigstore transparency entry: 1004939016
- Sigstore integration time:
-
Permalink:
Anarv2104/TraceIQ@d7c6df5dd9af2555bb25ff594edec72d40f6bd85 -
Branch / Tag:
refs/tags/v0.5.2 - Owner: https://github.com/Anarv2104
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@d7c6df5dd9af2555bb25ff594edec72d40f6bd85 -
Trigger Event:
push
-
Statement type: