Production-grade adaptive meta-learning framework for continual model improvement. Implements research DOI: 10.5281/zenodo.17839490.
Project description
AIRBORNE-ANTARA
Adaptive Neural Thinking Architecture For Recursive Autonomy
V8.1 // CODENAME: "SENTIENT" EDITION (PRODUCTION READY)
"Intelligence is no longer just trained. It is synthesized through awareness."
| Autonomous Consciousness | Unified Memory |
|---|---|
Recursive Workspace V2 |
Holographic Saliency Pooling |
🏆 SENTIENT CAPABILITIES (V8.0)
[!IMPORTANT] ANTARA V8.0 is a non-destructive cognitive wrapper. It does not replace your model weights; it builds a "conscious" manifold around them.
1. Unified Memory (SI + EWC + Universal OGD)
Result: Eliminated catastrophic forgetting across arbitrary architectures. V9.3 Update: Implemented Universal Tensor Projection, extending memory protection to Conv2d, Attention, and RNN layers, making the entire backbone effectively immortal.
See
airborne_antara/memory.py
2. Recursive Consciousness (System 2)
Result: Enabled slow, deliberative reasoning over complex tasks using the Recursive Global Workspace. The model now generates and evaluates thought traces before final execution.
See
airborne_antara/consciousness_v2.py
3. Perception Gateway (Multi-Modal)
Result: Native support for Vision, Audio, and Text via ViT-style encoders with Dynamic Positional Interpolation for variable input scales.
See
airborne_antara/perception.py
4. Autonomic Health (MoE-Aware)
Result: Self-healing neural substrate. V9.3 Update: The monitor is now MoE-Aware, surgically preserving dormant expert knowledge while rejuvenating truly dead neurons in active manifold paths.
See
airborne_antara/core.py
🧬 THE 4 PILLARS OF SENTIENCE
1. CONSCIOUSNESS V2 (Global Workspace)
Implements System 2 Thinking. Instead of a single forward pass, the model projects states into a recursive workspace to simulate "thinking about the problem."
- Thought Trace: Internal hidden state evolution logged as "telemetry" for debugging.
- Recursive Workspace: Dynamic number of internal reasoning loops based on task entropy.
2. HOLOGRAPHIC MEMORY (Unified Handler)
Combines Elastic Weight Consolidation (EWC), Synaptic Intelligence (SI), and Orthogonal Gradient Descent (OGD).
- Saliency Pooling: Dynamically prioritizes historical parameters to prevent erasure.
- Experience Replay: Generative replay of "dreams" during idle cycles to consolidate learning.
3. MULTI-MODAL PERCEPTION GATEWAY
Unified manifold for Vision (Transformers), Audio (Spectral-Temporal), and Text.
- Positional Interpolation: Scalable attention windows for high-resolution vision.
- Modality Fusion: Cross-modal attention tokens for joint reasoning.
4. AUTONOMIC HEALTH MONITOR
A background daemon tracking the "Neural Health" of the host model.
- Neural Shivering: Injecting controlled stochastic noise to prevent saturation.
- Gradient Centralization: Modern optimization to stabilize deep manifold learning.
🧪 RESEARCH / EXPERIMENTAL (V9.2)
[!CAUTION] These features are in preview for the NeurIPS ablation suite and may exhibit instability in production.
- Self-Awareness V2: Metacognitive engine calculating "Confidence" and "Competence" in real-time.
- I-JEPA World Model: Predictive foresight for world-dynamic modeling.
- Holographic Compression: Next-gen memory storage with $O(log N)$ retrieval complexity.
⚡ INTEGRATION PROTOCOL
The architecture is designed for "One-Line Cognitive Injection".
import torch
from airborne_antara import AdaptiveFramework, PRESETS
# 1. DEFINE YOUR PYTORCH MODEL (Transformer, CNN, etc.)
model = MySubstrate()
# 2. INJECT SENTIENT LAYER
# Uses the 'production' preset: Consciousness V2 + Unified Memory + MoE
agent = AdaptiveFramework(model, PRESETS.production())
# 3. CONSCIOUS TRAINING LOOP
# The agent handles Mixed Precision (AMP), Memory Consolidation, and Thought Tracing
for inputs, targets in dataloader:
metrics = agent.train_step(inputs, target_data=targets)
print(f"Loss: {metrics['loss']:.4f} | Surprise: {metrics['surprise']:.4f}")
print(f"Cognitive Mode: {metrics['mode']}") # [NORMAL, NOVELTY, PANIC]
🖥️ TELEMETRY INTERFACE
Visualizing the internal state (Surprise, Memory Adjacency, Expert Utilization) is possible via the CLI dashboard.
python -m airborne_antara --demo
📂 RESEARCH DOCUMENTATION
V8.0 "Sentient" Release // 2026
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file airborne_antara-0.1.35.tar.gz.
File metadata
- Download URL: airborne_antara-0.1.35.tar.gz
- Upload date:
- Size: 87.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.8
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
fd2b1205053c4047d448be993b1f5e61b1d3a1189eefff955c1564c9023dddea
|
|
| MD5 |
fff14f606ab22ee3ecd5027a22a69482
|
|
| BLAKE2b-256 |
e93464102894f4032f9f49574a6f22e61cda863ee22dce59b2f3e277854c7ac9
|
File details
Details for the file airborne_antara-0.1.35-py3-none-any.whl.
File metadata
- Download URL: airborne_antara-0.1.35-py3-none-any.whl
- Upload date:
- Size: 86.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.8
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
3880420b68256acf7bf69dcf1848a565b6ec995cce1d332acb42ca60a0baaab6
|
|
| MD5 |
a9cd91fc4b5906a20e0e6532d364d07e
|
|
| BLAKE2b-256 |
2f03bc665a459b2f499b3c31ef92cc1a1c49a8eea406b7dd78057ec93bc2dcf2
|