Natural language interface for quantum computing infrastructure
Project description
text4q Cortex
Natural language interface for quantum computing infrastructure.
Write quantum intent in plain language — Cortex compiles, schedules, and executes it on real QPUs.
from cortex import Cortex
cx = Cortex(backend="ibm_quantum")
result = cx.run("Simulate a Bell state with 2 qubits and measure 1024 times")
print(result.counts)
# {'00': 512, '11': 512}
What is text4q Cortex?
Cortex is an open-source quantum orchestration platform built on three pillars:
- NLP → Circuit: Translate natural language descriptions into OpenQASM 3.0 circuits
- Orchestration: Manage QPU resources via QRMI, integrating quantum and classical HPC
- Execution: Schedule and run jobs across IBM Quantum, Google, or custom lab QPUs
Architecture
User (natural language)
↓
Cortex NLP Engine ← text4q core: language → OpenQASM
↓
OQTOPUS Job Queue ← cloud layer: scheduling + auth
↓
QAOA Scheduler ← quantum-native optimization (roadmap)
↓
QRMI Resource Manager ← QPU as HPC node
↓
QPU / Simulator ← IBM Quantum, Google, Qiskit Aer
Installation
pip install text4q-cortex
Or from source:
git clone https://github.com/your-org/text4q-cortex
cd text4q-cortex
pip install -e ".[dev]"
Quick Start
from cortex import Cortex
from cortex.connectors import IBMQuantumConnector
# Connect to IBM Quantum
connector = IBMQuantumConnector(token="YOUR_IBM_TOKEN")
cx = Cortex(connector=connector)
# Run from natural language
result = cx.run(
"Create a GHZ state with 3 qubits, apply noise model T1=50us, measure 2048 shots"
)
print(result.circuit) # the generated OpenQASM circuit
print(result.counts) # measurement results
print(result.metadata) # backend, shots, execution time
Modules
| Module | Description | Status |
|---|---|---|
cortex.nlp |
NLP → OpenQASM translation engine | 🚧 v0.1 |
cortex.connectors |
IBM Quantum, Aer, Google backends | 🚧 v0.1 |
cortex.scheduler |
Job queue and QPU resource management | 📋 planned |
cortex.cloud |
Multi-user cloud layer (OQTOPUS-based) | 📋 planned |
Roadmap
- Project structure and architecture
- v0.1 — NLP engine (pattern-based) + IBM Quantum connector
- v0.2 — LLM-powered circuit generation + multi-backend
- v0.3 — OQTOPUS job queue integration
- v0.4 — QAOA Scheduler (quantum-native scheduling)
- v1.0 — text4q Cortex Cloud (SaaS)
Contributing
Contributions welcome. Please read CONTRIBUTING.md first.
License
Apache 2.0 — see LICENSE.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file text4q_cortex-0.1.0a0.tar.gz.
File metadata
- Download URL: text4q_cortex-0.1.0a0.tar.gz
- Upload date:
- Size: 33.3 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.13.5
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
51543e419f410ba2eb8754d54e010b1ccd8f1c8372aa999249e9f551b168cee8
|
|
| MD5 |
6f075902afffc4968155b11eb21679cb
|
|
| BLAKE2b-256 |
f10f1c8a14d9b7f627d74dbb81bbebe8b500469b6633da0f053684936a2520ff
|
File details
Details for the file text4q_cortex-0.1.0a0-py3-none-any.whl.
File metadata
- Download URL: text4q_cortex-0.1.0a0-py3-none-any.whl
- Upload date:
- Size: 38.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.13.5
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
3855217626c21df6225c433384ee036aed4a0ca5b823bc7311a0c696218a06e5
|
|
| MD5 |
8b9aa8a119925a407c8880d7ecb10f04
|
|
| BLAKE2b-256 |
b15b44eea9a9632ba19785ffe1a02ece27d008745ebe2c793db873aac1271748
|