Skip to main content

A modular and asynchronous framework for neuromorphic optimisation using spike-driven metaheuristics and heuristic-controlled spiking dynamics.

Project description

Status PyPI Version GitHub Tag Python Versions PyPI Downloads License arXiv DOI DOI

NeurOptimiser

NeurOptimiser is a neuromorphic optimisation framework in which metaheuristic search emerges from asynchronous spiking dynamics. It defines optimisation as a decentralised process executed by interconnected Neuromorphic Heuristic Units (NHUs), each embedding a spiking neuron model and a spike-triggered heuristic rule.

This framework enables fully event-driven, low-power optimisation by integrating spiking computation with local heuristic adaptation. It supports multiple neuron models, perturbation operators, and network topologies.


✨ Key Features

  • Modular and extensible architecture using Intel’s Lava.
  • Supports linear and Izhikevich neuron dynamics.
  • Implements random, fixed, directional, and Differential Evolution operators as spike-triggered perturbations.
  • Includes asynchronous neighbourhood management, tensor contraction layers, and greedy selectors.
  • Compatible with BBOB (COCO) suite.
  • Designed for scalability, reusability, and future deployment on Loihi-class neuromorphic hardware.

📖 Documentation

For detailed documentation, examples, and API reference, please visit the Neuroptimiser Documentation.

📦 Installation

pip install -e .

Ensure you have Python ≥ 3.10 and the Lava-NC environment configured.

You can also clone the repository and install it. Check the Makefile for additional options.

🚀 Example Usage

from neuroptimiser import NeuroOptimiser
import numpy as np

problem_function    = lambda x: np.linalg.norm(x)
problem_bounds      = np.array([[-5.0, 5.0], [-5.0, 5.0]])

optimiser = NeurOptimiser()

optimiser.solve(
    obj_func=problem_function,
    search_space=problem_bounds,
    debug_mode=True,
    num_iterations=1000,
)

For more examples, please, visit Neuroptimiser Usage

📊 Benchmarking

Neuroptimiser has been validated over the BBOB suite, showing:

  • Competitive convergence versus Random Search
  • Consistent results across function types and dimensions
  • Linear runtime scaling with number of units and problem size

🔬 Citation

@misc{neuroptimiser2025,
  author={Cruz-Duarte, Jorge M. and Talbi, El-Ghazali},
  title        = {Neuroptimiser: A neuromorphic optimisation framework},
  year         = {2025},
  url          = {https://github.com/neuroptimiser/neuroptimiser},
  note         = {Version 1.0.X, accessed on 20XX-XX-XX}
}

🔗 Resources

🛠️ License

MIT License — see LICENSE

🧑‍💻 Authors

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

neuroptimiser-1.0.1.post1.tar.gz (37.7 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

neuroptimiser-1.0.1.post1-py3-none-any.whl (33.6 kB view details)

Uploaded Python 3

File details

Details for the file neuroptimiser-1.0.1.post1.tar.gz.

File metadata

  • Download URL: neuroptimiser-1.0.1.post1.tar.gz
  • Upload date:
  • Size: 37.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.10.14

File hashes

Hashes for neuroptimiser-1.0.1.post1.tar.gz
Algorithm Hash digest
SHA256 d581eb20edc798e05f6e7f42c589e2b27042cd77f2652b519349a7a33aa5fa60
MD5 520d089964e0d0b8c27ee670f429a21e
BLAKE2b-256 ac3f1610d1275b97f280062925df0632f9e5179f718038ef74e5771977184934

See more details on using hashes here.

File details

Details for the file neuroptimiser-1.0.1.post1-py3-none-any.whl.

File metadata

File hashes

Hashes for neuroptimiser-1.0.1.post1-py3-none-any.whl
Algorithm Hash digest
SHA256 6ce3b0bd47c865831e40a1b59df459056429a046ae41f464c905b1f523eda537
MD5 054b9481aed5c44e061b34e6b8ec42bd
BLAKE2b-256 2e7255f37092dd0803b5a6378e712fe13f114d32c43d5c3e8f153b3d746f7f1d

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page