Skip to main content

vLLM plugin for Spyre hardware support

Project description

SenDNN Inference

| Documentation | Users Forum | #sig-spyre |


IBM Spyre is the first production-grade Artificial Intelligence Unit (AIU) accelerator born out of the IBM Research AIU family, and is part of a long-term strategy of developing novel architectures and full-stack technology solutions for the emerging space of generative AI. Spyre builds on the foundation of IBM's internal AIU research and delivers a scalable, efficient architecture for accelerating AI in enterprise environments.

SenDNN Inference (sendnn-inference) is a vLLM plugin that enables seamless integration of IBM Spyre Accelerator with vLLM. It follows the architecture described in vLLM's Plugin System, making it easy to integrate IBM's advanced AI acceleration into existing vLLM workflows.

For more information, check out the following:

Getting Started

Visit our documentation:

Contributing

We welcome and value any contributions and collaborations. Please check out Contributing to SenDNN Inference for how to get involved.

Contact

You can reach out for discussion or support in the #sig-spyre channel in the vLLM Slack workspace or by opening an issue.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

sendnn_inference-2.0.0rc11.tar.gz (1.1 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

sendnn_inference-2.0.0rc11-py3-none-any.whl (108.4 kB view details)

Uploaded Python 3

File details

Details for the file sendnn_inference-2.0.0rc11.tar.gz.

File metadata

  • Download URL: sendnn_inference-2.0.0rc11.tar.gz
  • Upload date:
  • Size: 1.1 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for sendnn_inference-2.0.0rc11.tar.gz
Algorithm Hash digest
SHA256 3f55b31a43070c08acff4b78c5e1c1cf4b4e6647658ee36fcd2905865c6498ed
MD5 54b5567aaadd59bd7bcf8a899730a861
BLAKE2b-256 2d59a65c5a1084bcaf31bf59bdb7ad37e420ad247dd62cbc1d035f8b1644f792

See more details on using hashes here.

Provenance

The following attestation bundles were made for sendnn_inference-2.0.0rc11.tar.gz:

Publisher: build_and_publish.yaml on torch-spyre/sendnn-inference

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file sendnn_inference-2.0.0rc11-py3-none-any.whl.

File metadata

File hashes

Hashes for sendnn_inference-2.0.0rc11-py3-none-any.whl
Algorithm Hash digest
SHA256 32c66650585bdf10cd11a4412070693dd7e34b57597398d07a3765649d7e506a
MD5 7c0ce1bc1195680bc13dee097759c687
BLAKE2b-256 027ed20450b330c01d63a92d8e6d1e3dd6b9b3ed5964f277adf7fc2b8739f985

See more details on using hashes here.

Provenance

The following attestation bundles were made for sendnn_inference-2.0.0rc11-py3-none-any.whl:

Publisher: build_and_publish.yaml on torch-spyre/sendnn-inference

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page