Skip to main content

vLLM plugin for Spyre hardware support

Project description

SenDNN Inference

| Documentation | Users Forum | #sig-spyre |


IBM Spyre is the first production-grade Artificial Intelligence Unit (AIU) accelerator born out of the IBM Research AIU family, and is part of a long-term strategy of developing novel architectures and full-stack technology solutions for the emerging space of generative AI. Spyre builds on the foundation of IBM's internal AIU research and delivers a scalable, efficient architecture for accelerating AI in enterprise environments.

SenDNN Inference (sendnn-inference) is a vLLM plugin that enables seamless integration of IBM Spyre Accelerator with vLLM. It follows the architecture described in vLLM's Plugin System, making it easy to integrate IBM's advanced AI acceleration into existing vLLM workflows.

For more information, check out the following:

Getting Started

Visit our documentation:

Contributing

We welcome and value any contributions and collaborations. Please check out Contributing to SenDNN Inference for how to get involved.

Contact

You can reach out for discussion or support in the #sig-spyre channel in the vLLM Slack workspace or by opening an issue.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

sendnn_inference-2.0.0rc12.tar.gz (1.1 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

sendnn_inference-2.0.0rc12-py3-none-any.whl (109.5 kB view details)

Uploaded Python 3

File details

Details for the file sendnn_inference-2.0.0rc12.tar.gz.

File metadata

  • Download URL: sendnn_inference-2.0.0rc12.tar.gz
  • Upload date:
  • Size: 1.1 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for sendnn_inference-2.0.0rc12.tar.gz
Algorithm Hash digest
SHA256 0fc05ea30d9d7b0b7e0bf76c5169100049f2cdef1e0673edc5981d2e46b73030
MD5 0929485a3441eb95f77ad2be6a149544
BLAKE2b-256 3462a40848bc1432dd297834313ab453338f3539146f607a5173dedbaa37cea5

See more details on using hashes here.

Provenance

The following attestation bundles were made for sendnn_inference-2.0.0rc12.tar.gz:

Publisher: build_and_publish.yaml on torch-spyre/sendnn-inference

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file sendnn_inference-2.0.0rc12-py3-none-any.whl.

File metadata

File hashes

Hashes for sendnn_inference-2.0.0rc12-py3-none-any.whl
Algorithm Hash digest
SHA256 5fb79257120c1b793aa011f82136bc041f8da1964ed4e23d962b368d80ef1cbe
MD5 e31e8e9851b0cf96ced8d5366922eb28
BLAKE2b-256 1d18e8fd1105aae508c349ea2dbd144a35956cb3e8f45d267eb2f2774be17a10

See more details on using hashes here.

Provenance

The following attestation bundles were made for sendnn_inference-2.0.0rc12-py3-none-any.whl:

Publisher: build_and_publish.yaml on torch-spyre/sendnn-inference

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page