Add your description here
Project description
TrustyAI LM-Eval as and Out-of-Tree Llama Stack Provider
About
This repository implements TrustyAI's LM-Eval as an out-of-tree Llama Stack remote provider.
Use
Prerequsites
- Admin access to an OpenShift cluster with RHOAI installed
- Login to your OpenShift cluster with
oc login --token=<TOKEN> --server=<SERVER> - Installation of
uv - Installation of
occli tool - Installation of
llama stackcli tool
-
Clone this repository
git clone https://github.com/trustyai-explainability/llama-stack-provider-lmeval.git -
Set
llama-stack-provider-lmevalas your working directory.cd llama-stack-provider-lmeval -
Deploy
microsoft/Phi-3-mini-4k-instructon vLLM Serving Runtimea. Create a namespace with a name of your choice
TEST_NS=<NAMESPACE> oc create ns $TEST_NS oc get ns $TEST_NS
b. Deploy the model via vLLM
oc apply -k demos/resources/kustomization.yaml
-
Before continuing, preform a sanity check to make sure the model was sucessfully deployed
oc get pods | grep "predictor"
Expected output:
phi-3-predictor-00002-deployment-794fb6b4b-clhj7 3/3 Running 0 5h55m -
Create and activate a virtual enviornment
uv venv .llamastack-venvsource .llamastack-venv/bin/activate -
Install the required libraries
uv pip install -e . -
Define the following ennvironment variables
export VLLM_URL=https://$(oc get $(oc get ksvc -o name | grep predictor) --template={{.status.url}})/v1/completions export TRUSTYAI_LM_EVAL_NAMESPACE=$(oc project | cut -d '"' -f2) -
Start the llama stack server in a virtual enviornment
llama stack run run.yaml --image-type venvExpected output:
INFO: Application startup complete. INFO: Uvicorn running on http://['::', '0.0.0.0']:8321 (Press CTRL+C to quit) -
Navigate to
demos/to run the demo notebooks
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file llama_stack_provider_lmeval-0.1.8.tar.gz.
File metadata
- Download URL: llama_stack_provider_lmeval-0.1.8.tar.gz
- Upload date:
- Size: 24.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.12.9
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
b4ff758b8f3c4c108b43af648b9a369a28a26a31aa1c6140da4c7d6ef8702025
|
|
| MD5 |
3c273a5dfe916de73c8cfcc71f2ead14
|
|
| BLAKE2b-256 |
c54b67e8d70458c3670cd4fbdfa8d2eb0b75a0112a2021ec8818463f5bc1a7bb
|
Provenance
The following attestation bundles were made for llama_stack_provider_lmeval-0.1.8.tar.gz:
Publisher:
build-and-push.yaml on trustyai-explainability/llama-stack-provider-lmeval
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
llama_stack_provider_lmeval-0.1.8.tar.gz -
Subject digest:
b4ff758b8f3c4c108b43af648b9a369a28a26a31aa1c6140da4c7d6ef8702025 - Sigstore transparency entry: 356097247
- Sigstore integration time:
-
Permalink:
trustyai-explainability/llama-stack-provider-lmeval@b36c7a53ad1c0a471c3ef333770f33ae1c76ba9e -
Branch / Tag:
refs/tags/v0.1.8 - Owner: https://github.com/trustyai-explainability
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
build-and-push.yaml@b36c7a53ad1c0a471c3ef333770f33ae1c76ba9e -
Trigger Event:
release
-
Statement type:
File details
Details for the file llama_stack_provider_lmeval-0.1.8-py3-none-any.whl.
File metadata
- Download URL: llama_stack_provider_lmeval-0.1.8-py3-none-any.whl
- Upload date:
- Size: 19.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.12.9
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
4acfc1ea0b1cccc9549e78443f8d51efeb168465281a8c5f9fbd2a6945e5af46
|
|
| MD5 |
06e3840aeb01b45ab18c54be27083738
|
|
| BLAKE2b-256 |
8c164c0e0e48fafe4915ae8cf39ca476b445a473410d47a39e8335cbc696855e
|
Provenance
The following attestation bundles were made for llama_stack_provider_lmeval-0.1.8-py3-none-any.whl:
Publisher:
build-and-push.yaml on trustyai-explainability/llama-stack-provider-lmeval
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
llama_stack_provider_lmeval-0.1.8-py3-none-any.whl -
Subject digest:
4acfc1ea0b1cccc9549e78443f8d51efeb168465281a8c5f9fbd2a6945e5af46 - Sigstore transparency entry: 356097276
- Sigstore integration time:
-
Permalink:
trustyai-explainability/llama-stack-provider-lmeval@b36c7a53ad1c0a471c3ef333770f33ae1c76ba9e -
Branch / Tag:
refs/tags/v0.1.8 - Owner: https://github.com/trustyai-explainability
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
build-and-push.yaml@b36c7a53ad1c0a471c3ef333770f33ae1c76ba9e -
Trigger Event:
release
-
Statement type: