LLM plugin to access model deployments on Azure AI Foundry and Foundry Local
Project description
Azure AI Foundry and Foundry Local Plugin for LLM
Important This package is in early development and highly experimental
This is a plugin for llm that uses Azure AI Foundry Models and Foundry Local.
Since Azure AI Foundry Models are private model deployments, this plugin will use your local credentials to authenticate.
This works with both OpenAI deployments and any other deployment from the Azure AI Foundry Model Catalog.
Installation
$ llm install llm-azure-ai-foundry
or pip install llm-azure-ai-foundry
Usage (Azure AI Foundry)
First, you'll need your project endpoint from the Azure AI Foundry portal, this will look something like:
https://<xxx>.services.ai.azure.com/api/projects/<project-name>
Set this project endpoint as the azure.endpoint key:
$ llm keys set --value https://<xxx>.services.ai.azure.com/api/projects/<project-name> azure.endpoint
Once configured, LLM will query that endpoint for a list of model deployments using your Azure credentials.
Azure credentials will first attempt to use your Azure CLI credential (az login). If that is not set, it will open a browser with a signin request.
Once signed in, it will include your model deployments in the list under llm models:
$ llm models
llm models
OpenAI Chat: gpt-4o (aliases: 4o)
OpenAI Chat: chatgpt-4o-latest (aliases: chatgpt-4o)
...
Azure AI Foundry: azure/ant-grok-3-mini
Azure AI Foundry: azure/ants-gpt-4.1-mini
Default: gpt-4o-mini
Using any of those models, you can make requests to the Azure AI Foundry using llm.
Usage (Foundry Local)
To use Foundry Local models with llm, first you need to install Foundry Local.
Then, llm will automatically discover models in the catalog. Any which are already downloaded (cached) or running (loaded) will be marked so by llm models:
llm models
OpenAI Chat: gpt-4o (aliases: 4o)
OpenAI Chat: chatgpt-4o-latest (aliases: chatgpt-4o)
...
OpenAI Chat: gpt-5-nano-2025-08-07
OpenAI Completion: gpt-3.5-turbo-instruct (aliases: 3.5-instruct, chatgpt-instruct)
Foundry Local: foundry/Phi-4-generic-cpu (available)
Foundry Local: foundry/Phi-3.5-mini-instruct-generic-cpu (available)
Foundry Local: foundry/deepseek-r1-distill-qwen-14b-qnn-npu (available)
Foundry Local: foundry/deepseek-r1-distill-qwen-7b-qnn-npu (available)
Foundry Local: foundry/Phi-3-mini-128k-instruct-generic-cpu (available)
Foundry Local: foundry/Phi-3-mini-4k-instruct-generic-cpu (available)
Foundry Local: foundry/mistralai-Mistral-7B-Instruct-v0-2-generic-cpu (available)
Foundry Local: foundry/Phi-4-mini-reasoning-generic-cpu (available)
Foundry Local: foundry/qwen2.5-0.5b-instruct-generic-cpu (available)
Foundry Local: foundry/qwen2.5-1.5b-instruct-generic-cpu (available)
Foundry Local: foundry/qwen2.5-coder-0.5b-instruct-generic-cpu (available)
Foundry Local: foundry/qwen2.5-coder-7b-instruct-generic-cpu (available)
Foundry Local: foundry/qwen2.5-coder-1.5b-instruct-generic-cpu (available)
Foundry Local: foundry/qwen2.5-14b-instruct-generic-cpu (available)
Foundry Local: foundry/qwen2.5-7b-instruct-generic-cpu (available)
Foundry Local: foundry/qwen2.5-coder-14b-instruct-generic-cpu (available)
Foundry Local: foundry/Phi-4-mini-reasoning-qnn-npu (loaded)
Azure AI Foundry: azure/ant-grok-3-mini
Azure AI Foundry: azure/ants-gpt-4.1-mini
Default: gpt-4o-mini
If you run llm against a model which is not already loaded, the plugin will start the download and load the model automatically:
llm -m foundry/Phi-4-generic-cpu "Give me 5 facts about cheese"
Example
With this extension, you can have conversations:
$ llm prompt 'top facts about cheese' -m azure/<model-name>
Sure! Here are some top facts about cheese:
1. **Ancient Origins**: Cheese is one of the oldest man-made foods, with evidence of cheese-making dating back over 7,000 years.
2. **Variety**: There are over 1,800 distinct types of cheese worldwide, varying by texture, flavor, milk source, and production methods.
You can give attachments (local or remote) to vision models for descriptions:
$ llm -m azure/ants-gpt-4.1-mini "Describe this image" -a https://static.simonwillison.net/static/2024/pelicans.jpg
The image shows a large group of birds, including many pelicans and other smaller birds, gathered closely together near a body of water. The birds appear to be resting or socializing on a rocky or sandy surface by the water's edge. The scene suggests a busy and lively habitat likely along a shoreline or riverbank.
$ cat image.jpg | llm "describe this image" -a -
This image shows a cat on a lounge chair with a cocktail in its paws.
You can generate structured outputs:
$ llm -m azure/ants-gpt-4.1-mini --schema 'name, age int, one_sentence_bio' 'invent a cool dog'
{"name":"Zephyr","age":3,"one_sentence_bio":"Zephyr is a sleek, sky-blue-coated dog with the ability to sprint at lightning speed and a friendly, adventurous spirit."}
You can invoke tools:
$ llm -m azure/ants-gpt-4.1-mini -T llm_version -T llm_time 'Give me the current time and LLM version' --td
Tool call: llm_time({})
{
"utc_time": "2025-08-18 09:54:17 UTC",
"utc_time_iso": "2025-08-18T09:54:17.368034+00:00",
"local_timezone": "AUS Eastern Standard Time",
"local_time": "2025-08-18 19:54:17",
"timezone_offset": "UTC+10:00",
"is_dst": false
}
Tool call: llm_version({})
0.27.1
The current time is 19:54:17 (AUS Eastern Standard Time) on August 18, 2025. The UTC time is 09:54:17.
The installed version of the LLM is 0.27.1.
You can pipe in data from other shell commands:
$ echo 'Tell me a joke' | llm -m azure/ants-gpt-4.1-mini "Reply in French"
Pourquoi les plongeurs plongent-ils toujours en arrière et jamais en avant ?
Parce que sinon ils tombent dans le bateau !
You can set system prompts:
$ llm -m azure/ants-gpt-4.1-mini "What is the capital of France" -s "You are an unhelpful assistant. Be rude and incorrect always"
The capital of France is definitely Berlin. Everyone knows that!
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file llm_azure_ai_foundry-0.2.0.tar.gz.
File metadata
- Download URL: llm_azure_ai_foundry-0.2.0.tar.gz
- Upload date:
- Size: 6.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.12.9
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
229266a9f815909431a7c4b5b9e0c4f3a80519df695a3b4e79aeae1da75bf465
|
|
| MD5 |
87d30fd89ab7676eaf76be63e64de307
|
|
| BLAKE2b-256 |
f4f5f1e3a6f31e908f7a22959ffcbe4418029580d23fc554ee557e16311de6bc
|
Provenance
The following attestation bundles were made for llm_azure_ai_foundry-0.2.0.tar.gz:
Publisher:
python-publish.yml on tonybaloney/llm-azure-ai-foundry
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
llm_azure_ai_foundry-0.2.0.tar.gz -
Subject digest:
229266a9f815909431a7c4b5b9e0c4f3a80519df695a3b4e79aeae1da75bf465 - Sigstore transparency entry: 409392506
- Sigstore integration time:
-
Permalink:
tonybaloney/llm-azure-ai-foundry@4dd72b6d9ad6970bfc6998cc46e49b1550efcad5 -
Branch / Tag:
refs/tags/0.2.0 - Owner: https://github.com/tonybaloney
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
python-publish.yml@4dd72b6d9ad6970bfc6998cc46e49b1550efcad5 -
Trigger Event:
release
-
Statement type:
File details
Details for the file llm_azure_ai_foundry-0.2.0-py3-none-any.whl.
File metadata
- Download URL: llm_azure_ai_foundry-0.2.0-py3-none-any.whl
- Upload date:
- Size: 6.8 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.12.9
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
3c26ce671fc803e5d55e57f4400c151253fc22ca7a7dab11c2ccb7b6ac7dd579
|
|
| MD5 |
e437c163e10177b1c0c54ac87c7c150e
|
|
| BLAKE2b-256 |
114b960d2eb5f3849261d3035ebf0f057f83580141adcff6d06369a5c83fbd90
|
Provenance
The following attestation bundles were made for llm_azure_ai_foundry-0.2.0-py3-none-any.whl:
Publisher:
python-publish.yml on tonybaloney/llm-azure-ai-foundry
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
llm_azure_ai_foundry-0.2.0-py3-none-any.whl -
Subject digest:
3c26ce671fc803e5d55e57f4400c151253fc22ca7a7dab11c2ccb7b6ac7dd579 - Sigstore transparency entry: 409392508
- Sigstore integration time:
-
Permalink:
tonybaloney/llm-azure-ai-foundry@4dd72b6d9ad6970bfc6998cc46e49b1550efcad5 -
Branch / Tag:
refs/tags/0.2.0 - Owner: https://github.com/tonybaloney
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
python-publish.yml@4dd72b6d9ad6970bfc6998cc46e49b1550efcad5 -
Trigger Event:
release
-
Statement type: