Skip to main content

LLM plugin to access GitHub Models API

Project description

GitHub Models Plugin for LLM

PyPI Changelog

This is a plugin for llm that uses GitHub Models via the Azure AI Inference SDK. GitHub Models is available to all GitHub users and offers free usage of many AI LLMs.

Features

  • Support for all >30 models, including GPT-4o, 4.1, o3, DeepSeek-R1, Llama3.x and more
  • Support for schemas
  • Output token usage
  • Support for Embedding Models
  • Async and streaming outputs (model dependent)
  • Support for model attachments
  • Support for tools

Installation

$ llm install llm-github-models

or pip install llm-github-models

Usage

To set the API key, use the llm keys set github command or use the GITHUB_MODELS_KEY environment variable.

To get an API key, create a personal access token (PAT) inside GitHub Settings.

Learn about rate limits here

All model names are affixed with github/ to distinguish the OpenAI ones from the builtin models.

Example

$ llm prompt 'top facts about cheese' -m github/gpt-4.1-mini
Sure! Here are some top facts about cheese:

1. **Ancient Origins**: Cheese is one of the oldest man-made foods, with evidence of cheese-making dating back over 7,000 years.

2. **Variety**: There are over 1,800 distinct types of cheese worldwide, varying by texture, flavor, milk source, and production methods.

Image attachments

Multi-modal vision models can accept image attachments using the LLM attachments options:

llm -m github/Llama-3.2-11B-Vision-Instruct "Describe this image" -a https://static.simonwillison.net/static/2024/pelicans.jpg

Produces

This image depicts a dense gathering of pelicans, with the largest birds situated in the center, showcasing their light brown plumage and long, pointed beaks. The pelicans are standing on a rocky shoreline, with a serene body of water behind them, characterized by its pale blue hue and gentle ripples. In the background, a dark, rocky cliff rises, adding depth to the scene.

The overall atmosphere of the image exudes tranquility, with the pelicans seemingly engaging in a social gathering or feeding activity. The photograph's clarity and focus on the pelicans' behavior evoke a sense of observation and appreciation for the natural world.

Supported Models

Chat Models

Model Name Schemas Tools Input Modalities Output Modalities
AI21-Jamba-1.5-Large text text
AI21-Jamba-1.5-Mini text text
Codestral-2501 text text
Cohere-command-r text text
Cohere-command-r-08-2024 text text
Cohere-command-r-plus text text
Cohere-command-r-plus-08-2024 text text
DeepSeek-R1 text text
DeepSeek-R1-0528 text text
DeepSeek-V3 text text
DeepSeek-V3-0324 text text
Llama-3.2-11B-Vision-Instruct text, image, audio text
Llama-3.2-90B-Vision-Instruct text, image, audio text
Llama-3.3-70B-Instruct text text
Llama-4-Maverick-17B-128E-Instruct-FP8 text, image text
Llama-4-Scout-17B-16E-Instruct text, image text
MAI-DS-R1 text text
Meta-Llama-3-70B-Instruct text text
Meta-Llama-3-8B-Instruct text text
Meta-Llama-3.1-405B-Instruct text text
Meta-Llama-3.1-70B-Instruct text text
Meta-Llama-3.1-8B-Instruct text text
Ministral-3B text text
Mistral-Large-2411 text text
Mistral-Nemo text text
Mistral-large-2407 text text
Mistral-small text text
Phi-3-medium-128k-instruct text text
Phi-3-medium-4k-instruct text text
Phi-3-mini-128k-instruct text text
Phi-3-mini-4k-instruct text text
Phi-3-small-128k-instruct text text
Phi-3-small-8k-instruct text text
Phi-3.5-MoE-instruct text text
Phi-3.5-mini-instruct text text
Phi-3.5-vision-instruct text, image text
Phi-4 text text
Phi-4-mini-instruct text text
Phi-4-mini-reasoning text text
Phi-4-multimodal-instruct audio, image, text text
Phi-4-reasoning text text
cohere-command-a text text
gpt-4.1 text, image text
gpt-4.1-mini text, image text
gpt-4.1-nano text, image text
gpt-4o text, image, audio text
gpt-4o-mini text, image, audio text
grok-3 text text
grok-3-mini text text
jais-30b-chat text text
mistral-medium-2505 text, image text
mistral-small-2503 text, image text
o1 text, image text
o1-mini text text
o1-preview text text
o3 text, image text
o3-mini text text
o4-mini text, image text

AI21 Jamba 1.5 Large

Usage: llm -m github/AI21-Jamba-1.5-Large

Publisher: AI21 Labs

Description: A 398B parameters (94B active) multilingual model, offering a 256K long context window, function calling, structured output, and grounded generation.

AI21 Jamba 1.5 Mini

Usage: llm -m github/AI21-Jamba-1.5-Mini

Publisher: AI21 Labs

Description: A 52B parameters (12B active) multilingual model, offering a 256K long context window, function calling, structured output, and grounded generation.

Codestral 25.01

Usage: llm -m github/Codestral-2501

Publisher: Mistral AI

Description: Codestral 25.01 by Mistral AI is designed for code generation, supporting 80+ programming languages, and optimized for tasks like code completion and fill-in-the-middle

Cohere Command A

Usage: llm -m github/cohere-command-a

Publisher: Cohere

Description: Command A is a highly efficient generative model that excels at agentic and multilingual use cases.

Cohere Command R

Usage: llm -m github/Cohere-command-r

Publisher: Cohere

Description: Command R is a scalable generative model targeting RAG and Tool Use to enable production-scale AI for enterprise.

Cohere Command R 08-2024

Usage: llm -m github/Cohere-command-r-08-2024

Publisher: Cohere

Description: Command R is a scalable generative model targeting RAG and Tool Use to enable production-scale AI for enterprise.

Cohere Command R+

Usage: llm -m github/Cohere-command-r-plus

Publisher: Cohere

Description: Command R+ is a state-of-the-art RAG-optimized model designed to tackle enterprise-grade workloads.

Cohere Command R+ 08-2024

Usage: llm -m github/Cohere-command-r-plus-08-2024

Publisher: Cohere

Description: Command R+ is a state-of-the-art RAG-optimized model designed to tackle enterprise-grade workloads.

Cohere Embed v3 English

Usage: llm -m github/Cohere-embed-v3-english

Publisher: Cohere

Description: Cohere Embed English is the market's leading text representation model used for semantic search, retrieval-augmented generation (RAG), classification, and clustering.

Cohere Embed v3 Multilingual

Usage: llm -m github/Cohere-embed-v3-multilingual

Publisher: Cohere

Description: Cohere Embed Multilingual is the market's leading text representation model used for semantic search, retrieval-augmented generation (RAG), classification, and clustering.

DeepSeek-R1

Usage: llm -m github/DeepSeek-R1

Publisher: DeepSeek

Description: DeepSeek-R1 excels at reasoning tasks using a step-by-step training process, such as language, scientific reasoning, and coding tasks.

DeepSeek-R1-0528

Usage: llm -m github/DeepSeek-R1-0528

Publisher: DeepSeek

Description: The DeepSeek R1 0528 model has improved reasoning capabilities, this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding.

DeepSeek-V3

Usage: llm -m github/DeepSeek-V3

Publisher: DeepSeek

Description: A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token.

DeepSeek-V3-0324

Usage: llm -m github/DeepSeek-V3-0324

Publisher: DeepSeek

Description: DeepSeek-V3-0324 demonstrates notable improvements over its predecessor, DeepSeek-V3, in several key aspects, including enhanced reasoning, improved function calling, and superior code generation capabilities.

Cohere Embed 4

Usage: llm -m github/embed-v-4-0

Publisher: Cohere

Description: Embed 4 transforms texts and images into numerical vectors

OpenAI GPT-4.1

Usage: llm -m github/gpt-4.1

Publisher: OpenAI

Description: gpt-4.1 outperforms gpt-4o across the board, with major gains in coding, instruction following, and long-context understanding

OpenAI GPT-4.1-mini

Usage: llm -m github/gpt-4.1-mini

Publisher: OpenAI

Description: gpt-4.1-mini outperform gpt-4o-mini across the board, with major gains in coding, instruction following, and long-context handling

OpenAI GPT-4.1-nano

Usage: llm -m github/gpt-4.1-nano

Publisher: OpenAI

Description: gpt-4.1-nano provides gains in coding, instruction following, and long-context handling along with lower latency and cost

OpenAI GPT-4o

Usage: llm -m github/gpt-4o

Publisher: OpenAI

Description: OpenAI's most advanced multimodal model in the gpt-4o family. Can handle both text and image inputs.

OpenAI GPT-4o mini

Usage: llm -m github/gpt-4o-mini

Publisher: OpenAI

Description: An affordable, efficient AI solution for diverse text and image tasks.

Grok 3

Usage: llm -m github/grok-3

Publisher: xAI

Description: Grok 3 is xAI's debut model, pretrained by Colossus at supermassive scale to excel in specialized domains like finance, healthcare, and the law.

Grok 3 Mini

Usage: llm -m github/grok-3-mini

Publisher: xAI

Description: Grok 3 Mini is a lightweight model that thinks before responding. Trained on mathematic and scientific problems, it is great for logic-based tasks.

JAIS 30b Chat

Usage: llm -m github/jais-30b-chat

Publisher: Core42

Description: JAIS 30b Chat is an auto-regressive bilingual LLM for Arabic & English with state-of-the-art capabilities in Arabic.

Llama-3.2-11B-Vision-Instruct

Usage: llm -m github/Llama-3.2-11B-Vision-Instruct

Publisher: Meta

Description: Excels in image reasoning capabilities on high-res images for visual understanding apps.

Llama-3.2-90B-Vision-Instruct

Usage: llm -m github/Llama-3.2-90B-Vision-Instruct

Publisher: Meta

Description: Advanced image reasoning capabilities for visual understanding agentic apps.

Llama-3.3-70B-Instruct

Usage: llm -m github/Llama-3.3-70B-Instruct

Publisher: Meta

Description: Llama 3.3 70B Instruct offers enhanced reasoning, math, and instruction following with performance comparable to Llama 3.1 405B.

Llama 4 Maverick 17B 128E Instruct FP8

Usage: llm -m github/Llama-4-Maverick-17B-128E-Instruct-FP8

Publisher: Meta

Description: Llama 4 Maverick 17B 128E Instruct FP8 is great at precise image understanding and creative writing, offering high quality at a lower price compared to Llama 3.3 70B

Llama 4 Scout 17B 16E Instruct

Usage: llm -m github/Llama-4-Scout-17B-16E-Instruct

Publisher: Meta

Description: Llama 4 Scout 17B 16E Instruct is great at multi-document summarization, parsing extensive user activity for personalized tasks, and reasoning over vast codebases.

MAI-DS-R1

Usage: llm -m github/MAI-DS-R1

Publisher: Microsoft

Description: MAI-DS-R1 is a DeepSeek-R1 reasoning model that has been post-trained by the Microsoft AI team to fill in information gaps in the previous version of the model and improve its harm protections while maintaining R1 reasoning capabilities.

Meta-Llama-3-70B-Instruct

Usage: llm -m github/Meta-Llama-3-70B-Instruct

Publisher: Meta

Description: A powerful 70-billion parameter model excelling in reasoning, coding, and broad language applications.

Meta-Llama-3-8B-Instruct

Usage: llm -m github/Meta-Llama-3-8B-Instruct

Publisher: Meta

Description: A versatile 8-billion parameter model optimized for dialogue and text generation tasks.

Meta-Llama-3.1-405B-Instruct

Usage: llm -m github/Meta-Llama-3.1-405B-Instruct

Publisher: Meta

Description: The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.

Meta-Llama-3.1-70B-Instruct

Usage: llm -m github/Meta-Llama-3.1-70B-Instruct

Publisher: Meta

Description: The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.

Meta-Llama-3.1-8B-Instruct

Usage: llm -m github/Meta-Llama-3.1-8B-Instruct

Publisher: Meta

Description: The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.

Ministral 3B

Usage: llm -m github/Ministral-3B

Publisher: Mistral AI

Description: Ministral 3B is a state-of-the-art Small Language Model (SLM) optimized for edge computing and on-device applications. As it is designed for low-latency and compute-efficient inference, it it also the perfect model for standard GenAI applications that have

Mistral Large (2407)

Usage: llm -m github/Mistral-large-2407

Publisher: Mistral AI

Description: Mistral Large (2407) is an advanced Large Language Model (LLM) with state-of-the-art reasoning, knowledge and coding capabilities.

Mistral Large 24.11

Usage: llm -m github/Mistral-Large-2411

Publisher: Mistral AI

Description: Mistral Large 24.11 offers enhanced system prompts, advanced reasoning and function calling capabilities.

Mistral Medium 3 (25.05)

Usage: llm -m github/mistral-medium-2505

Publisher: Mistral AI

Description: Mistral Medium 3 is an advanced Large Language Model (LLM) with state-of-the-art reasoning, knowledge, coding and vision capabilities.

Mistral Nemo

Usage: llm -m github/Mistral-Nemo

Publisher: Mistral AI

Description: Mistral Nemo is a cutting-edge Language Model (LLM) boasting state-of-the-art reasoning, world knowledge, and coding capabilities within its size category.

Mistral Small

Usage: llm -m github/Mistral-small

Publisher: Mistral AI

Description: Mistral Small can be used on any language-based task that requires high efficiency and low latency.

Mistral Small 3.1

Usage: llm -m github/mistral-small-2503

Publisher: Mistral AI

Description: Enhanced Mistral Small 3 with multimodal capabilities and a 128k context length.

OpenAI o1

Usage: llm -m github/o1

Publisher: OpenAI

Description: Focused on advanced reasoning and solving complex problems, including math and science tasks. Ideal for applications that require deep contextual understanding and agentic workflows.

OpenAI o1-mini

Usage: llm -m github/o1-mini

Publisher: OpenAI

Description: Smaller, faster, and 80% cheaper than o1-preview, performs well at code generation and small context operations.

OpenAI o1-preview

Usage: llm -m github/o1-preview

Publisher: OpenAI

Description: Focused on advanced reasoning and solving complex problems, including math and science tasks. Ideal for applications that require deep contextual understanding and agentic workflows.

OpenAI o3

Usage: llm -m github/o3

Publisher: OpenAI

Description: o3 includes significant improvements on quality and safety while supporting the existing features of o1 and delivering comparable or better performance.

OpenAI o3-mini

Usage: llm -m github/o3-mini

Publisher: OpenAI

Description: o3-mini includes the o1 features with significant cost-efficiencies for scenarios requiring high performance.

OpenAI o4-mini

Usage: llm -m github/o4-mini

Publisher: OpenAI

Description: o4-mini includes significant improvements on quality and safety while supporting the existing features of o3-mini and delivering comparable or better performance.

Phi-3-medium instruct (128k)

Usage: llm -m github/Phi-3-medium-128k-instruct

Publisher: Microsoft

Description: Same Phi-3-medium model, but with a larger context size for RAG or few shot prompting.

Phi-3-medium instruct (4k)

Usage: llm -m github/Phi-3-medium-4k-instruct

Publisher: Microsoft

Description: A 14B parameters model, proves better quality than Phi-3-mini, with a focus on high-quality, reasoning-dense data.

Phi-3-mini instruct (128k)

Usage: llm -m github/Phi-3-mini-128k-instruct

Publisher: Microsoft

Description: Same Phi-3-mini model, but with a larger context size for RAG or few shot prompting.

Phi-3-mini instruct (4k)

Usage: llm -m github/Phi-3-mini-4k-instruct

Publisher: Microsoft

Description: Tiniest member of the Phi-3 family. Optimized for both quality and low latency.

Phi-3-small instruct (128k)

Usage: llm -m github/Phi-3-small-128k-instruct

Publisher: Microsoft

Description: Same Phi-3-small model, but with a larger context size for RAG or few shot prompting.

Phi-3-small instruct (8k)

Usage: llm -m github/Phi-3-small-8k-instruct

Publisher: Microsoft

Description: A 7B parameters model, proves better quality than Phi-3-mini, with a focus on high-quality, reasoning-dense data.

Phi-3.5-mini instruct (128k)

Usage: llm -m github/Phi-3.5-mini-instruct

Publisher: Microsoft

Description: Refresh of Phi-3-mini model.

Phi-3.5-MoE instruct (128k)

Usage: llm -m github/Phi-3.5-MoE-instruct

Publisher: Microsoft

Description: A new mixture of experts model

Phi-3.5-vision instruct (128k)

Usage: llm -m github/Phi-3.5-vision-instruct

Publisher: Microsoft

Description: Refresh of Phi-3-vision model.

Phi-4

Usage: llm -m github/Phi-4

Publisher: Microsoft

Description: Phi-4 14B, a highly capable model for low latency scenarios.

Phi-4-mini-instruct

Usage: llm -m github/Phi-4-mini-instruct

Publisher: Microsoft

Description: 3.8B parameters Small Language Model outperforming larger models in reasoning, math, coding, and function-calling

Phi-4-mini-reasoning

Usage: llm -m github/Phi-4-mini-reasoning

Publisher: Microsoft

Description: Lightweight math reasoning model optimized for multi-step problem solving

Phi-4-multimodal-instruct

Usage: llm -m github/Phi-4-multimodal-instruct

Publisher: Microsoft

Description: First small multimodal model to have 3 modality inputs (text, audio, image), excelling in quality and efficiency

Phi-4-Reasoning

Usage: llm -m github/Phi-4-reasoning

Publisher: Microsoft

Description: State-of-the-art open-weight reasoning model.

OpenAI Text Embedding 3 (large)

Usage: llm -m github/text-embedding-3-large

Publisher: OpenAI

Description: Text-embedding-3 series models are the latest and most capable embedding model from OpenAI.

OpenAI Text Embedding 3 (small)

Usage: llm -m github/text-embedding-3-small

Publisher: OpenAI

Description: Text-embedding-3 series models are the latest and most capable embedding model from OpenAI.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

llm_github_models-0.17.0.tar.gz (16.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

llm_github_models-0.17.0-py3-none-any.whl (12.1 kB view details)

Uploaded Python 3

File details

Details for the file llm_github_models-0.17.0.tar.gz.

File metadata

  • Download URL: llm_github_models-0.17.0.tar.gz
  • Upload date:
  • Size: 16.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.12.9

File hashes

Hashes for llm_github_models-0.17.0.tar.gz
Algorithm Hash digest
SHA256 b992f59baba6ca67145e967069ea42d5eabe17843b46a5f43bb28a86ccfdf6db
MD5 281b4be3daa51ba61250825ee6cdea09
BLAKE2b-256 4ccc9ad6ecd555dbe654c874c0de6cb7ae31668daae9e4c78cd6d50c806b31fe

See more details on using hashes here.

Provenance

The following attestation bundles were made for llm_github_models-0.17.0.tar.gz:

Publisher: python-publish.yml on tonybaloney/llm-github-models

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file llm_github_models-0.17.0-py3-none-any.whl.

File metadata

File hashes

Hashes for llm_github_models-0.17.0-py3-none-any.whl
Algorithm Hash digest
SHA256 363f8a4cb4078af73b12aa84b39947e4420e9b51e3b5414c177c4997a1ea66fc
MD5 f67fac3159725e01fda640b4b124309b
BLAKE2b-256 78cc57bc1a59b07532c345b166cc498b3a383e7a5f39aa97ccfb034e7c9dc870

See more details on using hashes here.

Provenance

The following attestation bundles were made for llm_github_models-0.17.0-py3-none-any.whl:

Publisher: python-publish.yml on tonybaloney/llm-github-models

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page