LLM plugin to access Google's Gemini family of models
Project description
llm-gemini
API access to Google's Gemini models
Installation
Install this plugin in the same environment as LLM.
llm install llm-gemini
Usage
Configure the model by setting a key called "gemini" to your API key:
llm keys set gemini
<paste key here>
Now run the model using -m gemini-1.5-pro-latest
, for example:
llm -m gemini-1.5-pro-latest "A joke about a pelican and a walrus"
A pelican walks into a seafood restaurant with a huge fish hanging out of its beak. The walrus, sitting at the bar, eyes it enviously.
"Hey," the walrus says, "That looks delicious! What kind of fish is that?"
The pelican taps its beak thoughtfully. "I believe," it says, "it's a billfish."
Images, audio and video
Gemini models are multi-modal. You can provide images, audio or video files as input like this:
llm -m gemini-1.5-flash-latest 'extract text' -a image.jpg
Or with a URL:
llm -m gemini-1.5-flash-8b-latest 'describe image' \
-a https://static.simonwillison.net/static/2024/pelicans.jpg
Audio works too:
llm -m gemini-1.5-pro-latest 'transcribe audio' -a audio.mp3
And video:
llm -m gemini-1.5-pro-latest 'describe what happens' -a video.mp4
Code execution
Gemini models can write and execute code - they can decide to write Python code, execute it in a secure sandbox and use the result as part of their response.
To enable this feature, use -o code_execution 1
:
llm -m gemini-1.5-pro-latest -o code_execution 1 \
'use python to calculate (factorial of 13) * 3'
Chat
To chat interactively with the model, run llm chat
:
llm chat -m gemini-1.5-pro-latest
Other models are:
gemini-1.5-flash-latest
- gemini-1.5-flash-8b-latest` - the least expensive
Embeddings
The plugin also adds support for the text-embedding-004
embedding model.
Run that against a single string like this:
llm embed -m text-embedding-004 -c 'hello world'
This returns a JSON array of 768 numbers.
This command will embed every README.md
file in child directories of the current directory and store the results in a SQLite database called embed.db
in a collection called readmes
:
llm embed-multi readmes --files . '*/README.md' -d embed.db -m text-embedding-004
You can then run similarity searches against that collection like this:
llm similar readmes -c 'upload csvs to stuff' -d embed.db
See the LLM embeddings documentation for further details.
Development
To set up this plugin locally, first checkout the code. Then create a new virtual environment:
cd llm-gemini
python3 -m venv venv
source venv/bin/activate
Now install the dependencies and test dependencies:
llm install -e '.[test]'
To run the tests:
pytest
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file llm_gemini-0.3.tar.gz
.
File metadata
- Download URL: llm_gemini-0.3.tar.gz
- Upload date:
- Size: 9.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.1 CPython/3.13.0
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 5e9fbe25910c61a3ea0434d94f2d07123f7f76468231f6e8a7b8ca17c855e37f |
|
MD5 | 7503f83d4ae62e0fed0a14aca21de243 |
|
BLAKE2b-256 | 5bd85ba3f9fd6228ab91a96308c6da62414357c9944ed6fa1576975d5c41df66 |
File details
Details for the file llm_gemini-0.3-py3-none-any.whl
.
File metadata
- Download URL: llm_gemini-0.3-py3-none-any.whl
- Upload date:
- Size: 9.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.1 CPython/3.13.0
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | b795c7c809790f2b1f5337398b89c2bb81576e0f6f6dc48049f63444f64f7757 |
|
MD5 | 3dc757ec10b10077eeeaf865a5bfd4a4 |
|
BLAKE2b-256 | 65807b14867041437955add9ce530be4a6bd955502a411c63e8dd87640fd5d88 |