Self-hosted AI observability backend
Project description
🔭 Logmera
Self-hosted monitoring for AI applications
Log prompts, responses, and latency from your AI apps and view them in a simple dashboard.
What is Logmera?
Logmera is a self-hosted observability tool for AI / LLM applications.
It helps developers understand what their AI systems are doing.
Instead of printing logs to the console, Logmera stores:
- prompts
- responses
- model name
- latency
in a PostgreSQL database and shows them in a web dashboard.
Your data stays on your infrastructure.
Dashboard
Source Code: https://github.com/ThilakKumar-A/Logmera/
Why use Logmera?
When building AI applications it becomes hard to know:
- what prompts were sent
- what responses were returned
- how long requests took
- which model was used
- when errors happened
Logmera helps you see and monitor all AI activity in one place.
How it works
Your AI App
│
▼
Logmera SDK
│
▼
Logmera Server
│
▼
PostgreSQL
│
▼
Dashboard
Your application sends logs to Logmera. Logmera stores them and displays them in the dashboard.
Quick Start
1. Install
pip install logmera
2. Start PostgreSQL
You can use any PostgreSQL database.
Example connection string:
postgresql://username:password@localhost:5432/database
If you do not have PostgreSQL installed, you can run it with Docker.
docker run --name logmera-postgres \
-e POSTGRES_USER=postgres \
-e POSTGRES_PASSWORD=postgres \
-e POSTGRES_DB=logmera \
-p 5432:5432 -d postgres:16
3. Start Logmera
logmera --db-url "postgresql://postgres:postgres@localhost:5432/logmera"
Server starts at:
http://127.0.0.1:8000
Open the Dashboard
Open your browser:
http://127.0.0.1:8000
You will see your AI logs in the dashboard.
Python Example
Add Logmera to your AI code.
import logmera
logmera.log(
project_id="chatbot",
prompt="Hello",
response="Hi there",
model="gpt-4o",
latency_ms=120,
status="success"
)
Now the request will appear in the Logmera dashboard.
API Example
You can also send logs using the API.
curl -X POST http://127.0.0.1:8000/logs \
-H "Content-Type: application/json" \
-d '{
"project_id":"my-app",
"prompt":"Hello",
"response":"Hi",
"model":"gpt-4o",
"latency_ms":95,
"status":"success"
}'
API Endpoints
| Method | Endpoint | Description |
|---|---|---|
| GET | /health |
Health check |
| POST | /logs |
Create log |
| GET | /logs |
Get logs |
Configuration
Example CLI:
logmera --host 127.0.0.1 --port 8000
Environment variables:
DATABASE_URL
LOGMERA_URL
LOGMERA_TIMEOUT_SECONDS
LOGMERA_RETRIES
Deployment
Logmera can run on:
- Local machines
- Docker
- VPS servers
- Kubernetes
- Cloud VMs
Because Logmera is self-hosted, your AI data remains private.
License
MIT License
Links
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file logmera-0.1.4.tar.gz.
File metadata
- Download URL: logmera-0.1.4.tar.gz
- Upload date:
- Size: 94.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.13.9
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
a4941d4a3f33b38a2c21df9cbf514f6eb1969b8563458895c9e9697efd35177f
|
|
| MD5 |
353cf053b9705fb89bf754d8ecd26649
|
|
| BLAKE2b-256 |
c35499c1086c0a63c78c7f412bc603973c707a61330d98b9db519b643bf3b643
|
File details
Details for the file logmera-0.1.4-py3-none-any.whl.
File metadata
- Download URL: logmera-0.1.4-py3-none-any.whl
- Upload date:
- Size: 93.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.13.9
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
3c8ba235ecc0068143d90608073f534907880cd90127ccc6b4b656f723e67570
|
|
| MD5 |
f4dcc118670aaccd1058f5a410073ec5
|
|
| BLAKE2b-256 |
91b5974eed7376069f6993927e689acafae19710689bf75b6883cd3992dd944f
|