A dead simple FastAPI profiler with CSV export for Google Sheets.
Project description
FastAPI Simple Profiler
A dead simple profiler for FastAPI applications, designed to provide per-request performance metrics and export them to a CSV format easily importable into Google Sheets or other spreadsheet software.
Features
- Middleware-based: Easily integrate into your FastAPI application with a single middleware.
- Per-request Metrics: Capture total request wall clock time (TotalTimeMs) and CPU time (CPUTimeMs) for each API call.
- Conditional Activation: Enable profiling via a URL query parameter (?profile=true) or by setting the FASTAPI_SIMPLE_PROFILER_ENABLED=true environment variable to control overhead.
- In-Memory Storage: Temporarily stores recent profiling data in memory, with a configurable retention policy.
- CSV Export Endpoint: Access a dedicated endpoint (/profiler/metrics.csv) to download collected metrics as a CSV file.
- Google Sheets Ready: CSV format is optimized for direct import into spreadsheet applications.
- Lightweight: Designed for minimal overhead, especially when profiling is not active.
Installation
You can install the package using pip:
pip install fastapi-simple-profiler
This package depends on pyinstrument for detailed CPU time measurement, pandas for CSV generation, and fastapi/starlette for the web framework integration. These dependencies will be automatically installed.
Usage
1. Integrate the Middleware
Add ProfilerMiddleware to your FastAPI application instance.
from fastapi import FastAPI
from fastapi.responses import StreamingResponse
from fastapi_simple_profiler import ProfilerMiddleware, profiler_instance
import uvicorn
import time
import asyncio
app = FastAPI()
# Add the profiler middleware to your FastAPI application.
# You can configure its behavior:
# - `enable_by_default`: Set to `True` to profile all requests by default.
# (Default: `False`)
# - `profile_query_param`: The query parameter name to toggle profiling.
# (Default: "profile")
# - `max_retained_requests`: The maximum number of requests to keep in memory.
# Older requests are automatically pruned.
# (Default: 1000)
app.add_middleware(
ProfilerMiddleware,
enable_by_default=False, # Set to True to enable profiling for all requests by default
profile_query_param="profile", # e.g., use `?profile=true` in URL
max_retained_requests=500 # Keep data for the last 500 requests in memory
)
@app.get("/")
async def read_root():
"""A simple root endpoint."""
await asyncio.sleep(0.01) # Simulate some async I/O work
return {"message": "Hello World"}
@app.get("/items/{item_id}")
async def read_item(item_id: int):
"""An endpoint that simulates some compute-bound or I/O work."""
if item_id % 2 == 0:
await asyncio.sleep(0.05) # Simulate longer async work for even IDs
else:
# Simulate some blocking CPU work (e.g., heavy computation)
# This will be reflected in CPUTimeMs by pyinstrument
_ = [i*i for i in range(100000)] # CPU-bound loop
time.sleep(0.005) # Small blocking sleep to show in TotalTimeMs too
return {"item_id": item_id, "message": "Item processed"}
@app.get("/slow-endpoint")
async def slow_endpoint():
"""An intentionally slow endpoint."""
await asyncio.sleep(0.5) # Simulate significant async delay
return {"message": "This was a slow request!"}
@app.get("/profiler/metrics.csv")
async def get_profiler_metrics_csv():
"""
Dedicated endpoint to download the collected profiling metrics as a CSV file.
This uses FastAPI's StreamingResponse for efficient file download.
"""
csv_buffer = profiler_instance.export_to_csv()
return StreamingResponse(
csv_buffer,
media_type="text/csv",
headers={"Content-Disposition": "attachment; filename=fastapi_profile_metrics.csv"}
)
@app.get("/profiler/clear")
async def clear_profiler_data():
"""
Endpoint to clear all collected profiling data from memory.
Useful for resetting the collected metrics.
"""
profiler_instance.clear_data()
return {"message": "Profiler data cleared."}
if __name__ == "__main__":
# To run this example:
# 1. Save the above code as `main.py` in your project root.
# 2. Ensure `fastapi-simple-profiler` is installed (`pip install fastapi-simple-profiler`).
# 3. Run from your terminal: `uvicorn main:app --reload --port 8000`
#
# To enable profiling for ALL requests via environment variable:
# FASTAPI_SIMPLE_PROFILER_ENABLED=true uvicorn main:app --reload --port 8000
uvicorn.run(app, host="0.0.0.0", port=8000)
2. Run your FastAPI Application
Run your FastAPI application using Uvicorn (recommended ASGI server for FastAPI):
uvicorn your_app_module:app --reload --port 8000
(Replace your_app_module with the name of your Python file, e.g., main).
3. Generate Profiled Requests
Make some requests to your FastAPI application.
-
Profile specific requests: If enable_by_default is False (the default), append ?profile=true to the URL for requests you want to profile:
-
Profile all requests:
- Set enable_by_default=True when adding ProfilerMiddleware to your app.
- OR set the environment variable before running your app: FASTAPI_SIMPLE_PROFILER_ENABLED=true uvicorn your_app_module:app --reload --port 8000
4. Export Metrics to CSV
Once you have made some requests (with profiling active), open your web browser and navigate to:
http://localhost:8000/profiler/metrics.csv
This will trigger a direct download of a CSV file (e.g., fastapi_profile_metrics.csv) containing your collected profiling data.
5. Import into Google Sheets
- Go to Google Sheets (or your preferred spreadsheet software).
- Go to File > Import > Upload.
- Choose the downloaded fastapi_profile_metrics.csv file.
- Ensure "Detect automatically" is selected for the separator type (usually the default).
- Click "Import data".
Your profiling metrics will now be available in a clean, tabular format for analysis!
Columns in the CSV Export
The exported CSV file will include the following columns:
- Timestamp: The exact time the request completed (YYYY-MM-DD HH:MM:SS).
- RequestPath: The URL path of the API endpoint (e.g., /items/{item_id}).
- HTTPMethod: The HTTP method used for the request (e.g., GET, POST).
- StatusCode: The HTTP response status code (e.g., 200, 404, 500).
- TotalTimeMs: The total "wall clock" time for the request-response cycle in milliseconds.
- CPUTimeMs: The actual CPU time spent processing the request in milliseconds, as reported by pyinstrument. This excludes time spent waiting on I/O.
Contributing
Contributions are welcome! If you find bugs, have feature requests, or want to improve the code, please feel free to open issues or submit pull requests on the GitHub repository.
License
This project is licensed under the MIT License - see the LICENSE file for details.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file fastapi_simple_profiler-0.1.0.tar.gz.
File metadata
- Download URL: fastapi_simple_profiler-0.1.0.tar.gz
- Upload date:
- Size: 9.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.0
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
92a5bd6ac0da93bc8c94effc5e6f9821b7e86525d642a18ff2eab26adc7ef8cc
|
|
| MD5 |
25913ad006396cbe6505ad8515a9479b
|
|
| BLAKE2b-256 |
9f50aeecc9aaa59531751e6922e98de20c7b50d390f1fba172f3bfcfff1de953
|
File details
Details for the file fastapi_simple_profiler-0.1.0-py3-none-any.whl.
File metadata
- Download URL: fastapi_simple_profiler-0.1.0-py3-none-any.whl
- Upload date:
- Size: 9.5 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.0
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
6d0916be774efb9a58d4c3d26679ee3b44845dfdbf3df4e36abced59b4205b92
|
|
| MD5 |
33d490a35649c061e36e77749c185f83
|
|
| BLAKE2b-256 |
d45a64406e8d910ec04ed1c22e0612426a52c373fcba4bd78c0fc89ecf1d0de4
|