GitHub Copilot API Proxy - A Flask application serving as a proxy server for GitHub Copilot API
Project description
GitHub Copilot API Proxy (ghc-api)
A Python Flask application that serves as a proxy server for GitHub Copilot API, providing OpenAI and Anthropic API compatibility with caching and monitoring capabilities.
Features
- OpenAI API Compatibility:
/v1/chat/completionsendpoint - Anthropic API Compatibility:
/v1/messagesendpoint with automatic request/response translation - Model Listing:
/v1/modelsendpoint listing available models - Model Name Mapping: Translate model names with exact and prefix-based matching
- Token Management: Automatic GitHub Copilot token refresh
- Vision Support: Handle image inputs and enable vision capabilities
- Memory Caching: Cache all requests and responses (up to 1000 entries)
- Web Dashboard: Real-time statistics and request browser
- Request Details: View full request/response bodies with JSON formatting
- Export/Import: Export and import request history as JSON Lines files
- Optional Request File Logging: Save completed requests to daily JSON Lines files
- Content Filtering: Remove or add content from system prompts and tool results
- Code Agent Manager UI: Install Codex/Claude/Copilot CLI and manage config sync from dashboard
- Config Sync: Sync Claude Code, Codex, and ghc-api config files with OneDrive
- Safe Backups: Auto backup overwritten config files as
*.YYYYMMDD_HHMMSS.bak - Machine Token Usage Logs: Periodic token usage JSONL per machine with cross-machine overview in dashboard
Installation
Install the package using pip:
pip install ghc-api
Or install from source:
pip install .
Usage
Start the server with the ghc-api command:
ghc-api
By default, the server will start on http://localhost:8313.
Command Line Options
-p PORTor--port PORT: Specify the port to listen on (default: 8313)-a ADDRESSor--address ADDRESS: Specify the address to listen on (default: localhost)-cor--config: Generate a YAML config file in~/.ghc-api/config.yaml-vor--version: Show version (for exampleghc-api 1.0.8)--help: Show help message
Configuration
The application looks for a configuration file at ~/.ghc-api/config.yaml. You can generate this file using:
ghc-api --config
The config file contains:
# Server Settings
address: localhost
port: 8313
debug: false
# GitHub Copilot Account Type
# Options: "individual", "business", "enterprise"
account_type: individual
# Version settings (used to build request headers)
vscode_version: "1.93.0"
api_version: "2025-04-01"
copilot_version: "0.26.7"
# Model Name Mappings
model_mappings:
# Exact match mappings
exact:
opus: claude-opus-4.5
sonnet: claude-sonnet-4.5
haiku: claude-haiku-4.5
# Prefix match mappings
prefix:
claude-sonnet-4-: claude-sonnet-4
claude-opus-4.5-: claude-opus-4.5
# Content Filtering
system_prompt_remove: [] # Strings to remove from system prompts
system_prompt_add: [] # Strings to append to system prompts
tool_result_suffix_remove: [] # Strings to remove from end of tool results
# Optional request persistence
save_request_to_file: false # If true, save completed requests to requests/YYYY-MM-DD.jl
# Optional OneDrive access gate
disable_onedrive_access: true # If true, skip all OneDrive detection/sync/shared reads
Token Management
The application follows this priority for getting the GitHub token:
GITHUB_TOKENenvironment variable- Token file at
~/.ghc-api/github_token.txt - Interactive GitHub Device Flow authentication
Config Sync and OneDrive
ghc-api can manage and sync these files:
- Claude Code:
~/.claude/settings.json - Codex:
~/.codex/config.toml - ghc-api:
~/.ghc-api/config.yaml(or%APPDATA%/ghc-api/config.yamlon Windows)
OneDrive detection priority:
~/OneDrive - *~/OneDrive- In WSL:
/mnt/c/Users/<username>/OneDrive - *then/mnt/c/Users/<username>/OneDrive
To disable all OneDrive-dependent operations, set disable_onedrive_access: true in config.yaml.
When enabled, ghc-api skips OneDrive detection, config sync actions, and shared OneDrive hash reads.
Sync target folder:
.ghc-api/configSyncunder detected OneDrive root
Machine folder:
.ghc-api/agents/{hostname}_{os}whereosisWin,Linux, orWSL
Hash files:
.ghc-api/configSync/config.sha1.ghc-api/agents/{hostname}_{os}/ghc-api/config.sha1
Hashes are recalculated when local config file timestamp is newer than the hash file. On startup, ghc-api checks synced files and prints config differences to stdout (and UI indicator if different).
Token Usage Logging
Every 5 minutes, ghc-api writes token usage delta (if non-zero) to:
- OneDrive mode:
.ghc-api/agents/{hostname}_{os}/token_usage.jl - Fallback when OneDrive is unavailable:
~/.ghc-api/token_usage.jl
Also flushes pending usage on shutdown (Ctrl+C/termination/normal exit).
Each JSONL line includes:
timestamp(unix seconds)modelslist with:modelrequest_countinput_tokensoutput_tokenstotal_tokens
Request File Logging
When save_request_to_file: true, ghc-api appends each completed request to:
<ghc-api config dir>/requests/YYYY-MM-DD.jl
The saved .jl line format is the same as dashboard export (/api/requests/export) and can be imported by dashboard import (/api/requests/import).
API Endpoints
OpenAI Compatible
POST /v1/chat/completions- Chat completionsPOST /chat/completions- Chat completions (without v1 prefix)GET /v1/models- List available modelsGET /models- List available models (without v1 prefix)
Anthropic Compatible
POST /v1/messages- Messages API (Anthropic format)
Dashboard & Monitoring
GET /- Web dashboard with statisticsGET /requests- Request browser pageGET /api/runtime-config- Read in-memory runtime configPOST /api/runtime-config- Update in-memory runtime config (no file write)GET /api/stats- JSON statistics endpointGET /api/requests- Paginated list of requestsGET /api/requests/search- Full-text search in request/response bodiesGET /api/requests/export- Export all requests as JSON Lines filePOST /api/requests/import- Import requests from JSON Lines fileGET /api/request/<id>- Individual request detailsGET /api/request/<id>/request-body- Request body onlyGET /api/request/<id>/response-body- Response body onlyGET /api/config-manager/status- Config manager status and diff infoPOST /api/config-manager/install-tools- Install Codex/Claude/Copilot CLIPOST /api/config-manager/sync-to-onedrive- Sync local config to OneDrivePOST /api/config-manager/sync-from-onedrive- Copy OneDrive config to local machine with backupsGET /api/config-manager/token-usage?range=all|day|week|month- Cross-machine token usage overviewGET /api/config-manager/config-hashes- Config hash overview for shared OneDrive and each machine (with create time)
Example Usage
With OpenAI Python SDK
from openai import OpenAI
client = OpenAI(
base_url="http://localhost:8313/v1",
api_key="not-needed" # Token is managed by the proxy
)
response = client.chat.completions.create(
model="gpt-4o",
messages=[{"role": "user", "content": "Hello!"}]
)
print(response.choices[0].message.content)
With Anthropic Python SDK
import anthropic
client = anthropic.Anthropic(
base_url="http://localhost:8313",
api_key="not-needed" # Token is managed by the proxy
)
message = client.messages.create(
model="claude-sonnet-4",
max_tokens=1024,
messages=[{"role": "user", "content": "Hello!"}]
)
print(message.content[0].text)
With cURL
# Chat completions
curl http://localhost:8313/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "gpt-4o",
"messages": [{"role": "user", "content": "Hello!"}]
}'
# List models
curl http://localhost:8313/v1/models
Dashboard
Access the web dashboard at http://localhost:8313/ to:
- View overall statistics (total requests, data transfer)
- See per-model usage statistics
- See per-endpoint analytics
- Browse recent requests
- View detailed request/response bodies
- Use Code Agent Manager to:
- Install code-agent CLIs
- Sync config files to/from OneDrive
- See config mismatch alerts
- View token usage overview by machine/model with time-range and machine filters
- View config hash overview by machine and shared OneDrive hash with create times
Architecture
- Modular Design: Organized into separate modules for maintainability
main.py- Entry point and configuration loadingapp.py- Flask application factoryconfig.py- Configuration constants and model mappingscache.py- Request caching and statisticstranslator.py- OpenAI/Anthropic format translationstreaming.py- Streaming response handlingtoken_manager.py- GitHub token managementroutes/- API endpoint handlers (openai, anthropic, dashboard)
- Thread-Safe Caching: Uses threading locks for concurrent access
- Memory-Based Storage: No external database dependencies
- RESTful API Design: Follows REST conventions
License
MIT License
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file ghc_api-1.0.14.tar.gz.
File metadata
- Download URL: ghc_api-1.0.14.tar.gz
- Upload date:
- Size: 85.3 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
07b0d784fa6ab809832f21c940eb9654ff6bf46c4c2c5b9e2eb619bd53721c57
|
|
| MD5 |
5c669c113f23fc539e41334f25a0bbf3
|
|
| BLAKE2b-256 |
e507df6bc6c09915330fef16b9c68b3e2e38ac3b64f603eba9b490b850619c99
|
File details
Details for the file ghc_api-1.0.14-py3-none-any.whl.
File metadata
- Download URL: ghc_api-1.0.14-py3-none-any.whl
- Upload date:
- Size: 88.8 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
ece271e976bf1a89ecb1c186939d1f19cb0cf9683ee086bcf3aaf561d5a7368b
|
|
| MD5 |
738680d6b00afbd97409b09d81946a87
|
|
| BLAKE2b-256 |
24e5dee78025d42b52d005af2997ffb6bc9bbacb4693f440ad226a9a03a1786b
|