TokenScope: Token-Aware Directory Explorer for LLMs
Project description
TokenScope
TokenScope is a token-aware directory explorer for Large Language Models.
A Model Context Protocol (MCP) server for token-aware directory exploration and analysis, designed for Large Language Models (LLMs).
Overview
TokenScope provides intelligent directory structure analysis and token-aware file content exploration. It helps LLMs like Claude understand codebases and directory structures by:
- Scanning directory structures with token-efficient summaries
- Extracting and analyzing file contents with token awareness
- Finding important files for codebase understanding
- Generating reports with relevant information
Features
-
Token-Aware Directory Scanning
- Explores directories recursively with configurable depth
- Provides intelligent summaries for large directories
- Respects .gitignore files and custom ignore patterns
-
File Content Analysis
- Smart extraction of file contents that respects token limits
- Special handling for JSON and other structured files
- File selection prioritization based on importance
-
Token Usage Statistics
- Estimates tokens required to process directories
- Breaks down token usage by file extension
- Identifies token-heavy files
-
Comprehensive Reporting
- Generates markdown reports with directory structure
- Includes token usage statistics
- Shows samples of important files
-
Security Features
- Path validation to restrict operations to a specified base directory
- Prevents access to files outside the allowed base path
Installation
Prerequisites
- Python 3.10 or higher
- uv (recommended for easy dependency management)
1. Main Installation (PyPI)
This is the recommended method for most users who just want to use TokenScope:
# Install from PyPI using uv (recommended)
uv pip install tokenscope
Running TokenScope
The --base-path argument is mandatory for security reasons. It restricts all file operations to the specified directory.
# Run using the installed package
uv run --with tokenscope tokenscope --base-path /path/to/allowed/directory
Configuring in Claude Desktop
-
Locate Claude Desktop's configuration file (typically in
~/.config/claude/config.json) -
Add TokenScope to the
mcpServerssection:"mcpServers": { "TokenScope": { "command": "uv", "args": [ "run", "--with", "tokenscope", "tokenscope", "--base-path", "/your/secure/base/path" ] } }
-
Replace
/your/secure/base/pathwith the directory you want to restrict operations to -
Save the configuration file and restart Claude Desktop
2. Development Installation (from GitHub)
For contributors or users who want to modify the code:
# Clone the repository
git clone https://github.com/cdgaete/token-scope-mcp.git
cd token-scope-mcp
# Install development dependencies with uv
uv pip install -e ".[dev]"
Running in Development Mode
# Run the server directly with uv
uv run --with fastmcp --with tiktoken src/server.py --base-path /path/to/allowed/directory
Configuring Development Version in Claude Desktop
-
Locate Claude Desktop's configuration file
-
Add TokenScope to the
mcpServerssection with development paths:"mcpServers": { "TokenScope (Dev)": { "command": "uv", "args": [ "run", "--with", "fastmcp", "--with", "tiktoken", "/path/to/your/token-scope-mcp/src/server.py", "--base-path", "/your/secure/base/path" ] } }
-
Replace
/path/to/your/token-scope-mcp/src/server.pywith the actual path to the server.py file -
Replace
/your/secure/base/pathwith your secure directory
Security Features
The --base-path argument is mandatory for security reasons:
- All file operations are validated to ensure they're within the specified directory
- Attempts to access or modify files outside the base path will be rejected
- The base path is set once when starting the server and cannot be changed without restart
Example Prompts
Here are some examples of how to use TokenScope with Claude:
Please scan my project directory at /path/to/project and tell me about its structure, focusing on the most important files.
Analyze the token usage in my project directory at /path/to/project and tell me how many tokens would be needed to process the entire codebase with an LLM.
Generate a comprehensive directory report about my project at /path/to/project, including structure, token statistics, and samples of the most important files.
Available Tools
The server provides the following MCP tools:
scan_directory_structure
Scans a directory and returns its structure in a token-efficient way.
scan_directory_structure(
path: str,
depth: int = 3,
max_tokens: int = 10000,
ignore_patterns: list[str] | None = None,
include_gitignore: bool = True,
include_default_ignores: bool = True
)
extract_file_content
Extracts the content of a specific file, respecting token limits and format.
extract_file_content(
file_path: str,
max_tokens: int = 10000,
sample_only: bool = False
)
search_files_by_pattern
Searches for files matching specified patterns within a directory structure.
search_files_by_pattern(
directory: str,
patterns: list[str],
max_depth: int = 5,
include_content: bool = False,
max_files: int = 100,
max_tokens_per_file: int = 1000,
sample_only: bool = False,
ignore_patterns: list[str] | None = None,
include_gitignore: bool = True,
include_default_ignores: bool = True
)
analyze_token_usage
Analyzes token usage for a directory or file to estimate LLM processing requirements.
analyze_token_usage(
path: str,
include_file_details: bool = False,
ignore_patterns: list[str] | None = None,
include_gitignore: bool = True,
include_default_ignores: bool = True
)
generate_directory_report
Generates a comprehensive markdown report about a directory with token statistics.
generate_directory_report(
directory: str,
depth: int = 3,
include_file_content: bool = True,
max_files_with_content: int = 5,
max_tokens_per_file: int = 1000,
sample_only: bool = False,
ignore_patterns: list[str] | None = None,
include_gitignore: bool = True,
include_default_ignores: bool = True
)
copy_file_to_destination
Copy a file from source path to destination path.
copy_file_to_destination(
source_path: str,
destination_path: str
)
Default Ignore Patterns
TokenScope automatically ignores common directories and files:
DEFAULT_IGNORE_PATTERNS = [
".git/",
".venv/",
"venv/",
"__pycache__/",
"node_modules/",
".pytest_cache/",
".ipynb_checkpoints/",
".DS_Store",
"*.pyc",
"*.pyo",
"*.pyd",
"*.so",
"*.dll",
"*.class",
"build/",
"dist/",
"*.egg-info/",
".tox/",
".coverage",
".idea/",
".vscode/",
".mypy_cache/",
]
License
This project is licensed under the MIT License - see the LICENSE file for details.
Acknowledgments
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file tokenscope-0.1.1.tar.gz.
File metadata
- Download URL: tokenscope-0.1.1.tar.gz
- Upload date:
- Size: 23.8 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.1
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
d09c5777ab1554a1a3cfd049432bb9eee1c4444d9cf7f3b168f3983567430ce9
|
|
| MD5 |
29630a2c3fc176a734198167f691b439
|
|
| BLAKE2b-256 |
5c9333dc367639c875a3267df9d3de0daad8e9403a66e0dbafcd2df6cf8a4583
|
File details
Details for the file tokenscope-0.1.1-py3-none-any.whl.
File metadata
- Download URL: tokenscope-0.1.1-py3-none-any.whl
- Upload date:
- Size: 22.8 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.1
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
620f42cefe7fc15051d1d5dc86a28648faca710b5fb80cc4b280bd2c0abb3517
|
|
| MD5 |
4ba39350bb74b081e7b7f7a307e0bd8d
|
|
| BLAKE2b-256 |
0e538ae975f018471e33b351e7b3658a79b6159fec2ae1f3c1732ebe8f1f60bc
|