MCP server for LLM training, fine-tuning, and experimentation
Project description
LLM MCP Server
MCP server for LLM training, fine-tuning, and experimentation. Part of the scicomp-mcp suite.
Features
- Model Architectures: GPT (Transformer decoder) and Mamba (State Space Model)
- Tokenizers: tiktoken, BPE, SentencePiece, character-level
- Training: AdamW, learning rate scheduling, gradient checkpointing, mixed precision
- Evaluation: Perplexity, loss, text generation
Installation
uv sync --all-extras
Usage
scicomp-llm-mcp
Tools
Model Management
create_model- Create GPT or Mamba architectureget_model_config- Get model configurationlist_models- List all models
Tokenizers
create_tokenizer- Create or load tokenizertokenize_text- Tokenize text
Datasets
load_dataset- Load training datasetprepare_dataset- Prepare for training
Training
create_trainer- Configure trainingtrain_step- Execute training stepsget_training_status- Monitor progress
Evaluation
evaluate_model- Evaluate on datasetgenerate_text- Generate textcompute_perplexity- Compute perplexity
Checkpoints
save_checkpoint- Save model checkpointload_checkpoint- Load from checkpoint
Analysis
analyze_attention- Analyze attention patternscompute_gradient_norms- Compute gradient normsestimate_memory- Estimate training memory requirementscompute_model_flops- Compute model FLOPsanalyze_weights- Analyze weight distributionsanalyze_sparsity- Compute model sparsityanalyze_norms- Analyze layer normscompare_models- Compare model architectures
Dataset Ablation
analyze_data_influence- Compute sample influenceanalyze_token_distribution- Analyze token frequenciesanalyze_sequences- Compute sequence statisticsrun_data_ablation- Run ablation studiessuggest_augmentations- Suggest data augmentation strategies
Attention Visualization
visualize_attention- Extract attention summaryanalyze_attention_patterns- Detect attention patternscompute_head_rankings- Rank heads by importancecompare_heads- Compare attention heads
License
MIT
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file scicomp_llm_mcp-0.1.6.tar.gz.
File metadata
- Download URL: scicomp_llm_mcp-0.1.6.tar.gz
- Upload date:
- Size: 33.7 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
fe6a7f91227429dfddf9732a4b7930f7fc1b2f816304205cf3887dfaa8d12a2f
|
|
| MD5 |
bab1f7c2df671ddc095fa55f686677b8
|
|
| BLAKE2b-256 |
8c3fe07508c8ce2c13798ea7ab41cc6758afce4dae8f37139b6e5435cc3d978a
|
Provenance
The following attestation bundles were made for scicomp_llm_mcp-0.1.6.tar.gz:
Publisher:
publish.yml on andylbrummer/math-mcp
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
scicomp_llm_mcp-0.1.6.tar.gz -
Subject digest:
fe6a7f91227429dfddf9732a4b7930f7fc1b2f816304205cf3887dfaa8d12a2f - Sigstore transparency entry: 791289756
- Sigstore integration time:
-
Permalink:
andylbrummer/math-mcp@6e17f969bf98556277f43d38b5fe933f35af8798 -
Branch / Tag:
refs/tags/v0.1.6 - Owner: https://github.com/andylbrummer
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@6e17f969bf98556277f43d38b5fe933f35af8798 -
Trigger Event:
push
-
Statement type:
File details
Details for the file scicomp_llm_mcp-0.1.6-py3-none-any.whl.
File metadata
- Download URL: scicomp_llm_mcp-0.1.6-py3-none-any.whl
- Upload date:
- Size: 32.5 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
c3d7044115157e99721689179a2bc3f3eed9b07a15e22ca9a45104d0f475c93e
|
|
| MD5 |
f81dcab08405344f2972e779c1125972
|
|
| BLAKE2b-256 |
39a54f0155aba10fcd09d6ebc9025024cd8c07decc44494be250052deb63a564
|
Provenance
The following attestation bundles were made for scicomp_llm_mcp-0.1.6-py3-none-any.whl:
Publisher:
publish.yml on andylbrummer/math-mcp
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
scicomp_llm_mcp-0.1.6-py3-none-any.whl -
Subject digest:
c3d7044115157e99721689179a2bc3f3eed9b07a15e22ca9a45104d0f475c93e - Sigstore transparency entry: 791289760
- Sigstore integration time:
-
Permalink:
andylbrummer/math-mcp@6e17f969bf98556277f43d38b5fe933f35af8798 -
Branch / Tag:
refs/tags/v0.1.6 - Owner: https://github.com/andylbrummer
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@6e17f969bf98556277f43d38b5fe933f35af8798 -
Trigger Event:
push
-
Statement type: