Mistral Voxtral plugin for the cjm-transcription-plugin-system library - provides local speech-to-text transcription through vLLM with configurable model selection and parameter control.
Project description
cjm-transcription-plugin-voxtral-vllm
Install
pip install cjm_transcription_plugin_voxtral_vllm
Project Structure
nbs/
├── meta.ipynb # Metadata introspection for the Voxtral VLLM plugin used by cjm-ctl to generate the registration manifest.
└── plugin.ipynb # Plugin implementation for Mistral Voxtral transcription through vLLM server
Total: 2 notebooks
Module Dependencies
graph LR
meta[meta<br/>Metadata]
plugin[plugin<br/>Voxtral VLLM Plugin]
plugin --> meta
1 cross-module dependencies detected
CLI Reference
No CLI commands found in this project.
Module Overview
Detailed documentation for each module in the project:
Metadata (meta.ipynb)
Metadata introspection for the Voxtral VLLM plugin used by cjm-ctl to generate the registration manifest.
Import
from cjm_transcription_plugin_voxtral_vllm.meta import (
get_plugin_metadata
)
Functions
def get_plugin_metadata() -> Dict[str, Any]: # Plugin metadata for manifest generation
"""Return metadata required to register this plugin with the PluginManager."""
# Calculate default DB path relative to the environment
base_path = os.path.dirname(os.path.dirname(sys.executable))
data_dir = os.path.join(base_path, "data")
db_path = os.path.join(data_dir, "voxtral_vllm_transcriptions.db")
# Ensure data directory exists
os.makedirs(data_dir, exist_ok=True)
return {
"name": "cjm-transcription-plugin-voxtral-vllm",
"Return metadata required to register this plugin with the PluginManager."
Voxtral VLLM Plugin (plugin.ipynb)
Plugin implementation for Mistral Voxtral transcription through vLLM server
Import
from cjm_transcription_plugin_voxtral_vllm.plugin import (
VLLMServer,
VoxtralVLLMPluginConfig,
VoxtralVLLMPlugin
)
Functions
@patch
def supports_streaming(
self: VoxtralVLLMPlugin # The plugin instance
) -> bool: # True if streaming is supported
"Check if this plugin supports streaming transcription."
@patch
def execute_stream(
self: VoxtralVLLMPlugin, # The plugin instance
audio: Union[AudioData, str, Path], # Audio data or path to audio file
**kwargs # Additional plugin-specific parameters
) -> Generator[str, None, TranscriptionResult]: # Yields text chunks, returns final result
"Stream transcription results chunk by chunk."
Classes
class VLLMServer:
def __init__(
self,
model: str = "mistralai/Voxtral-Mini-3B-2507", # Model name to serve
port: int = 8000, # Port for the server
host: str = "0.0.0.0", # Host address to bind to
gpu_memory_utilization: float = 0.85, # Fraction of GPU memory to use
log_level: str = "INFO", # Logging level (DEBUG, INFO, WARNING, ERROR)
capture_logs: bool = True, # Whether to capture and display server logs
**kwargs # Additional vLLM server arguments
)
"vLLM server manager for Voxtral models."
def __init__(
self,
model: str = "mistralai/Voxtral-Mini-3B-2507", # Model name to serve
port: int = 8000, # Port for the server
host: str = "0.0.0.0", # Host address to bind to
gpu_memory_utilization: float = 0.85, # Fraction of GPU memory to use
log_level: str = "INFO", # Logging level (DEBUG, INFO, WARNING, ERROR)
capture_logs: bool = True, # Whether to capture and display server logs
**kwargs # Additional vLLM server arguments
)
def add_log_callback(
self,
callback: Callable[[str], None] # Function that receives log line strings
) -> None: # Returns nothing
"Add a callback function to receive each log line."
def start(
self,
wait_for_ready: bool = True, # Wait for server to be ready before returning
timeout: int = 120, # Maximum seconds to wait for server readiness
show_progress: bool = True # Show progress indicators during startup
) -> None: # Returns nothing
"Start the vLLM server."
def stop(self) -> None: # Returns nothing
"""Stop the vLLM server."""
if self.process and self.process.poll() is None
"Stop the vLLM server."
def restart(self) -> None: # Returns nothing
"""Restart the server."""
self.stop()
time.sleep(2)
self.start()
def is_running(self) -> bool: # True if server is running and responsive
"Restart the server."
def is_running(self) -> bool: # True if server is running and responsive
"Check if server is running and responsive."
def get_recent_logs(
self,
n: int = 100 # Number of recent log lines to retrieve
) -> List[str]: # List of recent log lines
"Get the most recent n log lines."
def get_metrics_from_logs(self) -> dict: # Dictionary with performance metrics
"""Parse recent logs to extract performance metrics."""
metrics = {
"prompt_throughput": 0.0,
"Parse recent logs to extract performance metrics."
def tail_logs(
self,
follow: bool = True, # Continue displaying new logs as they arrive
n: int = 10 # Number of initial lines to display
) -> None: # Returns nothing
"Tail the server logs (similar to tail -f)."
@dataclass
class VoxtralVLLMPluginConfig:
"Configuration for Voxtral VLLM transcription plugin."
model_id: str = field(...)
device: str = field(...)
server_mode: str = field(...)
server_url: str = field(...)
server_port: int = field(...)
gpu_memory_utilization: float = field(...)
max_model_len: int = field(...)
language: Optional[str] = field(...)
temperature: float = field(...)
streaming: bool = field(...)
server_startup_timeout: int = field(...)
auto_start_server: bool = field(...)
capture_server_logs: bool = field(...)
dtype: str = field(...)
tensor_parallel_size: int = field(...)
class VoxtralVLLMPlugin:
def __init__(self):
"""Initialize the Voxtral VLLM plugin with default configuration."""
self.logger = logging.getLogger(f"{__name__}.{type(self).__name__}")
self.config: VoxtralVLLMPluginConfig = None
"Mistral Voxtral transcription plugin via vLLM server."
def __init__(self):
"""Initialize the Voxtral VLLM plugin with default configuration."""
self.logger = logging.getLogger(f"{__name__}.{type(self).__name__}")
self.config: VoxtralVLLMPluginConfig = None
"Initialize the Voxtral VLLM plugin with default configuration."
def name(self) -> str: # The plugin name identifier
"""Get the plugin name identifier."""
return "voxtral_vllm"
@property
def version(self) -> str: # The plugin version string
"Get the plugin name identifier."
def version(self) -> str: # The plugin version string
"""Get the plugin version string."""
return "1.0.0"
@property
def supported_formats(self) -> List[str]: # List of supported audio formats
"Get the plugin version string."
def supported_formats(self) -> List[str]: # List of supported audio formats
"""Get the list of supported audio file formats."""
return ["wav", "mp3", "flac", "m4a", "ogg", "webm", "mp4", "avi", "mov"]
def get_current_config(self) -> Dict[str, Any]: # Current configuration as dictionary
"Get the list of supported audio file formats."
def get_current_config(self) -> Dict[str, Any]: # Current configuration as dictionary
"""Return current configuration state."""
if not self.config
"Return current configuration state."
def get_config_schema(self) -> Dict[str, Any]: # JSON Schema for configuration
"""Return JSON Schema for UI generation."""
return dataclass_to_jsonschema(VoxtralVLLMPluginConfig)
@staticmethod
def get_config_dataclass() -> VoxtralVLLMPluginConfig: # Configuration dataclass
"Return JSON Schema for UI generation."
def get_config_dataclass() -> VoxtralVLLMPluginConfig: # Configuration dataclass
"""Return dataclass describing the plugin's configuration options."""
return VoxtralVLLMPluginConfig
def initialize(
self,
config: Optional[Any] = None # Configuration dataclass, dict, or None
) -> None
"Return dataclass describing the plugin's configuration options."
def initialize(
self,
config: Optional[Any] = None # Configuration dataclass, dict, or None
) -> None
"Initialize or re-configure the plugin (idempotent)."
def execute(
self,
audio: Union[AudioData, str, Path], # Audio data or path to audio file to transcribe
**kwargs # Additional arguments to override config
) -> TranscriptionResult: # Transcription result with text and metadata
"Transcribe audio using Voxtral via vLLM."
def is_available(self) -> bool: # True if vLLM and dependencies are available
"""Check if vLLM and required dependencies are available."""
if not OPENAI_AVAILABLE
"Check if vLLM and required dependencies are available."
def cleanup(self) -> None:
"""Clean up resources."""
self.logger.info("Cleaning up Voxtral VLLM plugin")
# Stop managed server if running
if self.config and self.config.server_mode == "managed" and self.server
"Clean up resources."
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file cjm_transcription_plugin_voxtral_vllm-0.0.10.tar.gz.
File metadata
- Download URL: cjm_transcription_plugin_voxtral_vllm-0.0.10.tar.gz
- Upload date:
- Size: 22.1 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.11.13
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
b6b3559032721de601e45d03ad21c667325d0ab432fa26d5b504b24db1328639
|
|
| MD5 |
b4c94c4c775e8f8289e0dd641bbe4137
|
|
| BLAKE2b-256 |
797c53a5f8a915ebc44316aa9c795638847ae2785a3c88f7360547d4db57a1eb
|
File details
Details for the file cjm_transcription_plugin_voxtral_vllm-0.0.10-py3-none-any.whl.
File metadata
- Download URL: cjm_transcription_plugin_voxtral_vllm-0.0.10-py3-none-any.whl
- Upload date:
- Size: 20.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.11.13
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
77dd2dc004a8a328f59dec09bcc9f9b94f041b362ace12f5bb790e2845e2aa6d
|
|
| MD5 |
1e6d21f8d49fb825787ba2d9c98160cd
|
|
| BLAKE2b-256 |
50968fc81f870e5311f2dc14a674fee4ce66224948688798656d1fa9fe6a436a
|