Skip to main content

Provides data acquisition and processing assets shared between Sollertia platform libraries.

Project description

sollertia-shared-assets

Provides data acquisition and processing assets shared between Sollertia platform libraries.

PyPI - Version PyPI - Python Version uv Ruff type-checked: mypy PyPI - License PyPI - Status PyPI - Wheel


Detailed Description

This library is part of the Sollertia AI-assisted scientific data acquisition and processing platform, built on the Ataraxis framework and developed in the Sun (NeuroAI) lab at Cornell University. It keeps the two main Sollertia libraries used for data acquisition (sollertia-experiment) and processing (sollertia-forgery) independent of each other by providing the shared assets both depend on.

The library stores dataclasses used to save data acquired with the Sollertia platform (sessions, subjects, hardware state) and configure data acquisition and processing runtimes. It also provides a CLI (slsa) for platform configuration and an MCP server with tools for agentic configuration management, session operations, and Unity Editor integration. A subset of the MCP tools relay commands to a running Unity Editor instance via the McpBridge plugin from sollertia-unity-tasks, enabling agents to generate task prefabs, manage scenes, and control Play Mode.


Table of Contents


Dependencies

For users, all library dependencies are installed automatically by all supported installation methods. For developers, see the Developers section for information on installing additional development dependencies.


Installation

Source

Note, installation from source is highly discouraged for anyone who is not an active project developer.

  1. Download this repository to the local machine using the preferred method, such as git-cloning. Use one of the stable releases that include precompiled binary and source code distribution (sdist) wheels.
  2. If the downloaded distribution is stored as a compressed archive, unpack it using the appropriate decompression tool.
  3. cd to the root directory of the prepared project distribution.
  4. Run pip install . to install the project and its dependencies.

pip

Use the following command to install the library and all of its dependencies via pip: pip install sollertia-shared-assets


Usage

Most library components are intended to be used via other Sollertia platform libraries. For details on using shared assets for data acquisition and preprocessing, see the sollertia-experiment library. For details on using shared assets for data processing and dataset formation, see the sollertia-forgery library.

Warning! End users should not use any component of this library directly or install this library into any Python environment. All assets from this library are intended to be used exclusively by developers working on other Sollertia platform libraries.

CLI Commands

This library provides the slsa CLI that exposes the following commands and command groups:

Command Description
mcp Starts the MCP server for agentic configuration management
configure directory Sets the local Sollertia platform working directory
configure google Sets the path to the Google service account credentials file
configure templates Sets the path to the sollertia-unity-tasks task templates directory
configure project Creates a project directory structure for data acquisition
configure experiment Creates an experiment configuration from a task template

Use slsa --help, slsa configure --help, or slsa COMMAND --help for detailed usage information.

MCP Server

This library provides an MCP server that exposes configuration management, session and dataset operations, and Unity Editor relay tools for AI agent integration. The server enables agents to query and configure shared Sollertia platform workflow components.

Starting the Server

Start the MCP server using the CLI:

slsa mcp

Available Tools

Tool Description
create_experiment_config_tool Creates an experiment configuration from a task template using sensible defaults
create_scene_tool Creates a new Unity scene by copying ExperimentTemplate and optionally adding a task prefab
describe_experiment_configuration_schema_tool Returns the schema for the experiment configuration of a given acquisition system
describe_session_data_schema_tool Returns the schema for the SessionData dataclass
describe_session_descriptor_schema_tool Returns the schema for the descriptor associated with a given session type
describe_session_hardware_state_schema_tool Returns the hardware-state schema for a given acquisition system
describe_surgery_data_schema_tool Returns the schema for SurgeryData and its nested subclasses
describe_template_schema_tool Returns the schema for TaskTemplate and nested Cue, Segment, TrialStructure, and VREnvironment
discover_experiments_tool Discovers all experiment configuration YAML files under the data root
discover_templates_tool Lists all task templates in the configured templates directory
enter_play_mode_tool Enters Play Mode in the Unity Editor
exit_play_mode_tool Exits Play Mode in the Unity Editor
generate_task_prefab_tool Generates a Task prefab in Unity from a YAML task template
get_data_root_overview_tool Builds the project/animal/session hierarchy from SessionData and per-session lifecycle status
get_platform_environment_status_tool Reports the status of the working directory, templates directory, and Google credentials
get_play_state_tool Returns the current Unity Editor play state and active scene name
inspect_prefab_tool Returns the full hierarchy, components, transforms, and collider details of a prefab
inspect_sessions_tool Produces a detailed health and inventory report for one or more sessions
list_scenes_tool Lists all Unity scene assets and identifies the currently active scene
list_supported_acquisition_systems_tool Enumerates the acquisition systems supported by the Sollertia platform
list_supported_session_types_tool Enumerates the session types supported by the Sollertia platform
list_supported_trial_types_tool Enumerates the trial classes supported by experiment configurations
list_supported_trigger_types_tool Enumerates the trigger type values supported by trial structures
list_unity_assets_tool Lists Unity assets of a given type within a search path
open_scene_tool Opens a Unity scene in the Editor
read_experiment_configuration_tool Loads an experiment configuration YAML (project source or per-session frozen snapshot)
read_google_credentials_tool Returns the configured path to the Google service account credentials file
read_session_data_tool Loads a session_data.yaml file via the SessionData schema (file-path based)
read_session_descriptor_tool Detects the appropriate descriptor class and loads the descriptor YAML
read_session_hardware_state_tool Loads a hardware-state YAML for a session using the class for the given acquisition system
read_surgery_data_tool Loads the full SurgeryData payload from a session's raw_data/surgery_metadata.yaml snapshot
read_task_templates_directory_tool Returns the configured path to the task templates directory
read_template_tool Loads a TaskTemplate YAML by name from the configured templates directory
read_working_directory_tool Returns the configured Sollertia platform working directory path
set_google_credentials_tool Sets the path to the Google service account credentials file
set_task_templates_directory_tool Sets the path to the task templates directory
set_working_directory_tool Sets the local Sollertia platform working directory
validate_experiment_configuration_tool Validates an experiment configuration YAML for a project
validate_prefab_against_template_tool Validates that Unity prefab zone positions match the template configuration
validate_template_tool Validates a TaskTemplate against its schema and cross-reference constraints
write_experiment_configuration_tool Creates or replaces an experiment configuration YAML for a project
write_session_data_tool Creates or replaces a session_data.yaml file, validated against the SessionData schema
write_session_descriptor_tool Creates or replaces a session descriptor YAML for a session
write_session_hardware_state_tool Creates or replaces a hardware-state YAML for a session using the class for the given acquisition system
write_template_tool Creates or replaces a TaskTemplate YAML in the configured templates directory

Note, tools that interact with Unity (create_scene_tool, enter_play_mode_tool, exit_play_mode_tool, generate_task_prefab_tool, get_play_state_tool, inspect_prefab_tool, list_scenes_tool, list_unity_assets_tool, open_scene_tool, validate_prefab_against_template_tool) require the Unity Editor to be running on the local machine with the McpBridge plugin from sollertia-unity-tasks active. These tools relay commands to the Editor via HTTP.

Client Registration

MCP server registration and Claude Code skill assets for this library are distributed through the sollertia marketplace as part of the configuration plugin. Install the plugin from the marketplace to automatically register the MCP server with compatible clients and make all associated skills available.


API Documentation

See the API documentation for the detailed description of the methods and classes exposed by components of this library.

Note, the API documentation includes additional details about the slsa CLI commands and their parameters beyond what is covered in the CLI Commands section above.


Developers

This section provides installation, dependency, and build-system instructions for the developers that want to modify the source code of this library.

Installing the Project

Note, this installation method requires mamba version 2.3.2 or above. Currently, all Sun lab automation pipelines require that mamba is installed through the miniforge3 installer.

  1. Download this repository to the local machine using the preferred method, such as git-cloning.
  2. If the downloaded distribution is stored as a compressed archive, unpack it using the appropriate decompression tool.
  3. cd to the root directory of the prepared project distribution.
  4. Install the core Sun lab development dependencies into the base mamba environment via the mamba install tox uv tox-uv command.
  5. Use the tox -e create command to create the project-specific development environment followed by tox -e install command to install the project into that environment as a library.

Additional Dependencies

In addition to installing the project and all user dependencies, install the following dependencies:

  1. Python distributions, one for each version supported by the developed project. Currently, this library supports Python 3.14 only. It is recommended to use a tool like pyenv to install and manage the required versions.

Development Automation

This project uses tox for development automation. The following tox environments are available:

Environment Description
lint Runs ruff formatting, ruff linting, and mypy type checking
stubs Generates py.typed marker and .pyi stub files
py314-test Runs the test suite via pytest for Python 3.14
coverage Aggregates test coverage into an HTML report
docs Builds the API documentation via Sphinx
build Builds sdist and wheel distributions
upload Uploads distributions to PyPI via twine
install Builds and installs the project into its mamba environment
uninstall Uninstalls the project from its mamba environment
create Creates the project's mamba development environment
remove Removes the project's mamba development environment
provision Recreates the mamba environment from scratch
export Exports the mamba environment as .yml and spec.txt files
import Creates or updates the mamba environment from a .yml file

Run any environment using tox -e ENVIRONMENT. For example, tox -e lint.

Note, all pull requests for this project have to successfully complete the tox task before being merged. To expedite the task's runtime, use the tox --parallel command to run some tasks in parallel.

Adding New Acquisition Systems

This library owns the shared vocabulary that identifies acquisition systems (the AcquisitionSystems enum) and the experiment configuration factory registry used to build per-system experiment configuration dataclasses from a TaskTemplate. System-level hardware and software configuration classes live in the acquisition runtime package (sollertia-experiment), not in this library. The following steps outline how to add support for a new acquisition system.

Step 1: Add the system to the AcquisitionSystems enum

In configuration/configuration_utilities.py, add a new entry to the AcquisitionSystems enum:

from enum import StrEnum
class AcquisitionSystems(StrEnum):
    MESOSCOPE_VR = "mesoscope"
    NEW_SYSTEM = "new_system"  # Add new system here

Step 2: Create the experiment configuration module

Create a new file (e.g., new_system_configuration.py) in configuration/ containing an experiment configuration dataclass inheriting from YamlConfig that captures the runtime experiment parameters for the new system. Use MesoscopeExperimentConfiguration in mesoscope_configuration.py as a reference.

Step 3: Update the factory registry

In configuration/configuration_utilities.py:

  1. Extend the ExperimentConfigFactory type alias so its return type includes the new experiment configuration class
  2. Implement a private factory function (e.g., _create_new_system_experiment_config) that builds the new experiment configuration dataclass from a TaskTemplate and the converted trial structures dictionary
  3. Register the factory in _experiment_config_factory_registry under the new AcquisitionSystems key

Step 4: Update downstream libraries

Coordinate changes with sollertia-experiment (which owns the system-level hardware/software configuration classes and the acquisition runtime) and sollertia-forgery (data processing) as needed.

AI-Assisted Development

Claude Code skills and AI development assets for this project are distributed through two marketplaces:

  • sollertia marketplace: Provides MCP server registration, configuration-specific skills for working directory management, system and experiment configuration, session data, subject metadata, dataset management, task templates, and MCP environment setup via the configuration plugin. Install this plugin to register the slsa mcp server with compatible MCP clients and make all configuration workflow skills available.
  • ataraxis marketplace: Provides shared development skills that enforce Sun Lab coding conventions (Python style, README style, commit messages, pyproject.toml, tox configuration) and general-purpose codebase exploration tools via the automation plugin.

Install both marketplace plugins to make all associated skills and development tools available to compatible AI coding agents.

Automation Troubleshooting

Many packages used in tox automation pipelines (uv, mypy, ruff) and tox itself may experience runtime failures. In most cases, this is related to their caching behavior. If an unintelligible error is encountered with any of the automation components, deleting the corresponding cache directories (.tox, .ruff_cache, .mypy_cache, etc.) manually or via a CLI command typically resolves the issue.


Versioning

This project uses semantic versioning. See the tags on this repository for the available project releases.


Authors


License

This project is licensed under the Apache 2.0 License: see the LICENSE file for details.


Acknowledgments

  • All Sun lab members for providing the inspiration and comments during the development of this library.
  • The creators of all other dependencies and projects listed in the pyproject.toml file.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

sollertia_shared_assets-8.0.0rc7.tar.gz (79.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

sollertia_shared_assets-8.0.0rc7-py3-none-any.whl (65.9 kB view details)

Uploaded Python 3

File details

Details for the file sollertia_shared_assets-8.0.0rc7.tar.gz.

File metadata

File hashes

Hashes for sollertia_shared_assets-8.0.0rc7.tar.gz
Algorithm Hash digest
SHA256 7472e5c32343c6f6f546f4fc162891b8bd40fee3ac8214cbc2ef3bc0f778632b
MD5 c15dc9f23e984ca4678f8f93c40ec08c
BLAKE2b-256 cfc43ae1cb66773254bc023ad78066bd54adc28bd6b6f1fae77864b0365ddbcd

See more details on using hashes here.

File details

Details for the file sollertia_shared_assets-8.0.0rc7-py3-none-any.whl.

File metadata

File hashes

Hashes for sollertia_shared_assets-8.0.0rc7-py3-none-any.whl
Algorithm Hash digest
SHA256 f07fc0b00c35a86ef3c044d822503126bbfcb8b5524d93ab36a8cc4d60352dda
MD5 1bcaaab054f504c517815e0b560b0e15
BLAKE2b-256 e0ae0a51bd339f6ae6e2e6d12ce1c3631f9d363d13a25449212c7347fb637202

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page