CNF/SAT-based information-theoretic online action model learning
Project description
information-gain-aml
A CNF/SAT-based information-theoretic approach for online action model learning in PDDL planning domains.
The algorithm maintains uncertainty over preconditions and effects using CNF formulas, selects actions that maximize expected information gain, and converges toward the true action model through online interaction with the environment.
Installation
pip install information-gain-aml
Quick Start
from information_gain_aml.algorithms.information_gain import InformationGainLearner
learner = InformationGainLearner(
domain_file="path/to/domain.pddl",
problem_file="path/to/problem.pddl",
)
# Select an action based on expected information gain
action_name, objects = learner.select_action(current_state)
# After observing the outcome, update the model
learner.update_model()
Key Features
- CNF-based uncertainty representation -- precondition and effect knowledge encoded as SAT formulas
- Information-theoretic action selection -- picks actions that maximize expected information gain
- Lifted learning -- learns at the operator level, generalizing across object instances
- Object subset selection -- scales to large domains by focusing on relevant object subsets
- Parallel gain computation -- optional multiprocessing for large action spaces
- MCTS-based action selection -- lookahead and full UCT strategies for deeper exploration
Action Selection Strategies
| Strategy | Description | Speed |
|---|---|---|
greedy |
Selects the action with the highest immediate information gain | Fast (default) |
lookahead |
Bounded depth-limited lookahead with discounted future gain | Moderate |
mcts |
Full UCT-based Monte Carlo Tree Search | Slow (see note below) |
Performance note: The
mctsstrategy performs SAT solving during rollouts, which makes it significantly slower than other strategies. For large domains it may be impractical. Performance improvements are planned for a future release. Uselookaheadfor a balance between exploration depth and speed.
Configuration
learner = InformationGainLearner(
domain_file="domain.pddl",
problem_file="problem.pddl",
max_iterations=1000, # max learning iterations
use_object_subset=True, # object subset selection (default: True)
spare_objects_per_type=2, # extra objects per type beyond minimum
num_workers=None, # parallel workers (None=auto, 0=sequential)
learn_negative_preconditions=True, # include negative precondition candidates
selection_strategy="greedy", # "greedy", "lookahead", or "mcts"
lookahead_depth=2, # depth for lookahead strategy
mcts_iterations=50, # iterations for mcts strategy
mcts_rollout_depth=5, # rollout depth for mcts strategy
)
Requirements
- Python >= 3.10
- python-sat -- SAT solver for CNF management
- unified-planning -- PDDL parsing and domain representation
License
MIT
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file information_gain_aml-0.3.1.tar.gz.
File metadata
- Download URL: information_gain_aml-0.3.1.tar.gz
- Upload date:
- Size: 60.6 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
698c393d0c44f267df7e011ae163781b14e3966457bdfa39a933b65a12eb6447
|
|
| MD5 |
fb5a4d2725ca3b80f32fddf32117d806
|
|
| BLAKE2b-256 |
0b7c5752f9070255421d922567b390e44f967186fad27ae9c360bcab731e1535
|
Provenance
The following attestation bundles were made for information_gain_aml-0.3.1.tar.gz:
Publisher:
publish.yml on omereliy/online_model_learning
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
information_gain_aml-0.3.1.tar.gz -
Subject digest:
698c393d0c44f267df7e011ae163781b14e3966457bdfa39a933b65a12eb6447 - Sigstore transparency entry: 1154995720
- Sigstore integration time:
-
Permalink:
omereliy/online_model_learning@6d183b4a9419b770dc81c8c9597bb28e2597232d -
Branch / Tag:
refs/tags/v0.3.1 - Owner: https://github.com/omereliy
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@6d183b4a9419b770dc81c8c9597bb28e2597232d -
Trigger Event:
push
-
Statement type:
File details
Details for the file information_gain_aml-0.3.1-py3-none-any.whl.
File metadata
- Download URL: information_gain_aml-0.3.1-py3-none-any.whl
- Upload date:
- Size: 67.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
c9cd8cf84208f5cddeed6991139dcd8b5371b4ab7cd434a9c3dc1b46fa903e50
|
|
| MD5 |
47bf9fb38eef115ba5e79c0aab8a57e7
|
|
| BLAKE2b-256 |
647decb54cd299ed0b7ff66e3f386fa1f363ac4d6002b4f99c3866c0c241d0bc
|
Provenance
The following attestation bundles were made for information_gain_aml-0.3.1-py3-none-any.whl:
Publisher:
publish.yml on omereliy/online_model_learning
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
information_gain_aml-0.3.1-py3-none-any.whl -
Subject digest:
c9cd8cf84208f5cddeed6991139dcd8b5371b4ab7cd434a9c3dc1b46fa903e50 - Sigstore transparency entry: 1154995722
- Sigstore integration time:
-
Permalink:
omereliy/online_model_learning@6d183b4a9419b770dc81c8c9597bb28e2597232d -
Branch / Tag:
refs/tags/v0.3.1 - Owner: https://github.com/omereliy
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@6d183b4a9419b770dc81c8c9597bb28e2597232d -
Trigger Event:
push
-
Statement type: