Run parameter sweeps and Monte Carlo dispersions over GMAT missions in parallel from Python.
Project description
gmat-sweep
Run parameter sweeps and Monte Carlo dispersions over GMAT missions in parallel from Python.
What this is
A parallel orchestrator on top of gmat-run's
single-run primitive. Point gmat-sweep at a working .script and either a parameter
grid, an explicit run table, or a perturbation distribution, and it fans the run set
across subprocess workers, aggregates each run's ReportFile (and any EphemerisFile
or ContactLocator outputs) into multi-indexed pandas DataFrames, and writes a JSON
Lines manifest alongside the results so any sweep is reproducible bit-for-bit. Killed
sweeps reload from the manifest and re-run only the missing or failed runs.
The four entry points cover the common shapes:
sweep(grid=...)— full-factorial grid over one or more dotted-path fields.sweep(samples=DataFrame)— explicit-row sweep where you pre-build the run set (Halton, Sobol, custom design).monte_carlo(perturb=...)— stochastic dispersion with named distributions and a deterministic seed contract.latin_hypercube(perturb=...)— stratified sampling for variance reduction at smalln.
What this is not
- Not a single-run runner — that's
gmat-run; everygmat-sweepworker calls into it. - Not a way to build GMAT missions from scratch in Python — see
gmatpyplus. - Not a
.scripttext generator — seepygmat. - Not an optimiser. Gradient-, Bayesian-, and population-based optimisation
(CasADi, pagmo2, scikit-optimize) is a different problem;
gmat-sweepmay serve as the parallel evaluator inside one, but it ships no optimiser of its own. - Not a workflow engine.
gmat-sweepruns homogeneous parametric sweeps of one mission; Snakemake / Nextflow / Hamilton manage DAGs of heterogeneous tasks. A workflow engine can schedule agmat-sweepstep; the converse is not interesting.
Requirements
- Python 3.10, 3.11, or 3.12.
gmat-run≥ 0.3 — installed as a transitive dependency from PyPI.gmat-sweepnever importsgmatpydirectly; the import happens inside each worker subprocess on first call.- A local GMAT install.
gmat-sweepdoes not ship GMAT binaries; it relies ongmat-run's install discovery, which honours$GMAT_ROOTor finds a build under a conventional path. Download GMAT from the SourceForge release page — seegmat-run's install guide for the unpack-and-discover steps.
Supported GMAT versions
| GMAT release | Status | CI |
|---|---|---|
| R2026a | Primary development target | Exercised on every PR (Ubuntu + Windows + macOS, Python 3.10/3.11/3.12) |
| R2025a | Supported | Exercised on every PR (Ubuntu + Windows + macOS, Python 3.10/3.11/3.12) |
R2023a and R2024a were never released by the upstream GMAT project; R2025a and R2026a are the only releases supported.
Installation
pip install gmat-sweep
The [examples] extra pulls in matplotlib for the example notebooks:
pip install gmat-sweep[examples]
Quick start
from gmat_sweep import LocalJoblibPool, sweep
df = sweep(
"mission.script",
grid={"Sat.SMA": [7000, 7100, 7200]},
backend=LocalJoblibPool(workers=8),
)
print(df)
That call runs mission.script three times — once per Sat.SMA value — each in a fresh
subprocess, and returns a (run_id, time)-MultiIndexed pandas.DataFrame containing
the rows from every run's ReportFile plus a __status column flagging
ok / failed / skipped. A single failed run lands as a failed row with the captured
GMAT stderr in the manifest — never as a silent zero-row DataFrame and never as an
unhandled exception that aborts the whole sweep.
For a stochastic dispersion, swap sweep
for monte_carlo and pass a
perturb mapping of named distributions:
from gmat_sweep import LocalJoblibPool, monte_carlo
df = monte_carlo(
"mission.script",
n=1000,
perturb={"Sat.SMA": ("normal", 7100.0, 50.0)},
backend=LocalJoblibPool(workers=8),
seed=42,
)
Returns the same DataFrame shape as sweep(). Per-run sub-seeds derive from seed via
numpy.random.SeedSequence.spawn, so the draw is bit-reproducible and a resumed sweep
samples the same values for any given run_id. See the
Monte Carlo guide for the full
determinism contract and latin_hypercube
for the stratified-sampling variant.
By default the per-run Parquet files and the manifest land in a temporary directory
whose lifetime is tied to the returned DataFrame. Pass out=Path(...) to keep them —
that's also what enables resuming a killed sweep
via Sweep.from_manifest(...).resume() or gmat-sweep resume <manifest>.
For multi-host sweeps, swap the local pool for DaskPool or RayPool — same
sweep() / monte_carlo() / latin_hypercube() call shape, different backend=:
from gmat_sweep import sweep
from gmat_sweep.backends import DaskPool
with DaskPool(n_workers=8) as pool:
df = sweep(
"mission.script",
grid={"Sat.SMA": [7000, 7100, 7200]},
backend=pool,
)
DaskPool and RayPool ship behind pip install gmat-sweep[dask] /
gmat-sweep[ray]. See the backends page
for the full set of pool patterns and the
cluster recipes for
Slurm / Kubernetes / Ray autoscaling wiring.
A gmat-sweep console script is also installed for shell-script and CI use:
gmat-sweep run --grid Sat.SMA=7000:7200:3 --workers 8 --out ./sweep mission.script
gmat-sweep run --grid Sat.SMA=7000:7200:3 --backend dask --workers 8 --out ./sweep mission.script
gmat-sweep monte-carlo --n 1000 --perturb 'Sat.SMA=normal:7100:50' --seed 42 --out ./mc mission.script
gmat-sweep resume --script mission.script --workers 8 ./mc/manifest.jsonl
gmat-sweep show ./sweep/manifest.jsonl
See the CLI reference in the docs for every subcommand and the full mini-grammar.
Outputs
Every sweep emits two artefacts:
- The returned DataFrame —
(run_id, time)-MultiIndexed, one column perReportFilechannel plus the__statuscolumn. Built lazily from per-run Parquet files via pyarrow's dataset API, so a 10,000-run sweep does not have to fit in memory at once. - A JSON Lines manifest (
manifest.jsonl) — append-only, fsync'd after every entry. Records the canonical script SHA-256, software-version fingerprint, full parameter spec, and per-run status, timing, output paths, and captured stderr. ACtrl-Cmid-sweep leaves the manifest in a parseable state. See the manifest schema for the full contract.
Documentation
Full docs at https://astro-tools.github.io/gmat-sweep/, including a getting-started guide, the parameter spec reference, the manifest schema, the supported-version matrix, the FAQ, and the API reference.
Runnable example notebooks:
- Single-axis SMA scan —
fifty runs across
np.linspace(7000, 8000, 50)ofSat.SMA, parallel-dispatched and overlaid on a single altitude-vs-time plot. - Two-axis epoch × time-of-flight grid —
cartesian product over
Sat.Epochand a script-levelVariable TOF, contoured by per-run miss distance. - Surviving a kill —
launch a sweep, send
SIGINTmid-run, walk through inspecting the partial manifest withgmat-sweep show, then complete the sweep withSweep.from_manifest(...).resume(). - Monte Carlo dispersion — 1000-run Monte Carlo around a nominal injection burn over a four-axis perturbation cube, with arrival-miss histogram and a 3-σ covariance ellipse.
- Latin hypercube vs Monte Carlo — 64-run Latin hypercube alongside a 64-run plain Monte Carlo on the same perturbation, pair-plotting the unit-cube samples to make the stratification visible.
- Dask cluster recipe —
100-run
Sat.SMAgrid dispatched through adistributed.LocalClusterwithDaskPool, same flow as a realdask.distributedcluster. - Ray autoscaling recipe —
100-run Monte Carlo dispatched through
RayPoolagainst a localray.init(), same task model as a real autoscaling Ray cluster.
Roadmap
| Release | Scope |
|---|---|
| v0.3 (current) | DaskPool (extra [dask]) and RayPool (extra [ray]) join LocalJoblibPool behind a single Pool ABC. CLI --backend {local,dask,ray} flag and rich gmat-sweep show --detail / --run modes. Three cluster-recipe pages (Slurm with srun, Kubernetes pod-per-worker, Ray autoscaling). Benchmark page comparing backends on a 1000-run reference sweep, with a per-backend throughput floor enforced in CI. Manifest header gains a backend field; reuse_gmat_context exposes the bootstrap-amortisation choice on every pool. |
| v0.4 (next) | Notebook-friendly __repr_html__ for Sweep / RunOutcome. Optional plotting helpers (sweep_corner, sweep_heatmap) behind a [plot] extra pulling matplotlib. A docs cookbook page on integrating sweep outputs into downstream consumers. Smoke-canary cell against the canonical ghcr.io/astro-tools/gmat image. |
Past releases live in CHANGELOG.md.
Development
To work on gmat-sweep itself:
git clone https://github.com/astro-tools/gmat-sweep.git
cd gmat-sweep
uv sync --all-groups
See CONTRIBUTING.md for the full branch / PR / test workflow.
Licence
MIT. See LICENSE.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file gmat_sweep-0.3.0.tar.gz.
File metadata
- Download URL: gmat_sweep-0.3.0.tar.gz
- Upload date:
- Size: 877.1 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
489a428504ed70df3680cc9a918fdc3cc959f4694e1c0b0487dfb20fadb207a6
|
|
| MD5 |
dde4c1c9d962f44a64d3b15d20069994
|
|
| BLAKE2b-256 |
1a2b65a217f3b3e72221249376d049d7ccd0dfc4e2ae87eaa8cbff08239899a5
|
Provenance
The following attestation bundles were made for gmat_sweep-0.3.0.tar.gz:
Publisher:
release.yml on astro-tools/gmat-sweep
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
gmat_sweep-0.3.0.tar.gz -
Subject digest:
489a428504ed70df3680cc9a918fdc3cc959f4694e1c0b0487dfb20fadb207a6 - Sigstore transparency entry: 1465769648
- Sigstore integration time:
-
Permalink:
astro-tools/gmat-sweep@781c3735c5b49756ab77444946149e13f9e3c6bf -
Branch / Tag:
refs/tags/v0.3.0 - Owner: https://github.com/astro-tools
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
release.yml@781c3735c5b49756ab77444946149e13f9e3c6bf -
Trigger Event:
push
-
Statement type:
File details
Details for the file gmat_sweep-0.3.0-py3-none-any.whl.
File metadata
- Download URL: gmat_sweep-0.3.0-py3-none-any.whl
- Upload date:
- Size: 61.9 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
d7e18df9fc1f20590ca21bcb4223d7429e2709b3df738bd654a4cdea46646162
|
|
| MD5 |
43142040f8dd0f26f380186dba46d114
|
|
| BLAKE2b-256 |
dd3bb2bb69a01902f6ad2814b0ef86dc51b8d212abf74d8a7c341f1af9a26357
|
Provenance
The following attestation bundles were made for gmat_sweep-0.3.0-py3-none-any.whl:
Publisher:
release.yml on astro-tools/gmat-sweep
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
gmat_sweep-0.3.0-py3-none-any.whl -
Subject digest:
d7e18df9fc1f20590ca21bcb4223d7429e2709b3df738bd654a4cdea46646162 - Sigstore transparency entry: 1465769748
- Sigstore integration time:
-
Permalink:
astro-tools/gmat-sweep@781c3735c5b49756ab77444946149e13f9e3c6bf -
Branch / Tag:
refs/tags/v0.3.0 - Owner: https://github.com/astro-tools
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
release.yml@781c3735c5b49756ab77444946149e13f9e3c6bf -
Trigger Event:
push
-
Statement type: