Skip to main content

Binary distribution of the CBC MILP solver (COIN-OR Branch and Cut)

Project description

cbcbox

PyPI version PyPI downloads CI Platforms License

cbcbox is a high-performance, self-contained Python distribution of the CBC MILP solver (COIN-OR Branch and Cut), built from the latest COIN-OR master branch.

On x86_64 (Linux, macOS, Windows) the wheel ships both a Haswell-optimised binary (AVX2/FMA) for maximum speed and a generic build with runtime CPU dispatch for compatibility with any x86_64 machine — selected automatically. All dynamic dependencies (OpenBLAS, libgfortran, etc.) are bundled; no system libraries or separate installation steps are needed.

Highlights

  • Haswell-optimised & generic builds — on x86_64 Linux, macOS, and Windows the wheel ships two complete solver stacks: a Haswell build (OpenBLAS AVX2/FMA kernel) for maximum throughput, and a generic build (DYNAMIC_ARCH runtime dispatch) for compatibility with any x86_64 CPU. The best available variant is selected automatically at import time (see Build variants).

  • Parallel branch-and-cut — built with --enable-cbc-parallel. Use -threads=N to distribute the search tree across N threads, giving significant speedups on multi-core machines for hard MIP instances.

  • AMD fill-reducing orderingSuiteSparse AMD is compiled in, enabling the high-quality UniversityOfFlorida Cholesky factorization for Clp's barrier (interior point) solver. AMD reordering produces much less fill-in on large sparse problems than the built-in native Cholesky, making barrier substantially faster. Activate with -cholesky UniversityOfFlorida -barrier (see barrier usage).

Performance (x86_64)

Auto-updated by CI after each successful workflow run. Single-threaded solve time — lower is better.

The AVX2/Haswell build is ~3.2× faster than the generic build on average (geometric mean across 19 instances, 2 x86_64 platforms: Darwin x86_64, Windows AMD64).

CBC solve time — generic vs AVX2/Haswell (Linux x86_64)

Single-threaded solve time across benchmark instances on Linux x86_64. Speedup factor shown above each pair. Lower is better.

See also: Windows AMD64 + macOS x86_64 summary

Build variants

On x86_64 Linux, macOS, and Windows, the wheel ships two complete sets of binaries:

Variant OpenBLAS kernel Clp SIMD Minimum CPU
generic DYNAMIC_ARCH=1 (runtime dispatch, Nehalem–Zen targets) standard any x86_64
avx2 DYNAMIC_ARCH=1 + DYNAMIC_LIST=HASWELL SKYLAKEX -march=haswell -DCOIN_AVX2=4 Haswell (2013+)

At import time cbcbox automatically selects avx2 when it is available and the running CPU supports AVX2; otherwise it falls back to generic.

You can override this selection with the CBCBOX_BUILD environment variable:

# Force generic (portable) build
CBCBOX_BUILD=generic cbc mymodel.mps -solve -quit

# Force AVX2-optimised build (raises an error if not available)
CBCBOX_BUILD=avx2 cbc mymodel.mps -solve -quit

When CBCBOX_BUILD is set, a short summary of the selected build is printed to stdout on every call — useful for tagging experiment results:

[cbcbox] CBCBOX_BUILD=avx2
[cbcbox]   binary  : .../cbcbox/cbc_dist_avx2/bin/cbc
[cbcbox]   lib dir : .../cbcbox/cbc_dist_avx2/lib
[cbcbox]   libs    : libCbc.so.3, libClp.so.3, libopenblas.so.0

Non-x86_64 platforms (Linux aarch64, macOS arm64) ship the generic build only. CBCBOX_BUILD=avx2 will raise a RuntimeError on those platforms.

Local debug builds

The released wheels are fully optimised and stripped. To debug CBC itself (e.g. with GDB or LLDB), use the scripts in scripts/ to build a local debug-enabled binary. These produce the same full feature set as the release wheels (OpenBLAS, AMD, Nauty, pthreads) but compiled with -O1 -g and, on x86_64, with -march=haswell -DCOIN_AVX2=4 so you can debug AVX2-specific code paths.

Script Platform Environment Output directory
scripts/build_debug.sh Linux, macOS native (host compiler) cbc_dist_debug_avx2/ (x86_64) or cbc_dist_debug/ (ARM64)
scripts/build_debug_manylinux.sh Linux Docker — manylinux2014 container (exact CI parity) same as above
scripts/build_debug_windows.ps1 Windows MSYS2 / MinGW64 cbc_dist_debug_avx2\

Quick start

Linux / macOS (native build):

# x86_64 → debug + AVX2 → cbc_dist_debug_avx2/bin/cbc
# ARM64  → debug only  → cbc_dist_debug/bin/cbc
./scripts/build_debug.sh

# With AddressSanitizer:
./scripts/build_debug.sh --asan

# With ThreadSanitizer:
./scripts/build_debug.sh --tsan

# Force a clean rebuild from scratch (required when switching sanitizers):
./scripts/build_debug.sh --asan --clean

Linux (manylinux2014 container — matches CI exactly):

# Requires Docker; the script prints install instructions if it is missing.
./scripts/build_debug_manylinux.sh
./scripts/build_debug_manylinux.sh --asan
./scripts/build_debug_manylinux.sh --tsan

Windows (PowerShell):

# Requires MSYS2 at C:\msys64.  Note: sanitizers are not supported on Windows/MinGW.
.\scripts\build_debug_windows.ps1
.\scripts\build_debug_windows.ps1 -Clean   # force full rebuild

Debugging

# GDB (Linux):
gdb cbc_dist_debug_avx2/bin/cbc
(gdb) run mymodel.mps -solve -quit

# LLDB (macOS):
lldb cbc_dist_debug/bin/cbc
(lldb) run mymodel.mps -solve -quit

Sanitizer tips

Sanitizer Flag What it catches Runtime env var
AddressSanitizer --asan heap/stack buffer overflows, use-after-free, memory leaks ASAN_OPTIONS=detect_leaks=0 to suppress system-lib false positives
ThreadSanitizer --tsan data races between threads TSAN_OPTIONS=halt_on_error=0 to log races without aborting

ASan and TSan are mutually exclusive. Neither is available on Windows/MinGW. Always pass --clean when switching from one sanitizer to another to avoid linking mismatched object files.

OpenBLAS is always built without sanitizer flags to avoid false positives from hand-optimised BLAS assembly; only the COIN-OR stack is instrumented.

Note: Debug binaries are not included in the published wheels because of their size. They are intended for local development only.

Supported platforms

Platform Wheel tag
Linux x86_64 manylinux2014_x86_64
Linux aarch64 manylinux2014_aarch64
macOS arm64 (Apple Silicon) macosx_11_0_arm64
macOS x86_64 macosx_10_9_x86_64
Windows AMD64 win_amd64

Installation

pip install cbcbox

Usage

Command line

After installation, CBC is available directly as the cbc command (pip installs the entry point into the environment's bin/ on Linux/macOS or Scripts/ on Windows, which is already on PATH):

cbc mymodel.lp -solve -quit
cbc mymodel.mps.gz -solve -quit
cbc mymodel.mps -seconds 60 -timem elapsed -solve -quit
cbc mymodel.mps -dualp pesteep -solve -quit

Alternatively, invoke via the Python module entry point:

python -m cbcbox mymodel.lp -solve -quit

CBC accepts LP, MPS and compressed MPS (.mps.gz) files. Pass -help for the full list of options, or -quit to exit after solving.

Parallel branch-and-cut

This build includes parallel branch-and-cut (--enable-cbc-parallel). Use -threads=N to distribute the search tree across N threads:

cbc mymodel.mps -threads=4 -solve -quit

Barrier (interior-point) solver

Clp's barrier solver can be faster than simplex for large LP relaxations. This build includes SuiteSparse AMD, which enables the high-quality UniversityOfFlorida Cholesky factorization — significantly reducing fill-in compared to the built-in native Cholesky:

# Solve LP relaxation with barrier + AMD Cholesky, then crossover to simplex basis
cbc mymodel.mps -cholesky UniversityOfFlorida -barrier -solve -quit

# Useful as a root-node strategy inside MIP (let CBC use simplex for B&B):
cbc mymodel.mps -cholesky UniversityOfFlorida -barrier -solve -quit

Without AMD, only -cholesky native (less efficient) is available.

Python API

The package exposes helpers to locate the installed files:

import cbcbox
import subprocess

# Path to the cbc binary (cbc.exe on Windows).
cbcbox.cbc_bin_path()
# e.g. '/home/user/.venv/lib/python3.13/site-packages/cbcbox/cbc_dist/bin/cbc'

# Directory containing the shared libraries.
cbcbox.cbc_lib_dir()
# e.g. '.../cbcbox/cbc_dist/lib'

# Directory containing the COIN-OR C/C++ headers.
cbcbox.cbc_include_dir()
# e.g. '.../cbcbox/cbc_dist/include/coin'

# Run CBC programmatically.
result = subprocess.run(
    [cbcbox.cbc_bin_path(), "mymodel.mps", "-solve", "-quit"],
    capture_output=True, text=True,
)
print(result.stdout)

What is built

The build pipeline compiles all components from source inside the CI runner, in the following order:

Component Version / branch Purpose
Cbc master Branch-and-cut MIP solver
Cgl master Cut generation library
Clp master Simplex LP solver (used as the MIP node relaxation)
Osi master Open Solver Interface
CoinUtils master Utility library (shared by all COIN-OR packages)
Nauty 2.8.9 Symmetry detection for MIP presolve
AMD (SuiteSparse v7.12.2) v7.12.2 Sparse matrix fill-reducing ordering
OpenBLAS v0.3.31 Optimised BLAS/LAPACK for LP basis factorisation

On x86_64 Linux, macOS, and Windows the entire stack is compiled twice: once for the generic variant (OpenBLAS DYNAMIC_ARCH=1 with a broad set of x86_64 targets for runtime dispatch) and once for the avx2 variant (OpenBLAS DYNAMIC_ARCH=1 restricted to Haswell/Skylake targets via DYNAMIC_LIST, COIN-OR compiled with -march=haswell -DCOIN_AVX2=4). Both variants use NO_CBLAS=1 (COIN-OR only calls the Fortran BLAS interface). AMD and Nauty are built only once (they are pure combinatorial code with no BLAS dependency) and reused by both COIN-OR variants.

All COIN-OR components are built as shared (.so / .dylib / .dll) libraries. The shared libraries are patched with self-relative RPATHs and bundled inside the wheel, making them directly usable via cffi or ctypes without any system installation.

Wheel contents

The wheel installs under cbcbox/ inside the site-packages directory. On x86_64 Linux, macOS, and Windows it contains two dist trees; other platforms contain only cbc_dist/:

cbc_dist/           ← generic build (all platforms)
cbc_dist_avx2/      ← AVX2-optimised build (x86_64 Linux/macOS/Windows)
├── bin/
│   ├── cbc           # CBC MIP solver binary  (cbc.exe on Windows)
│   └── clp           # Clp LP solver binary   (clp.exe on Windows)
├── lib/
│   ├── libCbc.so / libCbc.dylib / libCbc.dll  # CBC solver
│   ├── libCbcSolver.so ...
│   ├── libClp.so ...                          # Clp LP solver
│   ├── libCgl.so ...                          # Cut generation
│   ├── libOsi.so ...                          # Solver interface
│   ├── libOsiClp.so ...                       # Clp OSI binding
│   ├── libOsiCbc.so ...                       # CBC OSI binding (where available)
│   ├── libCoinUtils.so ...
│   ├── libopenblas.so / .dylib / .dll         # OpenBLAS BLAS/LAPACK
│   ├── pkgconfig/                             # .pc files for all libraries
│   └── <bundled runtime shared libs>          # Platform-specific — see below
└── include/
    ├── coin/      # COIN-OR headers (CoinUtils, Osi, Clp, Cgl, Cbc)
    ├── nauty/     # Nauty headers
    └── *.h        # SuiteSparse / AMD headers

Bundled dynamic libraries

Because OpenBLAS links to the Fortran runtime, the following shared libraries are bundled inside the wheel and their paths are rewritten so no system installation is required.

Linux (lib/ directory, RPATH set to $ORIGIN)

Library Description
libopenblas.so.0 OpenBLAS BLAS/LAPACK
libgfortran.so.5 GNU Fortran runtime
libquadmath.so.0 Quad-precision math (dependency of libgfortran)

macOS (lib/ directory, install names rewritten to @rpath/)

Library Description
libopenblas.dylib OpenBLAS BLAS/LAPACK
libgfortran.5.dylib GNU Fortran runtime
libgcc_s.1.1.dylib GCC runtime
libquadmath.0.dylib Quad-precision math

Windows (bin/ directory, DLLs placed next to the executable)

Library Description
libopenblas.dll OpenBLAS BLAS/LAPACK
libgfortran-5.dll GNU Fortran runtime
libgcc_s_seh-1.dll GCC SEH runtime
libquadmath-0.dll Quad-precision math
libstdc++-6.dll C++ standard library (MinGW64)
libwinpthread-1.dll POSIX thread emulation

CI / build pipeline

Wheels are built and tested automatically via GitHub Actions using cibuildwheel. The workflow (.github/workflows/wheel.yml) runs independent compile jobs in parallel, then packages each platform:

Compile jobs Runner Produces
compile-linux-x64-generic + compile-linux-x64-avx2 ubuntu-latest manylinux2014_x86_64 wheel
compile-linux-arm64-generic ubuntu-24.04-arm manylinux2014_aarch64 wheel
compile-macos-arm64-generic macos-15 macosx_11_0_arm64 wheel
compile-macos-intel-generic + compile-macos-intel-avx2 macos-15-intel macosx_10_9_x86_64 wheel
compile-windows-generic + compile-windows-avx2 windows-latest win_amd64 wheel

Each platform's compile jobs run in parallel. Once all compile jobs for a platform finish, the corresponding package-* job assembles the wheel via cibuildwheel and runs the test suite against the installed wheel.

A final combine_reports job collects per-platform performance results and commits the updated README.md to the repository.

Integration tests

The test suite (pytest) solves 20 MIP instances and checks the optimal objective values, in both single-threaded and parallel (3-thread) modes. On x86_64 Linux, macOS, and Windows each test is run twice — once against the generic binary and once against the avx2 binary — and a side-by-side performance comparison is recorded:

Instance Expected optimal Time limit
pp08a 7 350 2000 s
sprint_hidden06_j 130 2000 s
air03 340 160 2000 s
air04 56 137 2000 s
air05 26 374 2000 s
nw04 16 862 2000 s
mzzv11 −21 718 2000 s
trd445c −153 419.078836 2000 s
nursesched-sprint02 58 2000 s
stein45 30 2000 s
neos-810286 2 877 2000 s
neos-1281048 601 2000 s
j3050_8 1 2000 s
qiu −132.873136947 2000 s
gesa2-o 25 779 856.3717 2000 s
pk1 11 2000 s
mas76 40 005.054142 2000 s
app1-1 −3 2000 s
eil33-2 934.007916 2000 s
fiber 405 935.18 2000 s

Time limits are generous to avoid false failures on slow CI runners.

Performance results

Auto-updated by CI after each successful workflow run.

Summary

Geometric mean solve time (seconds) across all test instances.

1 thread

Platform generic (s) avx2 (s) avx2 speedup
Darwin x86_64 66.36 17.18 3.86×
Darwin arm64 40.17
Windows AMD64 54.47 20.16 2.70×

3 threads

Platform generic (s) avx2 (s) avx2 speedup
Darwin x86_64 38.08 16.29 2.34×
Darwin arm64 38.32
Windows AMD64 42.17 18.41 2.29×

Per-instance results

pp08a.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 4.49 10.52 0.43×
Darwin x86_64 generic 12.07 6.21 1.94×
Darwin arm64 generic 8.80 19.54 0.45×
Windows AMD64 avx2 5.25 8.43 0.62×
Windows AMD64 generic 12.41 17.70 0.70×

sprint_hidden06_j.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 45.73 45.99 0.99×
Darwin x86_64 generic 208.78 161.11 1.30×
Darwin arm64 generic 116.66 119.24 0.98×
Windows AMD64 avx2 94.33 91.44 1.03×
Windows AMD64 generic 237.12 224.04 1.06×

air03.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 1.56 1.71 0.91×
Darwin x86_64 generic 6.04 6.22 0.97×
Darwin arm64 generic 3.77 3.91 0.96×
Windows AMD64 avx2 3.03 3.16 0.96×
Windows AMD64 generic 6.03 6.26 0.96×

air04.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 49.78 39.04 1.28×
Darwin x86_64 generic 125.52 71.48 1.76×
Darwin arm64 generic 101.43 77.99 1.30×
Windows AMD64 avx2 43.46 36.37 1.19×
Windows AMD64 generic 153.25 87.31 1.76×

air05.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 23.03 19.70 1.17×
Darwin x86_64 generic 58.12 42.19 1.38×
Darwin arm64 generic 46.13 37.99 1.21×
Windows AMD64 avx2 21.68 17.39 1.25×
Windows AMD64 generic 56.63 44.08 1.28×

nw04.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 11.95 12.72 0.94×
Darwin x86_64 generic 36.83 35.88 1.03×
Darwin arm64 generic 32.69 33.70 0.97×
Windows AMD64 avx2 18.08 18.71 0.97×
Windows AMD64 generic 63.84 61.80 1.03×

mzzv11.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 115.86 78.14 1.48×
Darwin x86_64 generic 532.25 372.20 1.43×
Darwin arm64 generic 258.36 170.75 1.51×
Windows AMD64 avx2 135.28 206.71 0.65×
Windows AMD64 generic 255.32 319.91 0.80×

trd445c.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 108.17 96.74 1.12×
Darwin x86_64 generic 278.08 197.85 1.41×
Darwin arm64 generic 188.64 166.87 1.13×
Windows AMD64 avx2 122.33 142.31 0.86×
Windows AMD64 generic 231.04 228.85 1.01×

nursesched-sprint02.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 28.95 30.09 0.96×
Darwin x86_64 generic 105.55 88.64 1.19×
Darwin arm64 generic 97.27 87.85 1.11×
Windows AMD64 avx2 38.79 38.04 1.02×
Windows AMD64 generic 107.88 79.44 1.36×

stein45.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 8.42 8.40 1.00×
Darwin x86_64 generic 25.37 14.09 1.80×
Darwin arm64 generic 21.31 13.24 1.61×
Windows AMD64 avx2 8.72 6.50 1.34×
Windows AMD64 generic 26.21 15.65 1.67×

neos-810286.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 12.42 16.55 0.75×
Darwin x86_64 generic 45.56 38.75 1.18×
Darwin arm64 generic 28.88 79.34 0.36×
Windows AMD64 avx2 15.61 15.39 1.01×
Windows AMD64 generic 35.03 37.94 0.92×

neos-1281048.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 17.69 6.88 2.57×
Darwin x86_64 generic 111.58 16.17 6.90×
Darwin arm64 generic 43.51 17.92 2.43×
Windows AMD64 avx2 14.58 11.20 1.30×
Windows AMD64 generic 36.40 23.30 1.56×

j3050_8.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 3.90 3.70 1.05×
Darwin x86_64 generic 6.99 7.68 0.91×
Darwin arm64 generic 8.05 6.61 1.22×
Windows AMD64 avx2 2.37 2.62 0.90×
Windows AMD64 generic 8.23 6.92 1.19×

qiu.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 42.15 12.96 3.25×
Darwin x86_64 generic 67.23 28.16 2.39×
Darwin arm64 generic 97.17 26.15 3.72×
Windows AMD64 avx2 25.60 12.96 1.98×
Windows AMD64 generic 79.49 37.35 2.13×

gesa2-o.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 4.73 5.00 0.95×
Darwin x86_64 generic 15.48 9.50 1.63×
Darwin arm64 generic 9.95 10.21 0.97×
Windows AMD64 avx2 3.75 3.69 1.02×
Windows AMD64 generic 11.21 11.66 0.96×

pk1.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 33.89 37.78 0.90×
Darwin x86_64 generic 107.08 56.64 1.89×
Darwin arm64 generic 72.51 50.87 1.43×
Windows AMD64 avx2 36.98 33.96 1.09×
Windows AMD64 generic 97.39 56.13 1.74×

mas76.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 18.16 40.53 0.45×
Darwin x86_64 generic 62.26 60.85 1.02×
Darwin arm64 generic 46.06 44.65 1.03×
Windows AMD64 avx2 20.67 32.65 0.63×
Windows AMD64 generic 47.28 57.56 0.82×

app1-1.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 5.62 7.90 0.71×
Darwin x86_64 generic 683.38 52.48 13.02×
Darwin arm64 generic 13.34 122.01 0.11×
Windows AMD64 avx2 23.44 8.81 2.66×
Windows AMD64 generic 89.27 24.91 3.58×

eil33-2.mps.gz

Platform Build 1 thread (s) 3 threads (s) parallel speedup
Darwin x86_64 avx2 36.85 20.69 1.78×
Darwin x86_64 generic 151.21 74.97 2.02×
Darwin arm64 generic 114.47 58.20 1.97×
Windows AMD64 avx2 46.03 24.50 1.88×
Windows AMD64 generic 154.12 65.04 2.37×

NAQ — Never Asked Questions

Why not benchmark on the full MIPLIB 2017 library?

Several practical constraints shape the benchmark set:

  1. CI time limits. GitHub Actions enforces a 6-hour wall-clock limit per job. The full MIPLIB 2017 collection contains ~240 instances, many of which take hours even on fast hardware. Including all of them would make every CI run time out before producing any useful measurements.

  2. Comparing apples to apples requires instances solved to optimality. If some instances are only solved within a time limit (i.e., a gap > 0 %), a meaningful performance comparison must account for both solve time and solution quality simultaneously. This greatly complicates analysis and makes plots harder to interpret. Restricting to instances that CBC reliably solves to proven optimality keeps the comparison clean: a single elapsed-time number per instance is all that is needed.

  3. The instance set is intentionally biased toward set packing / covering / partitioning structure. Most instances in the benchmark (pp08a, sprint_hidden06_j, nw04, mzzv11, nursesched-sprint02, air0x, trd445c) contain large blocks of set packing, covering, or partitioning constraints. This structure arises naturally in applications such as crew scheduling, nurse scheduling, vehicle routing, and cutting stock — exactly the domain where column generation is most valuable. Since the benchmark focuses on this problem class rather than providing a general-purpose solver survey, it is a specially interesting use case.

License

CBC and all COIN-OR components are distributed under the Eclipse Public License 2.0. OpenBLAS is distributed under the BSD 3-Clause licence. SuiteSparse AMD is distributed under the BSD 3-Clause licence. Nauty is distributed under the Apache 2.0 licence.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distributions

If you're not sure about the file name format, learn more about wheel file names.

cbcbox-2.921-py3-none-win_amd64.whl (57.6 MB view details)

Uploaded Python 3Windows x86-64

cbcbox-2.921-py3-none-manylinux2014_x86_64.whl (72.7 MB view details)

Uploaded Python 3

cbcbox-2.921-py3-none-manylinux2014_aarch64.whl (35.9 MB view details)

Uploaded Python 3

cbcbox-2.921-py3-none-macosx_15_0_x86_64.whl (59.9 MB view details)

Uploaded Python 3macOS 15.0+ x86-64

cbcbox-2.921-py3-none-macosx_15_0_arm64.whl (30.1 MB view details)

Uploaded Python 3macOS 15.0+ ARM64

File details

Details for the file cbcbox-2.921-py3-none-win_amd64.whl.

File metadata

  • Download URL: cbcbox-2.921-py3-none-win_amd64.whl
  • Upload date:
  • Size: 57.6 MB
  • Tags: Python 3, Windows x86-64
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for cbcbox-2.921-py3-none-win_amd64.whl
Algorithm Hash digest
SHA256 397886284bd5c79fcf6f00fb25e6d01495d1ddc652ef8b901c2125145cc94777
MD5 258687b31386a6e75d9b2b60a83dbdc8
BLAKE2b-256 0c6f348fdb713bb35140a38c7f0995fdfc422bf81639d0ca7a999397f7940a18

See more details on using hashes here.

Provenance

The following attestation bundles were made for cbcbox-2.921-py3-none-win_amd64.whl:

Publisher: wheel.yml on h-g-s/cbcbox

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file cbcbox-2.921-py3-none-manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for cbcbox-2.921-py3-none-manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 781dbd612833daab29593a4b10f3f3794e3635f39ac768a925d4b593df4b492d
MD5 b54670972c7afa0a4efc3aab423b58b6
BLAKE2b-256 dac0e05273eccf170beeffc2c43559dd1ad637a51236077d6423a98f0d548477

See more details on using hashes here.

Provenance

The following attestation bundles were made for cbcbox-2.921-py3-none-manylinux2014_x86_64.whl:

Publisher: wheel.yml on h-g-s/cbcbox

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file cbcbox-2.921-py3-none-manylinux2014_aarch64.whl.

File metadata

File hashes

Hashes for cbcbox-2.921-py3-none-manylinux2014_aarch64.whl
Algorithm Hash digest
SHA256 42c14f8135bc6bb5f678824676b3afc8bfce49b7a09ffeb757015af3a85cc227
MD5 4b6da95d54d16915cdee55f0a4313700
BLAKE2b-256 211ec6a69a365b42e757d4b96e0a107517c6ebbcd7ac044074cc282ea55ff7b3

See more details on using hashes here.

Provenance

The following attestation bundles were made for cbcbox-2.921-py3-none-manylinux2014_aarch64.whl:

Publisher: wheel.yml on h-g-s/cbcbox

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file cbcbox-2.921-py3-none-macosx_15_0_x86_64.whl.

File metadata

File hashes

Hashes for cbcbox-2.921-py3-none-macosx_15_0_x86_64.whl
Algorithm Hash digest
SHA256 02cc89e0560f135a355cec8736f693feeee8987c69c641fbb7b7339a2f4a60fd
MD5 02a0322d86c7136e16b3407bb126745f
BLAKE2b-256 57098a3025946766abd9b95515af396cbe8d550f3a437db704626fa139965fec

See more details on using hashes here.

Provenance

The following attestation bundles were made for cbcbox-2.921-py3-none-macosx_15_0_x86_64.whl:

Publisher: wheel.yml on h-g-s/cbcbox

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file cbcbox-2.921-py3-none-macosx_15_0_arm64.whl.

File metadata

File hashes

Hashes for cbcbox-2.921-py3-none-macosx_15_0_arm64.whl
Algorithm Hash digest
SHA256 775f37b0f6eea2205193ac0bcd92d863aebd4655d0acc82b78ed4e4c7ed124d1
MD5 a1e3da06d3b208d1d66830334c3a8d86
BLAKE2b-256 a07d85bff1039533f124881d2c945bfb5d7c8a28d2e4747d0d5e3ac6b536cc76

See more details on using hashes here.

Provenance

The following attestation bundles were made for cbcbox-2.921-py3-none-macosx_15_0_arm64.whl:

Publisher: wheel.yml on h-g-s/cbcbox

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page