Evidence Theory Tools — belief functions, combination rules and contextual correction mechanisms for the Dempster-Shafer / Transferable Belief Model
Project description
evtools
Evidence Theory Tools — a Python library for working with belief functions in the Dempster-Shafer theory / Transferable Belief Model.
Modules
| Module | Description |
|---|---|
evtools.dsvector |
DSVector — unified container for any belief function representation |
evtools.conversions |
Low-level conversions via the Fast Möbius Transform |
evtools.combinations |
Combination rules: CRC, Dempster, DRC, Cautious, Bold |
evtools.corrections |
Correction mechanisms: discounting, reinforcement, negating |
evtools.dsvector
DSVector is the central object of evtools. It represents any belief function
as a vector on 2^Ω, in both sparse (dict) and dense (numpy array)
forms. The sparse representation is the master; the dense array is computed
on demand and cached.
Kind enum
Kind |
Symbol | Name |
|---|---|---|
Kind.M |
m |
Basic Belief Assignment (mass function) |
Kind.BEL |
bel |
Belief function |
Kind.PL |
pl |
Plausibility function |
Kind.B |
b |
Commonality function |
Kind.Q |
q |
Implicability function |
Kind.V |
v |
Disjunctive weight function |
Kind.W |
w |
Conjunctive weight function |
Constructors
from evtools.dsvector import DSVector, Kind
# Human-friendly: name focal elements as strings
m = DSVector.from_focal(["a", "b", "c"], {"a": 0.3, "b,c": 0.5})
# From a dense numpy array (binary index ordering, Smets 2002)
m = DSVector.from_dense(["a", "b", "c"], np.array([0, 0.3, 0, 0, 0.5, 0, 0, 0.2]))
# From a sparse dict of frozensets
m = DSVector.from_sparse(["a", "b", "c"], {
frozenset({"a"}): 0.3,
frozenset({"b", "c"}): 0.5,
frozenset({"a","b","c"}): 0.2,
})
Special constructors
# Simple MF A^β — focal sets Ω (mass β) and A (mass 1−β)
s = DSVector.simple(["a", "b", "c"], frozenset({"a"}), beta=0.6)
# Negative simple MF θ^β — focal sets ∅ (mass β) and θ (mass 1−β)
ns = DSVector.negative_simple(["a", "b", "c"], frozenset({"a"}), beta=0.4)
Conversions
pl = m.to(Kind.PL) # returns a new DSVector with kind=Kind.PL
bel = m.to_bel() # shortcut
b = m.to_b() # commonality
q = m.to_q() # implicability
v = m.to_v() # disjunctive weights (requires subnormal BBA)
w = m.to_w() # conjunctive weights (requires subnormal BBA)
Accessing values
m.sparse # dict[frozenset, float]
m.dense # np.ndarray of length 2^n
m.is_valid # True if all masses ≥ 0 and sum = 1 (Kind.M only)
m[frozenset({"a"})] # value for a given subset (0.0 if absent)
for subset, value in m: ... # iterate over non-zero focal elements
evtools.combinations
Combination rules for aggregating beliefs from multiple sources.
from evtools.combinations import crc, dempster, drc, cautious, bold
m12 = crc(m1, m2) # m1 & m2 — Conjunctive Rule (TBM), distinct reliable sources
m12 = dempster(m1, m2) # m1 @ m2 — Dempster's normalized rule
m12 = drc(m1, m2) # m1 | m2 — Disjunctive Rule, at least one reliable
m12 = cautious(m1, m2) # Cautious rule, nondistinct reliable sources
m12 = bold(m1, m2) # Bold disjunctive rule, nondistinct possibly unreliable
# Decombination (inverse operations — result may not be a valid BBA, check .is_valid)
m1 = decombine_crc(m12, m2) # m12 6∩ m2 — removes m2 from a conjunctive combination
m1 = decombine_drc(m12, m2) # m12 6∪ m2 — removes m2 from a disjunctive combination
Choice of rule:
| All sources reliable | At least one reliable | |
|---|---|---|
| Distinct sources | crc / dempster |
drc |
| Nondistinct sources | cautious |
bold |
Both crc and drc support method="sparse" (default) or method="dense".
evtools.corrections
Correction mechanisms for adjusting a BBA based on knowledge about the quality of a source (reliability, truthfulness).
from evtools.corrections import (
discount,
contextual_discount,
theta_contextual_discount,
contextual_reinforce,
contextual_dediscount,
contextual_dereinforce,
contextual_negate,
)
# Classical discounting — source reliable with degree 1-α
m_disc = discount(m, alpha=0.4)
# Contextual discounting — reliability depends on each singleton context
betas = {frozenset({"a"}): 0.6, frozenset({"h"}): 1.0, frozenset({"r"}): 1.0}
m_cd = contextual_discount(m, betas)
# Θ-contextual discounting — reliability per coarsening partition
betas_theta = {frozenset({"a"}): 0.4, frozenset({"h","r"}): 0.9}
m_theta = theta_contextual_discount(m, betas_theta)
# Contextual reinforcement — dual of discounting (uses CRC instead of DRC)
m_cr = contextual_reinforce(m, betas)
# Inverse operations (result may not be a valid BBA — check .is_valid)
m_cdd = contextual_dediscount(m_cd, betas) # reverses contextual_discount
m_cdr = contextual_dereinforce(m_cr, betas) # reverses contextual_reinforce
# Contextual negating — source lies for some contexts
m_cn = contextual_negate(m, {frozenset({"a"}): 0.7})
Hierarchy of discounting:
discount(m, α)
└── theta_contextual_discount(m, {Ω: 1-α})
contextual_discount(m, β)
└── theta_contextual_discount(m, β) [Θ = singletons]
theta_contextual_discount(m, β) [general Θ partition]
evtools.conversions
Low-level conversion functions operating on plain numpy arrays (length 2^n),
using the Fast Möbius Transform. Every conversion is available as
<source>to<target>, e.g. mtob, pltom, qtow, beltov, etc.
from evtools.conversions import mtob, mtopl, mtobel, mtoq
m = np.array([0.0, 0.5, 0.0, 0.0, 0.5, 0.0, 0.0, 0.0])
print(mtoq(m)) # commonality function
print(mtopl(m)) # plausibility function
Array indices follow the binary ordering of Smets (2002): index i corresponds
to the subset whose members are the frame atoms at the bit positions set in i.
Installation
pip install evtools-dst
Or from source:
git clone https://github.com/daviddavkanmercier/evtools.git
cd evtools
pip install -e .
Running tests
pip install -e ".[dev]"
pytest tests/
References
- P. Smets. The application of the matrix calculus to belief functions, International Journal of Approximate Reasoning, 31(1–2):1–30, 2002.
- T. Denœux. Conjunctive and disjunctive combination of belief functions induced by non-distinct bodies of evidence, Artificial Intelligence, 172:234–264, 2008.
- D. Mercier, B. Quost, T. Denœux, Refined modeling of sensor reliability in the belief function framework using contextual discounting, Information Fusion, Vol. 9, Issue 2, pp 246-258, April 2008.
- F. Pichon, D. Mercier, É. Lefèvre, F. Delmotte, Proposition and learning of some belief function contextual correction mechanisms, International Journal of Approximate Reasoning, Vol. 72, pp 4-42, May 2016.
License
MIT
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file evtools_dst-0.4.0.tar.gz.
File metadata
- Download URL: evtools_dst-0.4.0.tar.gz
- Upload date:
- Size: 27.3 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.10.11
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
d2ae3866fe1a4b4f54c5ce265ddfc2b2ae1e34251264403cbaf4dac17c03b890
|
|
| MD5 |
b0e55c9d8bfdb937ba20b41e313dce1d
|
|
| BLAKE2b-256 |
4ab4b1934cc896c9a136d6870dd401f81b44b7227905b732224b55fa76a21ee6
|
File details
Details for the file evtools_dst-0.4.0-py3-none-any.whl.
File metadata
- Download URL: evtools_dst-0.4.0-py3-none-any.whl
- Upload date:
- Size: 21.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.10.11
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
3361bc3d227919f7ba1d067a13e39324570294e2fb1719a7209a0ea080ecaf3f
|
|
| MD5 |
f10b27a70631e1c82cb00aaa5b2dba84
|
|
| BLAKE2b-256 |
67eb06de7a948176c1b8a9a217db5c48f415f91535bc8b8ea465b86060cd9709
|