Skip to main content

A Bayesian latent tree model of multivariate multinomial data

Project description

A Bayesian latent tree model

A Bayesian latent tree model

TreeCat

Build Status Latest Version DOI

Intended Use

TreeCat is an inference engine intended to power higher-level machine learning tools. TreeCat is appropriate for analyzing medium-sized tabular data with categorical and ordinal values, possibly with missing observations.

TreeCat supports

Feature Types

categorical, ordinal

# Rows (n)

1000-100K

# Features (p)

10-1000

# Cells (n × p)

10K-10M

# Categories

2-10ish

Max Ordinal

10ish

Missing obervations?

yes

Repeated observations?

yes

Sparse data?

no, use something else

Unsupervised

yes

Semisupervised

yes

Supervised

no, use something else

Installing

First install numba (conda makes this easy). Then

$ pip install pytreecat

Quick Start

  1. Format your data as a `data.csv <treecat/testdata/tiny_data.csv>`__ file with a header row. It’s fine to include extra columns that won’t be used.

    Contents of `data.csv <treecat/testdata/tiny_data.csv>`__:

    title

    genre

    decade

    rating

    vertigo

    thriller

    1950s

    5

    up

    family

    2000s

    3

    desk set

    comedy

    1950s

    4

    santapaws

    family

    2010s

  2. Generate two schema files `types.csv <treecat/testdata/tiny_types.csv>`__ and `values.csv <treecat/testdata/tiny_values.csv>`__ using TreeCat’s guess-schema command:

    $ treecat guess-schema data.csv types.csv values.csv

    You can manually fix any incorrectly guessed feature types, or add/remove feature values. TreeCat ignore features with an empty type field.

    Contents of `types.csv <treecat/testdata/tiny_types.csv>`__:

    name

    type

    total

    unique

    singletons

    title

    11

    11

    11

    genre

    categorical

    11

    7

    4

    decade

    categorical

    11

    6

    3

    rating

    ordinal

    10

    5

    2

    Contents of `values.csv <treecat/testdata/tiny_values.csv>`__:

    name

    value

    count

    title

    _OTHER

    11

    genre

    _OTHER

    11

    genre

    drama

    3

    genre

    family

    2

  3. Import your csv files into treecat’s internal format. We’ll call our dataset dataset.pkz (a gzipped pickle file).

    $ treecat import-data data.csv types.csv values dataset.pkz
  4. Train an ensemble model on your dataset. This typically takes ~15minutes for a 1M cell dataset.

    $ treecat train dataset.pkz ensemble.pkz
  5. Load your trained model into a server

    from treecat.serving import EnsembleServer
    
    server = EnsembleServer('ensemble.pkz')
  6. Run queries against the server. For example we can compute marginals

    server.sample(100, np.ones(V)).mean(axis=0)

    or compute a latent correlation matrix

    print(server.latent_correlation())

The Server Interface

TreeCat’s server interface currently supports the two basic Bayesian operations:

  • server.sample(N, counts, data=None) draws N samples from the joint posterior distribution, optionally conditioned on data.

  • server.logprob(data) computes posterior log probability of data.

TreeCat’s internal data representation is multinomial, and thus supports missing and repeated measurements, and even data adding. For example to compute conditional probability of data A given data B, we can simply compute

cond = server.logprob(A + B) - server.logprob(B)

The Model

Let V be a set of vertices (one vertex per feature). Let C[v] be the dimension of the vth feature. Let N be the number of datapoints. Let K[n,v] be the number of observations of feature v in row n (e.g. 1 for a categorical variable, 0 for missing data, or k for an ordinal value with minimum 0 and maximum k).

TreeCat is the following generative model:

E ~ UniformSpanningTree(V)    # An undirected tree.
for v in V:
    Pv[v] ~ Dirichlet(size = [M], alpha = 1/2)
for (u,v) in E:
    Pe[u,v] ~ Dirichlet(size = [M,M], alpha = 1/(2*M))
    assume(Pv[u] == sum(Pe[u,v], axis = 1))
    assume(Pv[v] == sum(Pe[u,v], axis = 0))
for v in V:
    for i in 1:M:
        Q[v,i] ~ Dirichlet(size = [C[v]])
for n in 1:N:
    for v in V:
        X[n,v] ~ Categorical(Pv[v])
    for (u,v) in E:
        (X[n,u],X[n,v]) ~ Categorical(Pe[u,v])
    for v in V:
        Z[n,v] ~ Multinomial(Q[v,X[n,v]], count = K[n,v])

where we’ve avoided adding an arbitrary root to the tree, and instead presented the model as a manifold with overlapping variables and constraints.

The Inference Algorithm

This package implements fully Bayesian MCMC inference using subsample-annealed Gibbs sampling. There are two pieces of latent state that are sampled:

  • Latent classes for each row for each vertex. These are sampled by single-site Gibbs sampling with a linear subsample-annealing schedule.

  • The latent tree structure is sampled by randomly removing an edge and replacing it. Since removing an edge splits the graph into two connected components, the only replacement locations that are feasible are those that re-connect the graph.

The single-site Gibbs sampler uses dynamic programming to simultaneously sample the complete latent assignment vector for each row. A dynamic programming program is created each time the tree structure changes. This program is interpreted by various virtual machines for different purposes (training the model, sampling from the posterior, computing log probability of the posterior). The virtual machine for training is jit-compiled using numba.

License

Copyright (c) 2017 Fritz Obermeyer. TreeCat is licensed under the Apache 2.0 License.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pytreecat-0.1.5.tar.gz (60.7 kB view details)

Uploaded Source

File details

Details for the file pytreecat-0.1.5.tar.gz.

File metadata

  • Download URL: pytreecat-0.1.5.tar.gz
  • Upload date:
  • Size: 60.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No

File hashes

Hashes for pytreecat-0.1.5.tar.gz
Algorithm Hash digest
SHA256 3e9a1a0e6eb8dcea8a45f20c1b11d222782923e0dc6703d34f11ff24407c1db9
MD5 4e2e3cd421974562d04d3b75bdee5112
BLAKE2b-256 0178e0267f32ccfbc35b581e94f539b8d54459f53bbfe17ee8f3ecae9f27528d

See more details on using hashes here.

Provenance

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page