Skip to main content

A python package for expanded topic modeling and metrics

Project description

STREAM

We present STREAM, a Simplified Topic Retrieval, Exploration, and Analysis Module for user-friendly topic modelling and especially subsequent interactive topic visualization and analysis.

Table of Contents

For better topic analysis, we implement multiple intruder-word based topic evaluation metrics. Additionally, we publicize multiple new datasets that can extend the so far very limited number of publicly available benchmark datasets in topic modeling. We integrate downstream interpretable analysis modules to enable users to easily analyse the created topics in downstream tasks together with additional tabular information.

Speed

Since most of STREAMs models are centered around Document embeddings, STREAM comes along with a set of pre-embedded datasets. Additionally, once a user fits a model that leverages document embeddings, the embeddings are saved and automatically loaded the next time the user wants to fit any model with the same set of embeddings.

Figure Description

Installation

stream_topic is available on PyPI. To install STREAM, you can either install it directly from the GitHub repository using the following command:

pip install git+https://github.com/AnFreTh/STREAM.git

or simply install via:

pip install stream_topic

Make additionally sure to download the necessary nltk ressources, e.g. via:

import nltk
nltk.download('averaged_perceptron_tagger')

Available Models

STREAM offers a variety of neural as well as non-neural topic models and we are always trying to incorporate more and new models. If you wish to incorporate your own model, or want another model incorporated please raise an issue with the required information. Currently, the following models are implemented:

Name Implementation
LDA Latent Dirichlet Allocation
WordCluTM Tired of topic models?
CEDC Topics in the Haystack
DCTE Human in the Loop
KMeansTM Simple Kmeans followed by c-tfidf
SomTM Self organizing map followed by c-tfidf
CBC Coherence based document clustering
TNTM Transformer-Representation Neural Topic Model
ETM Topic modeling in embedding spaces
CTM Combined Topic Model
CTMNeg Contextualized Topic Models with Negative Sampling
ProdLDA Autoencoding Variational Inference For Topic Models
NeuralLDA Autoencoding Variational Inference For Topic Models

Available Metrics

Since evaluating topic models, especially automatically, STREAM implements numerous evaluation metrics. Especially, the intruder based metrics, while they might take some time to compute, have shown great correlation with human evaluation.

Name Description
ISIM Average cosine similarity of top words of a topic to an intruder word.
INT For a given topic and a given intruder word, Intruder Accuracy is the fraction of top words to which the intruder has the least similar embedding among all top words.
ISH Calculates the shift in the centroid of a topic when an intruder word is replaced.
Expressivity Cosine Distance of topics to meaningless (stopword) embedding centroid
Embedding Topic Diversity Topic diversity in the embedding space
Embedding Coherence Cosine similarity between the centroid of the embeddings of the stopwords and the centroid of the topic.
NPMI Classical NPMi coherence computed on the source corpus.

Available Datasets

To integrate custom datasets for modeling with STREAM, please follow the example notebook in the examples folder. For benchmarking new models, STREAM already includes the following datasets:

Name # Docs # Words # Features Description
Spotify_most_popular 5,860 18,193 17 Spotify dataset comprised of popular song lyrics and various tabular features.
Spotify_least_popular 5,124 20,168 14 Spotify dataset comprised of less popular song lyrics and various tabular features.
Spotify 11,012 25,835 14 General Spotify dataset with song lyrics and various tabular features.
Reddit_GME 21,559 11,724 6 Reddit dataset filtered for "Gamestop" (GME) from the Subreddit "r/wallstreetbets".
Stocktwits_GME 300,000 14,707 3 Stocktwits dataset filtered for "Gamestop" (GME), covering the GME short squeeze of 2021.
Stocktwits_GME_large 600,000 94,925 0 Larger Stocktwits dataset filtered for "Gamestop" (GME), covering the GME short squeeze of 2021.
Reuters 10,788 19,696 - Preprocessed Reuters dataset.
Poliblogs 13,246 47,106 2 Preprocessed Poliblogs dataset suitable for STMs.
20NewsGroups 18,846 70,461 - preprocessed 20NewsGroups dataset
BBC_News 2,225 19,116 - preprocessed BBC News dataset
If you wish yo include and publish one of your datasets directly into the package, feel free to contact us.

Usage

To use one of the available models, follow the simple steps below:

  1. Import the necessary modules:

    from stream_topic.models import KmeansTM
    from stream_topic.utils import TMDataset
    

Preprocessing

  1. Get your dataset and preprocess for your model:
    dataset = TMDataset()
    dataset.fetch_dataset("20NewsGroup")
    dataset.preprocess(model_type="KmeansTM")
    

The specified model_type is optional and further arguments can be specified. Default steps are predefined for all included models. Steps like stopword removal and lemmatizing are automatically performed for models like e.g. LDA.

Model fitting

Fitting a model from STREAM follows a simple, sklearn-like logic and every model can be fit identically.

  1. Choose the model you want to use and train it:

    model = KmeansTM()
    model.fit(dataset, n_topics=20)
    

Depending on the model, check the documentation for hyperparameter settings. To get the topics, simply run:

  1. Get the topics:
    topics = model.get_topics()
    

Evaluation

In this section, we describe the three metrics used to evaluate topic models' performance: Intruder Shift (ISH), Intruder Accuracy (INT), and Average Intruder Similarity (ISIM).

Expressivity

Expressivity, evaluates the meaningfulness of a topic by leveraging stopwords. Stopwords primarily serve a grammatical role and don't contribute to the document's meaning. The steps to calculate Expressivity are as follows:

  1. Compute vector embeddings for all stopwords and calculate their centroid embedding, ${\psi}$.
  2. For each topic, compute the weighted centroid of the top $Z$ words, normalized so that their weights sum up to 1: ${\gamma}k = \frac{1}{Z}\sum{i=1}^{Z} \phi_{k,i}{\omega_i}$.
  3. Calculate the cosine similarity between each topic centroid ${\gamma}_k$ and the stopword centroid ${\psi}$.
  4. The Expressivity metric is then defined as the average similarity across all $K$ topics:

$$\small{EXPRS({\gamma}, {\psi}) = \frac{1}{K} \sum_{k=1}^{K} sim({\gamma}_k, {\psi})}$$

Note that ${\gamma}_k$ is different from ${\mu}_k$, where the latter is the centroid of the document cluster associated with topic $t_k$. Expressivity can vary based on the chosen stopwords, allowing for domain-specific adjustments to evaluate a topic's expressivity based on a custom stopword set.

This approach provides a quantifiable measure of how well a topic conveys meaningful information, distinct from grammatical structure alone.

Intruder Accuracy (INT)

The Intruder Accuracy (INT) metric aims to improve the identification of intruder words within a topic. Here's how it works:

  1. Given the top Z words of a topic, randomly select an intruder word from another topic.
  2. Calculate the cosine similarity between all possible pairs of words within the set of the top Z words and the intruder word.
  3. Compute the fraction of top words for which the intruder has the least similar word embedding using the following formula:

$$\small{INT(t_k) = \frac{1}{Z}\sum_{i=1}^Z {1}(\forall j: sim({\omega}_i, {\hat{\omega}}) < sim({\omega}_i, {\omega}_j))}$$

INT measures how effectively the intruder word can be distinguished from the top words in a topic. A larger value is better.

Average Intruder Similarity (ISIM)

The Average Intruder Similarity (ISIM) metric calculates the average cosine similarity between each word in a topic and an intruder word: $$ISIM(t_k) = \frac{1}{Z} \sum_{i=1}^{Z} sim({\omega}_i, {\hat{\omega}})$$

To enhance the metrics' robustness against the specific selection of intruder words, ISH, INT, and ISIM are computed multiple times with different randomly chosen intruder words, and the results are averaged.

These metrics provide insights into the performance of topic models and their ability to maintain topic coherence and diversity. A smaller value is better.

Intruder Shift (ISH)

The Intruder Shift (ISH) metric quantifies the shift in a topic's centroid when an intruder word is substituted. This process involves the following steps:

  1. Compute the unweighted centroid of a topic and denote it as $\tilde{\boldsymbol{\gamma}}_i$.
  2. Randomly select a word from that topic and replace it with a randomly selected word from a different topic.
  3. Recalculate the centroid of the resulting words and denote it as $\hat{\boldsymbol{\gamma}}_i$.
  4. Calculate the ISH score for a topic by averaging the cosine similarity between $\tilde{{\gamma}}_i$ and $\hat{\boldsymbol{\gamma}}_i$ for all topics using the formula:

$$ISH(T) = \frac{1}{K} \sum_{i=1}^{K} sim(\tilde{{\gamma}}_i, \hat{{\gamma}}_i)$$ A lower ISH score indicates a more coherent and diverse topic model.

To evaluate your model simply use one of the metrics.

from stream_topic.metrics import ISIM, INT, ISH,Expressivity, NPMI

metric = ISIM()
metric.score(topics)

Scores for each topic are available via:

metric.score_per_topic(topics)

To leverage one of the metrics available in octis, simply create a model output that fits within the octis' framework

from octis.evaluation_metrics.diversity_metrics import TopicDiversity

model_output = {"topics": model.get_topics(), "topic-word-matrix": model.get_beta(), "topic-document-matrix": model.get_theta()}

metric = TopicDiversity(topk=10) # Initialize metric
topic_diversity_score = metric.score(model_output)

Similarly to use one of STREAMS metrics for any model, use the topics and occasionally the $\beta$ (topic-word-matrix) of the model to calculate the score.

Hyperparameter optimization

If you want to optimize the hyperparameters, simply run:

model.optimize_and_fit(
    dataset,
    min_topics=2,
    max_topics=20,
    criterion="aic",
    n_trials=20,
)

Visualization

You can also specify to optimize with respect to any evaluation metric from stream_topic. Visualize the results:

from stream_topic.visuals import visualize_topic_model,visualize_topics
visualize_topic_model(
    model, 
    reduce_first=True, 
    port=8051,
    )

Figure Description

Downstream Tasks

The general formulation of a Neural Additive Model (NAM) can be summarized by the equation:

$$ E(y) = h(β + ∑_{j=1}^{J} f_j(x_j)), $$

where $h(·)$ denotes the activation function in the output layer, such as a linear activation for regression tasks or softmax for classification tasks. $x ∈ R^j$ represents the input features, and $β$ is the intercept. The function $f_j : R → R$ corresponds to the Multi-Layer Perceptron (MLP) for the $j$-th feature.

Let's consider $x$ as a combination of categorical and numerical features $x_{tab}$ and document features $x_{doc}$. After applying a topic model, STREAM extracts topical prevalences from documents, effectively transforming the input into $z ≡ (x_{tab}, x_{top})$, a probability vector over documents and topics. Here, $x_{j(tab)}^{(i)}$ indicates the $j$-th tabular feature of the $i$-th observation, and $x_{k(top)}^{(i)}$ represents the $i$-th document's topical prevalence for topic $k$.

For preserving interpretability, the downstream model is defined as:

$$ h(E[y]) = β + ∑_{j=1}^{J} f_j(x_{j(tab)}) + ∑_{k=1}^{K} f_k(x_{k(top)}), $$

In this setup, visualizing the shape function k reveals the impact of a topic on the target variable y. For example, in the context of the Spotify dataset, this could illustrate how a topic influences a song's popularity.

Fitting a downstream model with a pre-trained topic model is straightforward using the PyTorch Trainer class. Subsequently, visualizing all shape functions can be done similarly to the approach described by Agarwal et al. (2021).

How to use

from lightning import Trainer
from stream_topic.NAM import DownstreamModel

# Instantiate the DownstreamModel
downstream_model = DownstreamModel(
    trained_topic_model=topic_model,
    target_column='target',  # Target variable
    task='regression',  # or 'classification'
    dataset=dataset,  
    batch_size=128,
    lr=0.0005
)

# Use PyTorch Lightning's Trainer to train and validate the model
trainer = Trainer(max_epochs=10)
trainer.fit(downstream_model)

# Plotting
from stream_topic.visuals import plot_downstream_model
plot_downstream_model(downstream_model)

Contributing and Testing New Models

We welcome contributions to enhance the functionality of our topic modeling package. To ensure your new models integrate seamlessly, please follow the guidelines and testing instructions provided below.

Steps for Contributing

  1. Fork the Repository:

    • Fork the repository to your GitHub account.
    • Clone the forked repository to your local machine.
    git clone https://github.com/your-username/your-repository.git
    cd your-repository
    
  2. Create a New Branch:

    • Ensure you are on the develop branch and create a new branch for your model development.
    git checkout develop
    git checkout -b new-model-branch
    
  3. Develop Your Model:

    • Navigate to the mypackage/models/ directory.
    • Create your model class file, ensuring it follows the expected structure and naming conventions.
    • Implement the required methods (get_info, fit, predict) and attributes (topic_dict). Optionally, implement beta, theta, or corresponding methods (get_beta, get_theta).

Example Model Structure

Here is an example of how your model class should be structured:

import numpy as np
from mypackage.models.abstract_helper_models.base import BaseModel, TrainingStatus

class ExampleModel(BaseModel):
    def __init__(self, *args, **kwargs):
        super().__init__(*args, **kwargs)
        self._status = TrainingStatus.NOT_STARTED

    def get_info(self):
        return {"model_name": "ExampleModel", "trained": False}

    def any_other_processing_functions(self):
        pass

    def fit(self, dataset, n_topics=3):
        # do what you do during fitting the models
        self._status = TrainingStatus.INITIALIZED
        self._status = TrainingStatus.RUNNING
        self._status = TrainingStatus.SUCCEEDED

    def predict(self, texts):
        return [0] * len(texts)

    # If self.beta or self.theta are not assigned during fitting, plese include these two methods
    def get_beta(self):
        return self.beta

    def get_theta(self):
        return self.theta

Testing Your Model

  1. Install Dependencies:

    • Ensure all dependencies are installed.
    pip install -r requirements.txt
    
  2. Validate Your Model:

    • To validate your model, use tests/validate_new_model.py to include your new model class.
    from tests.model_validation import validate_model
    
    validate_model(NewModel)
    

If this validation fails, it will tell you

Validation Criteria

The following checks are performed during validation:

  • Presence of required methods (get_info, fit, predict).
  • Presence of required attributes (topic_dict).
  • Either presence of optional attributes (beta, theta) or corresponding methods (get_beta, get_theta).
  • Correct shape and sum of theta.
  • Proper status transitions during model fitting.
  • get_info method returns a dictionary with model_name and trained keys.

Refer to the tests/model_validation.py script for detailed validation logic.

Submitting Your Contribution

  1. Commit Your Changes:

    • Commit your changes to your branch.
    git add .
    git commit -m "Add new model: YourModelName"
    
  2. Push to GitHub:

    • Push your branch to your GitHub repository.
    git push origin new-model-branch
    
  3. Create a Pull Request:

    • Go to the original repository on GitHub.
    • Create a pull request from your forked repository and branch.
    • Provide a clear description of your changes and request a review.

We appreciate your contributions and strive to make the integration process as smooth as possible. If you encounter any issues or have questions, feel free to open an issue on GitHub. Happy coding!

If you want to include a new model where these guidelines are not approriate please mark this in your review request.

Citation

If you use this project in your research, please consider citing:

Paper 1 TBD

@article{your_paper_key1,
  title={Your Paper Title},
  author={Your Name and Co-Author's Name},
  journal={Journal/Conference Name},
  year={Year},
  volume={Volume},
  number={Number},
  pages={Pages},
  doi={link_to_doi}
}

Metrics and CEDC

@article{thielmann2024topics,
  title={Topics in the haystack: Enhancing topic quality through corpus expansion},
  author={Thielmann, Anton and Reuter, Arik and Seifert, Quentin and Bergherr, Elisabeth and S{\"a}fken, Benjamin},
  journal={Computational Linguistics},
  pages={1--37},
  year={2024},
  publisher={MIT Press One Broadway, 12th Floor, Cambridge, Massachusetts 02142, USA~…}
}

TNTM

@article{reuter2024probabilistic,
  title={Probabilistic Topic Modelling with Transformer Representations},
  author={Reuter, Arik and Thielmann, Anton and Weisser, Christoph and S{\"a}fken, Benjamin and Kneib, Thomas},
  journal={arXiv preprint arXiv:2403.03737},
  year={2024}
}

DCTE

@inproceedings{thielmann2024human,
  title={Human in the Loop: How to Effectively Create Coherent Topics by Manually Labeling Only a Few Documents per Class},
  author={Thielmann, Anton F and Weisser, Christoph and S{\"a}fken, Benjamin},
  booktitle={Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)},
  pages={8395--8405},
  year={2024}
}

CBC

@inproceedings{thielmann2023coherence,
  title={Coherence based document clustering},
  author={Thielmann, Anton and Weisser, Christoph and Kneib, Thomas and S{\"a}fken, Benjamin},
  booktitle={2023 IEEE 17th International Conference on Semantic Computing (ICSC)},
  pages={9--16},
  year={2023},
  organization={IEEE}

If you use one of the Reddit or GME datasets, consider citing:

@article{kant2024one,
  title={One-way ticket to the moon? An NLP-based insight on the phenomenon of small-scale neo-broker trading},
  author={Kant, Gillian and Zhelyazkov, Ivan and Thielmann, Anton and Weisser, Christoph and Schlee, Michael and Ehrling, Christoph and S{\"a}fken, Benjamin and Kneib, Thomas},
  journal={Social Network Analysis and Mining},
  volume={14},
  number={1},
  pages={121},
  year={2024},
  publisher={Springer}
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

stream_topic-0.1.4.tar.gz (106.6 kB view details)

Uploaded Source

Built Distribution

stream_topic-0.1.4-py3-none-any.whl (144.4 kB view details)

Uploaded Python 3

File details

Details for the file stream_topic-0.1.4.tar.gz.

File metadata

  • Download URL: stream_topic-0.1.4.tar.gz
  • Upload date:
  • Size: 106.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.8.18

File hashes

Hashes for stream_topic-0.1.4.tar.gz
Algorithm Hash digest
SHA256 a493e85cdae0090fafa3f41fea76c1f485795e0a0bbf78835846e96f15bc93a1
MD5 ee912e4e045598fa77d19ff56d0874fb
BLAKE2b-256 76528e3e1874fb0aee3fa3487eca7b61ae198f71d00da2db798959d78cb2ae6f

See more details on using hashes here.

File details

Details for the file stream_topic-0.1.4-py3-none-any.whl.

File metadata

  • Download URL: stream_topic-0.1.4-py3-none-any.whl
  • Upload date:
  • Size: 144.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.8.18

File hashes

Hashes for stream_topic-0.1.4-py3-none-any.whl
Algorithm Hash digest
SHA256 4ebc04eeb69774dba292a27df2129706c3602dc0aff6d9eb962e2e2345f0883f
MD5 629d7d74d87a600db38920551391c9b7
BLAKE2b-256 f82b602e2ed6965cfb178733a0bee0cb6be965bd75ec6c4e263cca667f8d016b

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page