Skip to main content

Benchmarking topic models for a paper

Project description

topic-benchmark

Just Benchmarking Topic Models :)

Todo:

  • Run benchmark with these models and upload the results:
    • [ x ] all-MiniLM-L6-v2
    • all-mpnet-base-v2 ⌛
    • sentence-transformers/average_word_embeddings_glove.6B.300d ⌛
    • intfloat/e5-large-v2 (OR intfloat/multilingual-e5-large-instruct, to my knowledge, they are the same size, but this one performs way better on MTEB)
  • [ x ] Implement pretty printing and formatting to Latex and MD tables for results.
  • [ x ] (Maybe) Implement speed tracking.

Usage:

pip install topic-benchmark

python3 -m topic_benchmark run -e "embedding_model_name"

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

topic_benchmark-0.2.2.tar.gz (9.8 kB view hashes)

Uploaded Source

Built Distribution

topic_benchmark-0.2.2-py3-none-any.whl (14.6 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page