Benchmarking topic models for a paper
Project description
topic-benchmark
Just Benchmarking Topic Models :)
Todo:
- Run benchmark with these models and upload the results:
- all-MiniLM-L6-v2 ⌛
- all-mpnet-base-v2 ⌛
- sentence-transformers/average_word_embeddings_glove.6B.300d ⌛
- intfloat/e5-large-v2 (OR intfloat/multilingual-e5-large-instruct, to my knowledge, they are the same size, but this one performs way better on MTEB)
- Implement pretty printing and formatting to Latex and MD tables for results.
- (Maybe) Implement speed tracking.
Usage:
pip install topic-benchmark
python3 -m topic_benchmark run -e "embedding_model_name"
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
topic_benchmark-0.2.1.tar.gz
(9.1 kB
view hashes)
Built Distribution
Close
Hashes for topic_benchmark-0.2.1-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 16309cef2d90b0e1899d26e1a6762aacd2da27a2c683b04c132054642d66192d |
|
MD5 | e39c45be949d3b75e7c82a1e75201ec9 |
|
BLAKE2b-256 | f76d48b12a203b3a96e359d003c420e1aba0661984aac6d6471c5153a4bc7fed |