Library to process dumps of knowledge graphs (Wikipedia, DBpedia, Wikidata)
Project description
kgdata
![Documentation](https://pypi-camo.freetls.fastly.net/e11316af6c3f7c7de9ada760f22285010eb40b92/68747470733a2f2f72656164746865646f63732e6f72672f70726f6a656374732f6b67646174612f62616467652f3f76657273696f6e3d6c6174657374267374796c653d666c6174)
KGData is a library to process dumps of Wikipedia, Wikidata. What it can do:
- Clean up the dumps to ensure the data is consistent (resolve redirect, remove dangling references)
- Create embedded key-value databases to access entities from the dumps.
- Extract Wikidata ontology.
- Extract Wikipedia tables and convert the hyperlinks to Wikidata entities.
- Create Pyserini indices to search Wikidata’s entities.
- and more
For a full documentation, please see the website.
Installation
From PyPI (using pre-built binaries):
pip install kgdata
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
kgdata-3.3.1.tar.gz
(62.6 kB
view hashes)
Built Distribution
kgdata-3.3.1-py3-none-any.whl
(88.3 kB
view hashes)