Wikidspark is an API for querying the Wikidata item database, allowing uses to search either directly by name or ID, or construct SPARQL queries
Project description
WikidSpark
WikidSpark is a python module with the aim of providing easy access to the WikiData SPARQL database. The aim is to provide a friendly method which allows those unfamiliar with SPARQL and databases in general to still access the information.
Querying WikiData
There are currently two methods for retrieving data from the WikiData site.
Direct Query
The first is query the site directly by passing the relevant id (QXYZ) to the query service. Two methods can be used to do this: get_by_id or get_by_name. Both of these methods come with two optional arguments for filtering the results:
from wikidspark.query import get_by_id
# Search for the Douglas Adams WikiData entry
dga_id = 'Q42'
dga_query_full = get_by_id(dga_id)
dga_query_language = get_by_id(dga_id, language='french')
the retrieve-by-name method currently makes use of the pre-requisite wikipedia model to find items relevant to the search, making it limited in terms of only displaying those entries which have an article attached to them. The function returns the first match it finds:
from wikidspark.query import get_by_name
dga_query_full = get_by_name('Douglas Adams')
dga_query_filtered = get_by_name('Douglas Adams', language='english', keys=['labels', 'descriptions'])
for a wider search it is recommended to use the find_item function to firstly return the relevant id, then use the get_by_id function. The find_item function has additional arguments to fetch only the first result (the default) or a list of matches, and to specify the wikipedia language:
dga_id = find_id('Douglas Adams')
london_ids = find_id('London', get_first=False, language='english')
Built Query
This method uses a class QueryBuilder to construct a SPARQL query to be sent to the WikiData query service. The query is built in stages, firstly by defining an instance of the builder (with the optional argument language),then adding conditions before finally fetching the results as a QItem object. This object allows you to view the data as XML, JSON or a Pandas Dataframe. The data returned currently has the additional properties Label, Description and AltLabel which are turned off by default but can be activated.
# Find first 100 books
from wikidspark.query import QueryBuilder
my_query = QueryBuilder("english")
my_query.instance_of('book')
my_query.Description = True
my_query.Label = True
result = my_query.get(limit=100)
json_res = result.json
xml_res = result.xml
df_res = result.dataframe
You can also search using a property ID and string value:
# Find Richmond station via CRS code 'RMD'
from wikidspark.query import QueryBuilder
my_query = QueryBuilder("english")
my_query.Description = True
my_query.Label = True
my_query.property_equals("P4755", "RMD")
df_res = result.dataframe
The functions available to the QueryBuilder class (e.g. instance_of) are based on an extensive list of WikiData properties, the full list can be fetched as a dataframe:
QueryBuilder().list_properties()
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file wikidspark-0.3.0.tar.gz.
File metadata
- Download URL: wikidspark-0.3.0.tar.gz
- Upload date:
- Size: 145.8 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.12.8
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
5b10a111369ef478f3a946254fcf755e843c866756ae34747244946f9f2992a3
|
|
| MD5 |
bb4452b135c7d50a2e6449d14c11e149
|
|
| BLAKE2b-256 |
c0be94bf1fa941331587058fae913c8df137536e610571707f2c6068937ada25
|
Provenance
The following attestation bundles were made for wikidspark-0.3.0.tar.gz:
Publisher:
deploy.yml on artemis-beta/wikidspark
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
wikidspark-0.3.0.tar.gz -
Subject digest:
5b10a111369ef478f3a946254fcf755e843c866756ae34747244946f9f2992a3 - Sigstore transparency entry: 165339147
- Sigstore integration time:
-
Permalink:
artemis-beta/wikidspark@fa0994feb1f006234793979865dffeeef59a1697 -
Branch / Tag:
refs/tags/v0.3.0 - Owner: https://github.com/artemis-beta
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
deploy.yml@fa0994feb1f006234793979865dffeeef59a1697 -
Trigger Event:
push
-
Statement type:
File details
Details for the file wikidspark-0.3.0-py3-none-any.whl.
File metadata
- Download URL: wikidspark-0.3.0-py3-none-any.whl
- Upload date:
- Size: 150.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.12.8
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
dd17d8fe4c344a2497312bae3c776a52095e97e43807fa87988cc88d53842b3a
|
|
| MD5 |
5be1237d0b812fbad6030709c3c13a79
|
|
| BLAKE2b-256 |
fc29492e96ac945521c807fe43167894f7018ddb606b43ad741795a7e07a0f6b
|
Provenance
The following attestation bundles were made for wikidspark-0.3.0-py3-none-any.whl:
Publisher:
deploy.yml on artemis-beta/wikidspark
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
wikidspark-0.3.0-py3-none-any.whl -
Subject digest:
dd17d8fe4c344a2497312bae3c776a52095e97e43807fa87988cc88d53842b3a - Sigstore transparency entry: 165339150
- Sigstore integration time:
-
Permalink:
artemis-beta/wikidspark@fa0994feb1f006234793979865dffeeef59a1697 -
Branch / Tag:
refs/tags/v0.3.0 - Owner: https://github.com/artemis-beta
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
deploy.yml@fa0994feb1f006234793979865dffeeef59a1697 -
Trigger Event:
push
-
Statement type: