Skip to main content

Wikidspark is an API for querying the Wikidata item database, allowing uses to search either directly by name or ID, or construct SPARQL queries

Project description

WikidSpark

WikidSparkcodecovCodeFactorGitHub

WikidSpark is a python module with the aim of providing easy access to the WikiData SPARQL database. The aim is to provide a friendly method which allows those unfamiliar with SPARQL and databases in general to still access the information.

Querying WikiData

There are currently two methods for retrieving data from the WikiData site.

Direct Query

The first is query the site directly by passing the relevant id (QXYZ) to the query service. Two methods can be used to do this: get_by_id or get_by_name. Both of these methods come with two optional arguments for filtering the results:

from wikidspark.query import get_by_id

# Search for the Douglas Adams WikiData entry
dga_id = 'Q42'

dga_query_full = get_by_id(dga_id)
dga_query_language = get_by_id(dga_id, language='french')

the retrieve-by-name method currently makes use of the pre-requisite wikipedia model to find items relevant to the search, making it limited in terms of only displaying those entries which have an article attached to them. The function returns the first match it finds:

from wikidspark.query import get_by_name

dga_query_full = get_by_name('Douglas Adams')
dga_query_filtered = get_by_name('Douglas Adams', language='english', keys=['labels', 'descriptions'])

for a wider search it is recommended to use the find_item function to firstly return the relevant id, then use the get_by_id function. The find_item function has additional arguments to fetch only the first result (the default) or a list of matches, and to specify the wikipedia language:

dga_id = find_id('Douglas Adams')
london_ids = find_id('London', get_first=False, language='english')

Built Query

This method uses a class QueryBuilder to construct a SPARQL query to be sent to the WikiData query service. The query is built in stages, firstly by defining an instance of the builder (with the optional argument language),then adding conditions before finally fetching the results as a QItem object. This object allows you to view the data as XML, JSON or a Pandas Dataframe. The data returned currently has the additional properties Label, Description and AltLabel which are turned off by default but can be activated.

# Find first 100 books
from wikidspark.query import QueryBuilder
my_query = QueryBuilder("english")

my_query.instance_of('book')
my_query.Description = True
my_query.Label = True

result = my_query.get(limit=100)
json_res = result.json
xml_res  = result.xml
df_res   = result.dataframe

You can also search using a property ID and string value:

# Find Richmond station via CRS code 'RMD'
from wikidspark.query import QueryBuilder
my_query = QueryBuilder("english")

my_query.Description = True
my_query.Label = True
my_query.property_equals("P4755", "RMD")

df_res   = result.dataframe

The functions available to the QueryBuilder class (e.g. instance_of) are based on an extensive list of WikiData properties, the full list can be fetched as a dataframe:

QueryBuilder().list_properties()

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

wikidspark-0.3.0.tar.gz (145.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

wikidspark-0.3.0-py3-none-any.whl (150.3 kB view details)

Uploaded Python 3

File details

Details for the file wikidspark-0.3.0.tar.gz.

File metadata

  • Download URL: wikidspark-0.3.0.tar.gz
  • Upload date:
  • Size: 145.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.12.8

File hashes

Hashes for wikidspark-0.3.0.tar.gz
Algorithm Hash digest
SHA256 5b10a111369ef478f3a946254fcf755e843c866756ae34747244946f9f2992a3
MD5 bb4452b135c7d50a2e6449d14c11e149
BLAKE2b-256 c0be94bf1fa941331587058fae913c8df137536e610571707f2c6068937ada25

See more details on using hashes here.

Provenance

The following attestation bundles were made for wikidspark-0.3.0.tar.gz:

Publisher: deploy.yml on artemis-beta/wikidspark

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file wikidspark-0.3.0-py3-none-any.whl.

File metadata

  • Download URL: wikidspark-0.3.0-py3-none-any.whl
  • Upload date:
  • Size: 150.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.12.8

File hashes

Hashes for wikidspark-0.3.0-py3-none-any.whl
Algorithm Hash digest
SHA256 dd17d8fe4c344a2497312bae3c776a52095e97e43807fa87988cc88d53842b3a
MD5 5be1237d0b812fbad6030709c3c13a79
BLAKE2b-256 fc29492e96ac945521c807fe43167894f7018ddb606b43ad741795a7e07a0f6b

See more details on using hashes here.

Provenance

The following attestation bundles were made for wikidspark-0.3.0-py3-none-any.whl:

Publisher: deploy.yml on artemis-beta/wikidspark

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page