Skip to main content

Build a search index across content from multiple SQLite database tables and run faceted searches against it using Datasette

Project description

datasette-metasearch

PyPI Changelog Tests License

Build a search index across content from multiple SQLite database tables and run faceted searches against it using Datasette.

Motivation

Different datasets may contain overlapping data but may not conform to the same schema. Some datasets may contain more information than others that we want to display when searching across datasets. We might also want to see statistics or facets across those datasettes (like how many records in a particular year or made by a particular person). datasette-metasearch enables this pattern with a config file, rather than trying to build a pipeline to transform each dataset into a common format then building a bespoke query interface for those, we can specify the fields we want to index and search, how to transform them into the com o

The motivation for this was to join government spending datasets so they can be easily queried.

Example

A live example of this plugin is running at https://datasette.io/-/beta - configured using this YAML file.

Read more about how this example works in Building a search engine for datasette.io.

Installation

Install this tool like so:

$ pip install datasette-metasearch

Usage

Run the indexer using the datasette-metasearch command-line tool:

$ datasette-metasearch index dogsheep.db config.yml

The config.yml file contains details of the databases and document types that should be indexed:

NOTE: the database storing the search index must be different from the ones containing the data to be indexed

twitter.db:
    tweets:
        sql: |-
            select
                tweets.id as key,
                'Tweet by @' || users.screen_name as title,
                tweets.created_at as timestamp,
                tweets.full_text as search_1
            from tweets join users on tweets.user = users.id
    users:
        sql: |-
            select
                id as key,
                name || ' @' || screen_name as title,
                created_at as timestamp,
                description as search_1
            from users

This will create a search_index table in the dogsheep.db database populated by data from those SQL queries.

By default the search index that this tool creates will be configured for Porter stemming. This means that searches for words like run will match documents containing runs or running.

If you don't want to use Porter stemming, use the --tokenize none option:

$ datasette-metasearch index dogsheep.db config.yml --tokenize none

You can pass other SQLite tokenize argumenst here, see the SQLite FTS tokenizers documentation.

Columns

The columns that can be returned by our query are:

  • key - a unique (within that type) primary key
  • title - the title for the item
  • timestamp - an ISO8601 timestamp, e.g. 2020-09-02T21:00:21
  • search_1 - a larger chunk of text to be included in the search index
  • category - an integer category ID, see below
  • is_public - an integer (0 or 1, defaults to 0 if not set) specifying if this is public or not

Public records are things like your public tweets, blog posts and GitHub commits.

Datasette plugin

Run datasette install datasette-metasearch (or use pip install datasette-metasearch in the same environment as Datasette) to install the Dogsheep Beta Datasette plugin.

Once installed, a custom search interface will be made available at /-/beta. You can use this interface to execute searches.

The Datasette plugin has some configuration options. You can set these by adding the following to your metadata.json configuration file:

{
    "plugins": {
        "datasette-metasearch": {
            "database": "beta",
            "config_file": "datasette-metasearch.yml",
            "template_debug": true
        }
    }
}

The configuration settings for the plugin are:

  • database - the database file that contains your search index. If the file is beta.db you should set database to beta.
  • config_file - the YAML file containing your Dogsheep Beta configuration.
  • template_debug - set this to true to enable debugging output if errors occur in your custom templates, see below.

Custom results display

Each indexed item type can define custom display HTML as part of the config.yml file. It can do this using a display key containing a fragment of Jinja template, and optionally a display_sql key with extra SQL to execute to fetch the data to display.

Here's how to define a custom display template for a tweet:

twitter.db:
    tweets:
        sql: |-
            select
                tweets.id as key,
                'Tweet by @' || users.screen_name as title,
                tweets.created_at as timestamp,
                tweets.full_text as search_1
            from tweets join users on tweets.user = users.id
        display: |-
            <p>{{ title }} - tweeted at {{ timestamp }}</p>
            <blockquote>{{ search_1 }}</blockquote>

This example reuses the value that were stored in the search_index table when the indexing query was run.

To load in extra values to display in the template, use a display_sql query like this:

twitter.db:
    tweets:
        sql: |-
            select
                tweets.id as key,
                'Tweet by @' || users.screen_name as title,
                tweets.created_at as timestamp,
                tweets.full_text as search_1
            from tweets join users on tweets.user = users.id
        display_sql: |-
            select
                users.screen_name,
                tweets.full_text,
                tweets.created_at
            from
                tweets join users on tweets.user = users.id
            where
                tweets.id = :key
        display: |-
            <p>{{ display.screen_name }} - tweeted at {{ display.created_at }}</p>
            <blockquote>{{ display.full_text }}</blockquote>

The display_sql query will be executed for every search result, passing the key value from the search_index table as the :key parameter and the user's search term as the :q parameter.

This performs well because many small queries are efficient in SQLite.

If an error occurs while rendering one of your templates the search results page will return a 500 error. You can use the template_debug configuration setting described above to instead output debugging information for the search results item that experienced the error.

Displaying maps

This plugin will eventually include a number of useful shortcuts for rendering interesting content.

The first available shortcut is for displaying maps. Make your custom content output something like this:

<div
    data-map-latitude="{{ display.latitude }}"
    data-map-longitude="{{ display.longitude }}"
    style="display: none; float: right; width: 250px; height: 200px; background-color: #ccc;"
></div>

JavaScript on the page will look for any elements with data-map-latitude and data-map-longitude and, if it finds any, will load Leaflet and convert those elements into maps centered on that location. The default zoom level will be 12, or you can set a data-map-zoom attribute to customize this.

Development

To set up this plugin locally, first checkout the code. Then create a new virtual environment:

cd datasette-metasearch
python3 -mvenv venv
source venv/bin/activate

Or if you are using pipenv:

pipenv shell

Now install the dependencies and tests:

pip install -e '.[test]'

To run the tests:

pytest

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

datasette_metasearch-1.0.1rc2.tar.gz (17.1 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

datasette_metasearch-1.0.1rc2-py3-none-any.whl (11.8 kB view details)

Uploaded Python 3

File details

Details for the file datasette_metasearch-1.0.1rc2.tar.gz.

File metadata

  • Download URL: datasette_metasearch-1.0.1rc2.tar.gz
  • Upload date:
  • Size: 17.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.12.9

File hashes

Hashes for datasette_metasearch-1.0.1rc2.tar.gz
Algorithm Hash digest
SHA256 3c193d482c3058576e17d5b5cc1ead5a08e3721591a1c0b88010f85a3414e423
MD5 6e41da45a0dd2e4c417c9365d620da45
BLAKE2b-256 0f6994e27068d5556609ce90f9e1d60bbe759eb8265afcbf9f2ed0efc4c2fb88

See more details on using hashes here.

Provenance

The following attestation bundles were made for datasette_metasearch-1.0.1rc2.tar.gz:

Publisher: publish.yml on xrendan/datasette-metasearch

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file datasette_metasearch-1.0.1rc2-py3-none-any.whl.

File metadata

File hashes

Hashes for datasette_metasearch-1.0.1rc2-py3-none-any.whl
Algorithm Hash digest
SHA256 6327feef3e127f1ab150625b364854f83380a64c8df4d170efeec94c0ab58517
MD5 065b5c93872ff34357b2beee7e13e057
BLAKE2b-256 2c5cde30d9cec319d8466872ab2b64bfc972f51095505bcaf20ddd2a66173931

See more details on using hashes here.

Provenance

The following attestation bundles were made for datasette_metasearch-1.0.1rc2-py3-none-any.whl:

Publisher: publish.yml on xrendan/datasette-metasearch

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page