Skip to main content

A python program that can be run from the command line, and used to search climate policy documents.

Project description

climate-search-cli

A python program that can be run from the command line, and used to search climate policy documents.

Task Overview

Created as an interview technical challenge. The task is to create a cli tool that can be used to search summaries of climate documents.

The cli needed to have the following functionality:

  • Load & validate documents into a database at the command line
  • Query the documents returning a sequence of document objects
  • Display the documents and some statistics about them as output
  • Order by relevance using a relevancy score

Evaluation Criteria

The following are the items that are being evaluated:

  • Readability
  • Maintainability
  • Functionality
  • Efficiency
  • Modularity
  • Commenting and documentation
  • Testing Strategy

Run

Run poetry to install dependencies, (see below for other ways of running):

poetry install
poetry shell

Data can be loaded via:

cs load

This will also output errors to the same directory, and load the data into a database. A custom file can also be loaded using the --localpath argument.

Data can then be queried by passing keywords with the retrieve command:

cs retrieve -k green -k energy

This will display the policies that match. Results can also be sorted with:

cs retrieve -k forests --sort

The test directory contains both unit and integration tests, these can be run with pytest:

pytest

Solution Overview

I decided to use click as the cli tool for this project. As well as sqlite as a backend, both of these are simple and portable, although if I could start again, I'd be keen to use a database that had support for arrays. Transformations and schema definitions where done in pandas for convenience, I originally started going down the path of having multiple tables in the database, but decided this was over optimising for what was needed with the given timeframe. Having just one table meant pandas was a straightoferward option for defining the table. The search relevency implementation is just a quick tfidf algorithm on the results.

Time taken

I worked intermittently on this over the course of a couple of days. I think the total time actively working on the solution was about 6 hours. (Not including time spent reading the brief, researching and planning). I could keep going, but I went over the suggested timeframe, so I'm leaving it here. Some key items I'd like to improve include error handling and the relevancy algorithm.

alternate ways of running

Docker

This can also be run via docker:

docker build -t climate-search-cli:latest .
docker run climate-search-cli:latest load
docker run climate-search-cli:latest retrieve -k cycling -k health --sort

pypi

Also available on pypi:

pip install climate-search-cli

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

climate_search_cli-0.1.4.tar.gz (624.0 kB view hashes)

Uploaded Source

Built Distribution

climate_search_cli-0.1.4-py3-none-any.whl (628.8 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page