Skip to main content

Record your llm calls and make your notebooks fast again.

Project description

reclm

When building AI based tooling and packaging we often call LLMs while prototyping and testing our code. A single LLM call can take 100’s of ms to run and the output isn’t deterministic. This can really slow down development especially if our notebook contains many LLM calls 😞.

While LLMs are new, working with external APIs in our code isn’t. Plenty of tooling already exists that make working with APIs much easier. For example, Python’s unittest mock object is commonly used to simulate or mock an API call so that it returns a hardcoded response. This works really well in the traditional Python development workflow and can make our tests fast and predictable.

However, it doesn’t work well in the nbdev workflow where oftentimes we’ll want to quickly run all cells in our notebook while we’re developing our code. While we can use mocks in our test cells we don’t want our exported code cells to be mocked. This leaves us with two choices:

  • we temporarily mock our exported code cells but undo the mocking before we export these cells.
  • we do nothing and just live with notebooks that take a long time to run.

Both options are pretty terrible as they pull us out of our flow state and slow down development 😞.

reclm builds on the underlying idea of mocks but adapts them to exploratory workflows. It initializes each sdk client (e.g. AsyncAnthropic) with a custom http_client. This http_client intercepts each LLM call, caches the response, and returns this cached response if the user makes the same LLM call again.

Usage

To use reclm

  • install the package: pip install git+https://github.com/AnswerDotAI/reclm.git
  • import the package from reclm.core import enable_reclm in each notebook
  • add enable_reclm() to the top of each notebook

Note: enable_reclm should be added after you import the OpenAI and/or Anthropic SDK.

Every LLM call you make using OpenAI/Anthropic will now be cached in reclm.json.

[!NOTE]

Cache Location

If you’re using reclm in an nbdev project the cache (i.e. reclm.json) will be added to your project’s root dir. For all other projects the cache will be added to the current working directory. You can set a custom location for the cache by passing cache_dir to enable_reclm (e.g. enable_reclm(cache_dir='/path/to/cache')).

Your Tests

nbdev_test will automatically read from the cache. However, if your notebooks contain LLM calls that haven’t been cached, nbdev_test will call the OpenAI/Anthropic APIs and then cache the responses.

Cleaning the cache

It is recommended that you clean the cache before committing it. This will remove any stale LLM requests you’ve accumulated during development.

To clean the cache, run update_reclm_cache from your project’s root directory.

Note: Your LLM request/response data is stored in a file called reclm.json. Depending on your setup, it might be stored in your project’s root dir or your current working directory.

Note: All request headers are removed so it is safe to include this file in your version control system (e.g. git). In fact, it is expected that you’ll include this file in your vcs. There is one scenario where your api key could end up in the cache. This occurs if you make an LLM call that includes your api key in the response such as an authentication error.

Reclm Tests

The reclm test suite is found in test_reclm.py. Use the command python test_reclm.py to run the test suite.

Each test checks a specific SDK endpoint (e.g. anthropic streaming). If reclm is working correctly the SDK endpoint won’t be called and instead the response will be pulled from reclm.json.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

reclm-0.0.6.tar.gz (11.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

reclm-0.0.6-py3-none-any.whl (10.7 kB view details)

Uploaded Python 3

File details

Details for the file reclm-0.0.6.tar.gz.

File metadata

  • Download URL: reclm-0.0.6.tar.gz
  • Upload date:
  • Size: 11.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.12.5

File hashes

Hashes for reclm-0.0.6.tar.gz
Algorithm Hash digest
SHA256 32212d6dad8cd419446dff22b6bc39bd27d28b552c8d5b47bf6a685b656c1cad
MD5 b3f11e0736019fb44d403801d4320f65
BLAKE2b-256 2b70c57e9fa277c142fc584e09780c0e9b31d4aa75e645137fcd3b7efd3c8bfc

See more details on using hashes here.

File details

Details for the file reclm-0.0.6-py3-none-any.whl.

File metadata

  • Download URL: reclm-0.0.6-py3-none-any.whl
  • Upload date:
  • Size: 10.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.12.5

File hashes

Hashes for reclm-0.0.6-py3-none-any.whl
Algorithm Hash digest
SHA256 7e667c66ee3636b0f96f96c1beea19c3636f376f113dfa8f5eb6e99ad8069ee5
MD5 65a51751602eca1906f4e3d0439fcbb0
BLAKE2b-256 3cde711e6ac46eb94cb5338ed741bde8a9a355fb53bd544c540c5d7d639b2a40

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page