Skip to main content

Collection of high-level tools to simplify everyday development tasks, with a focus on AI/ML

Project description

fmtr.tools

A collection of high-level tools to simplify everyday development tasks, with a slight focus on full-stack AI/ML.

This repository is an attempt to provide a one-stop source for a wide range of utilities and tools designed to streamline a typical, modern development workflow. There is an emphasis on a lean and nimble approach to dependencies, which tries to strike a balance between powerful functionality while avoiding unnecessary bloat.

Why?

Personally, I'm grossly impatient, and simply resent writing the same code, however simple, in multiple projects.

This could be trivial stuff like reading an integer from an environment variable (while handling errors gracefully) - or more complex ones (like just wanting a simple parallel-processing function without writing Queues, or remembering which libraries you need to do it for you).

At the same time, I find that traditional tools collections inevitably become bloated and unwieldy over time, so wanted something with a somewhat sophisticated approach to dependencies.

Key Features

  • Wide-Ranging Utilities: The collection includes tools for configuration, data types, environment management, functions, hashing, importing, iterating, JSON handling, path manipulation, platform-specific operations, randomness, and string operations.
  • Lean Dependencies: Dependencies are managed via extras, allowing you to install only what you need. Missing dependencies are handled in a clear way, telling you what's missing and how to install it.

Installing

The base library can be installed like this:

pip install fmtr.tools

Usage

Some simple import and usage examples

Read an integer from an environment variable and write it to a (human-readable) JSON file

from fmtr import tools
from fmtr.tools import Path

value=tools.env.get_int('MY_VALUE',default=None)
data=dict(value=value)
Path('data.json').write_json(data)

Zero-faff parallel multi-processing

Install the extra:

pip install fmtr.tools[parallel] --upgrade
from fmtr.tools import parallel

def expensive_computation(n):
    import math
    result = 0
    for i in range(1, n + 1):
        result += math.sqrt(i) * math.sin(i) * math.log(i)
    return result

if __name__ == '__main__':
    results=parallel.apply(expensive_computation, [10_000] * 1_000)

Extras

Most tools require no additional dependencies, but for any that do, you can add them like this:

pip install fmtr.tools[<extra>] --upgrade

If you try to use a module without the required extras, you'll get a message telling you which one is needed:

MissingExtraError: The current module is missing dependencies. To install them, run: `pip install fmtr.tools[logging] --upgrade`

Modules

The included modules, plus any extra requirements, are as follows:

  • tools.ai: Manages bulk inference for LLMs using dynamic batching. Includes classes for managing prompt encoding, generating outputs, and handling tool calls, with support for both local and remote models. Uses Pytorch and Transformers for model operations, and provides functionality for encoding prompts, generating responses, and applying tool functions.
    • Extras: ai
  • tools.config: Base config class with overridable field processors.
    • Extras: None
  • tools.dataclass: Utilities for extracting and filtering fields and metadata from dataclasses, with support for applying filters and retrieving enabled fields based on metadata attributes.
    • Extras: None
  • tools.datatype
    • Extras: None
  • tools.dm: Defines custom data modelling base classes for creating Pydantic models with error-tolerant deserialization from JSON (e.g. when output from an LLM).
    • Extras: dm
  • tools.environment: Tools for managing environment variables, including functions to retrieve variables with type conversions and default values. Features include environment variable fetching, handling missing variables, and creating type-specific getters for integers, floats, booleans, dates, and paths.
    • Extras: None
  • tools.env: Alias of tools.environment.
    • Extras: None
  • tools.function: Utilities for combining and splitting arguments and keyword arguments.
    • Extras: None
  • tools.hash: String hashing
    • Extras: None
  • tools.hfh: Utilities for caching and managing Hugging Face model repositories: setting tokens, downloading snapshots, tagging repositories, and retrieving local cache paths.
    • Extras: hfh
  • tools.html: Utilities for converting HTML documents to plain text.
    • Extras: html
  • tools.interface: Provides a base class for building Streamlit interfaces with a class-based structure.
    • Extras: interface
  • tools.iterator: Pivoting/unpivoting data structures
    • Extras: None
  • tools.json: Serialisation/deserialisation to human-readable, unicode JSON.
    • Extras: None
  • tools.merge: Utility for recursively merging multiple dictionaries or objects using the DeepMerge library.
    • Extras: merge
  • tools.name: Generates random memorable names (similar to Docker Container names) by combining an adjective with a surname.
    • Extras: None
  • tools.openai: Utilities for interacting with the OpenAI API, simple text-to-text output, etc.
    • Extras: openai.api
  • tools.Path: Enhanced pathlib.Path object with additional functionality for Windows-to-Unix path conversion, reading/writing JSON and YAML files, and convenient directory creation with parent directories. Includes methods for obtaining paths to modules and temporary directories.
    • Extras: None
  • tools.platform: Detecting if host is WSL, Docker etc.
    • Extras: None
  • tools.ContextProcess: Manages a function running in a separate process using a context manager. Provides methods to start, stop, and restart the process, with configurable restart delays. Useful for ensuring clean process management and automatic stopping when the context manager exits.
    • Extras: None
  • tools.random: Provides additional functions for random number generation and selection, useful for data augmentation.
    • Extras: None
  • tools.semantic: Manages semantic similarity operations using Sentence Transformers: loading a pre-trained model, vectorizing a text corpus, and retrieving the top matches based on similarity scores for a given query string.
    • Extras: semantic
  • tools.string: Provides utilities for handling string formatting.
    • Extras: None
  • tools.logging: Configures and initializes a logger using the Logfire library to log to an OpenTelemetry consumer.
    • Extras: logging
  • tools.logger: Prefabricated logger object, suitable for most projects: service name, colour-coded, timestamped, etc.
    • Extras: logging
  • tools.augmentation: Data augmentation stub.
    • Extras: augmentation
  • tools.Container: Runs a Docker container within a context manager, ensuring the container is stopped and removed when the context is exited.
    • Extras: docker.api
  • tools.parallel: Provides utilities for parallel computation using Dask. Supports executing functions across multiple workers or processes, handles different data formats, and options for progress display and parallelism configuration.
    • Extras: parallel
  • tools.profiling: Context-based code timing.
    • Extras: profiling
  • tools.tokenization: Provides utilities for creating and configuring tokenizers using the Tokenizers library. Iincludes functions for training both word-level and byte-pair encoding (BPE) tokenizers, applying special formatting and templates, and managing tokenizer configurations such as padding, truncation, and special tokens.
    • Extras: tokenization
  • tools.unicode: Simple unicode decoding (via Unidecode).
    • Extras: unicode

Contribution

Any contributions would be most welcome! If you have a utility that fits well within this collection, or improvements to existing tools, feel free to open a pull request.

License

This project is licensed under the Apache License Version 2.0. See the LICENSE file for more details.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

fmtr.tools-1.0.28.tar.gz (39.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

fmtr.tools-1.0.28-py3-none-any.whl (48.0 kB view details)

Uploaded Python 3

File details

Details for the file fmtr.tools-1.0.28.tar.gz.

File metadata

  • Download URL: fmtr.tools-1.0.28.tar.gz
  • Upload date:
  • Size: 39.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.12.3

File hashes

Hashes for fmtr.tools-1.0.28.tar.gz
Algorithm Hash digest
SHA256 4a9817d954c9994435db77be87c04776c56c7c40cec2bf62cdcd483d28145143
MD5 1675b2d5d2c3889be5b6e3eaf4a830d9
BLAKE2b-256 3766e097b96f40e2577d5c710f34fbb9a2502aa0a18f190862254d7ea6b3a031

See more details on using hashes here.

File details

Details for the file fmtr.tools-1.0.28-py3-none-any.whl.

File metadata

  • Download URL: fmtr.tools-1.0.28-py3-none-any.whl
  • Upload date:
  • Size: 48.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.12.3

File hashes

Hashes for fmtr.tools-1.0.28-py3-none-any.whl
Algorithm Hash digest
SHA256 82e1b88626258083bb2b2cf02f0298de415991c824f337bf77529c4b5de4658f
MD5 811ac8c93d16d4334aaf4456999eab43
BLAKE2b-256 2da06f68672b294dfceb65ff5f40b9ca037e61f5c3cb6fe73c11f91d92b0bd20

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page