Skip to main content

python interface for interacting with flashbots mempool dumpster

Project description

image

absorb 🧽🫧🫧

the sovereign dataset manager

absorb makes it easy to 1) collect, 2) query, 3) manage, and 4) customize datasets from nearly any data source

features

  • limitless dataset library: access to millions of datasets across 16 diverse data sources
  • intuitive cli+python interfaces: collect or query any dataset in a single line of code
  • maximal modularity: built on open standards for frictionless integration with other tools
  • easy extensibility: add new datasets or data sources with just a few lines of code

Contents

  1. Installation
  2. Example Usage i. Command Line ii. Python
  3. Supported Data sources
  4. Output Format
  5. Configuration

Installation

uv install paradigm_absorb

Example Usage

Example Command Line Usage

# collect dataset and save as local files
absorb collect kalshi

# list datasets that are collected or available
absorb ls

# show schemas of dataset
absorb schema kalshi

# create new custom dataset
absorb new custom_dataset

# upload custom dataset
absorb upload custom_dataset

Example Python Usage

import absorb

# collect dataset and save as local files
absorb.collect('kalshi')

# list datasets that are collected or available
datasets = absorb.list()

# get schemas of dataset
schema = absorb.schema('kalshi')

# load dataset as polars DataFrame
df = absorb.load('kalshi')

# scan dataset as polars LazyFrame
lf = absorb.scan('kalshi')

# create new custom dataset
absorb.new('custom_dataset')

# upload custom dataset
absorb.upload('custom_dataset')

Supported Data Sources

absorb collects data from each of these sources:

To list all available datasets and data sources, type absorb ls on the command line.

Output Format

To display information about the schema and other metadata of a dataset, type absorb help <DATASET> on the command line.

absorb stores each dataset as a collection of parquet files.

Datasets can be stored in any location on your disks, and absorb will use symlinks to organize those files in the TRUCK_ROOT tree.

the TRUCK_ROOT filesystem directory is organized as:

{TRUCK_ROOT}/
    datasets/
        <source>/
            tables/
                <datatype>/
                    {filename}.parquet
                table_metadata.json
            repos/
                {repo_name}/
    absorb_config.json

Configuration

absorb uses a config file to specify which datasets to track.

Schema of absorb_config.json:

{
    'tracked_tables': list[TrackedTable]
}

schema of dataset_config.json:

{
    "name": str,
    "definition": str,
    "parameters": dict[str, Any],
    "repos": [str]
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

paradigm_absorb-0.2.0.tar.gz (46.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

paradigm_absorb-0.2.0-py3-none-any.whl (68.2 kB view details)

Uploaded Python 3

File details

Details for the file paradigm_absorb-0.2.0.tar.gz.

File metadata

  • Download URL: paradigm_absorb-0.2.0.tar.gz
  • Upload date:
  • Size: 46.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: python-requests/2.32.3

File hashes

Hashes for paradigm_absorb-0.2.0.tar.gz
Algorithm Hash digest
SHA256 f662675bfe69b7445d1e719b2d9c77eb92cdeb29e64a7bd4c08a89c3948375a5
MD5 5043be0836469657c1d8ebad76ab59c1
BLAKE2b-256 a39de71cb65c317555fba8bd6af74dd2b284b6b9d7659414eb9fee7fbb336a06

See more details on using hashes here.

File details

Details for the file paradigm_absorb-0.2.0-py3-none-any.whl.

File metadata

File hashes

Hashes for paradigm_absorb-0.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 48a7983fa371c7d31d8fc7c6d99587c965cbd8b9168899b542752b66b7a8fae6
MD5 563e63e585f444c13c8cd3cf0f73cb4a
BLAKE2b-256 b40fb459b5422d0d946038f3a501b609d7ad4ca76143376087e8a3db69f85553

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page