Skip to main content

A Python client library to simplify robust mini-batch scoring against an H2O MLOps scoring endpoint.

Project description

H2O MLOps Scoring Client

A Python client library to simplify robust mini-batch scoring against an H2O MLOps scoring endpoint. It can run on your local PC, a stand alone server, Databricks, or a Spark 3 cluster.

Using it is as easy as:

import h2o_mlops_scoring_client

h2o_mlops_scoring_client.score_source_sink(
    mlops_endpoint_url="https://.../model/score",
    id_column="ID",
    source_data="file:///.../input.csv",
    source_format=h2o_mlops_scoring_client.Format.CSV,
    sink_location="file:///.../output/",
    sink_format=h2o_mlops_scoring_client.Format.PARQUET,
    sink_write_mode=h2o_mlops_scoring_client.WriteMode.OVERWRITE
)

Or if you want to work with Pandas or Spark data frames:

scores_df = h2o_mlops_scoring_client.score_data_frame(
    mlops_endpoint_url="https://.../model/score",
    id_column="ID",
    data_frame=input_df,
)

Installation

Requirements

  • Linux or Mac OS (Windows is not supported)
  • Java
  • Python 3.8 or greater

Install from PyPI

pip install h2o_mlops_scoring_client

FAQ

When should I use the MLOps Scoring Client?

Use when batch scoring processing (authenticating and connecting to source or sink, file/data processing or conversions, etc.) can happen external to H2O AI Cloud but you want to stay within the H2O MLOps workflow (projects, scoring, registry, monitoring, etc.).

Where does scoring take place?

As the batch scoring processing occurs, the data is sent to an H2O MLOps deployment for scoring. The scores are then returned for the batch scoring processing to complete.

What Source/Sinks are supported?

The MLOps scoring client can support many source/sinks, including:

  • ADLS Gen 2
  • Databases with a JDBC driver
  • Local file system
  • GBQ
  • S3
  • Snowflake

What file types are supported?

The MLOps scoring client can read and write:

  • CSV
  • Parquet
  • ORC
  • BigQuery tables
  • JDBC queries
  • JDBC tables
  • Snowflake queries
  • Snowflake tables

If there's a file type you would like to see supported, please let us know.

I want model monitoring for batch scoring, can I do that?

Yes. The MLOps Scoring Client uses MLOps scoring endpoints which are automatically monitored.

Is a Spark installation required?

No. If you're running locally and scoring local files or data frames, then no extra Spark install or configuration is needed. If you want to connect to an external source or sink, you'll need to do a small amount of configuration.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

h2o_mlops_scoring_client-0.1.5b1-py3-none-any.whl (11.8 kB view details)

Uploaded Python 3

File details

Details for the file h2o_mlops_scoring_client-0.1.5b1-py3-none-any.whl.

File metadata

File hashes

Hashes for h2o_mlops_scoring_client-0.1.5b1-py3-none-any.whl
Algorithm Hash digest
SHA256 7ebc79004671d1715ca68184678a94e05708f39fefbc7647437529793e96f430
MD5 509653c09b9e9868d289557e5d317897
BLAKE2b-256 27af83807e155c2e485d0348cf465d9c889825b281046db6645a43dea5628a18

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page