Skip to main content

Inference service package for IAPARC

Project description

iaparc_inference

PyPI version PyPI - License

The IA Parc inference plugin allows developers to easily integrate their inference pipeline into IA Parc's production module.

Installation

pip install iaparc-inference

Usage

  • If your inference pipeline support batching:

    from iaparc_inference import IAPListener
    
    # Define a callback to query your inference pipeline
    # To load your model only once it is recommended to use a class:
    class MyModel:
        def __init__(self, model_path: str):
            ## Load your model in pytorch, tensorflow or any other backend
        
        def batch_query(batch: list, parameters: Optional) -> list:
            ''' execute your pipeline on a batch input
                Note:   "parameters" is an optional argument.
                        It can be used to handle URL's query parameters
                        It's a list of key(string)/value(string) dictionaries
            '''
    
    if __name__ == '__main__':
        # Initiate your model class
        my_model = MyModel("path/to/my/model")
    
        # Initiate IAParc listener
        listener = IAPListener(my_model.batch_query)
        # Start the listener
        listener.run()
    
  • If your inference pipeline do not support batching:

    from iaparc_inference import IAPListener
    
    # Define a callback to query your inference pipeline
    # To load your model only once it is recommended to use a class:
    class MyModel:
        def __init__(self, model_path: str):
            ## Load your model in pytorch, tensorflow or any other backend
        
        def single_query(one_input, parameters: Optional):
            ''' execute your pipeline on a single input
                Note:   "parameters" is an optional argument.
                        It can be used to handle URL's query parameters
                        It's a key(string)/value(string) dictionary
            '''
    
    if __name__ == '__main__':
        # Initiate your model class
        my_model = MyModel("path/to/my/model")
    
        # Initiate IAParc listener
        listener = IAPListener(my_model.single_query, batch=1)  # Note that batch size is forced to 1 here
        # Start the listener
        listener.run()
    

Features

  • Dynamic batching
  • Autoscalling
  • Support both synchronous and asynchronous queries
  • Data agnostic

License

This project is licensed under the Apache License Version 2.0 - see the Apache LICENSE file for details.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

iaparc_inference-0.5.5.tar.gz (16.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

iaparc_inference-0.5.5-py3-none-any.whl (18.7 kB view details)

Uploaded Python 3

File details

Details for the file iaparc_inference-0.5.5.tar.gz.

File metadata

  • Download URL: iaparc_inference-0.5.5.tar.gz
  • Upload date:
  • Size: 16.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.10.14

File hashes

Hashes for iaparc_inference-0.5.5.tar.gz
Algorithm Hash digest
SHA256 ed31a5b03572238c2c76a7f71c41baf304bd39f8e6b5f6256cf0c1767382556b
MD5 4155ef10e3d6c2460a82d602ceec503c
BLAKE2b-256 6b67ecb1f0309c10a6ede9fbfa20278d17c82d9ed258d9a4446b5b4a7a77f38f

See more details on using hashes here.

File details

Details for the file iaparc_inference-0.5.5-py3-none-any.whl.

File metadata

File hashes

Hashes for iaparc_inference-0.5.5-py3-none-any.whl
Algorithm Hash digest
SHA256 827d7a246a6eed98855be49a25865ef4c6df45bf9a372021e54aaa9656b14f9f
MD5 76778d270a4c1333623119e35fc4f4a4
BLAKE2b-256 9f02014b6e39897a0919ae2007e54133682f395089e58e664cad1c5e505bb5f3

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page