Skip to main content

A python package to offload a function call to an http server automatically using a decorator.

Project description

auto-function-serving

A python package to offload a function call to an http server running on localhost automatically using a decorator. Compatible with multiprocessing, pickle, flask, fastapi, async etc..

Why

Imagine a case of a multi threaded or multiprocessing application where 1 or few functions are heavily resource (cpu or memory) intensive, but the other functions can run in parallel.
Example - an api call followed by tokenization and classification using a large DL model followed by further API calls.
In such a case, it would make sense to create a server (generally using torchserve or tfserving) to serve requests, and replace the function call with a post request to the server.
ServerHandler creates a synchronous server and replaces any calls to the function automatically during runtime.
Requests are made to 1 instance of a process running a http.server.HTTPServer which runs the function within it.
AsyncServerHandler is also available which makes the requests asynchronously.
Even calls made from different processes, threads, multiprocessing, flask, FastApi and async event loops are made to the same server process.

Usage

In general :

some code with a callable

can be replaced with an instance of Either ServerHandler or AsyncserverHandler that accepts the code as a string in it's first argument and the name of the callable as the second argument.

from auto_function_serving.ServerHandler import ServerHandler
callable_name = ServerHandler("""
some independent code with a callable
""", "callable_name")

Example :

import module1
import module2
def functionname(someinput):
    a = module1.function1(someinput)
    return module2.function2(a)

can be replaced with

from auto_function_serving.ServerHandler import AsyncserverHandler
functionname = AsyncServerHandler("""
import module1
import module2
def functionname(someinput):
    a = module1.function1(someinput)
    return module2.function2(a)
""", "functionname", port="Any")

Decorators (@AsyncserverHandler.decorator and @ServerHandler.decorator) and AsyncServerHandler details in more usage.

Arguments

from auto_function_serving.ServerHandler import ServerHandler
callable_name = ServerHandler("""
some independent code with a callable
""", "callable_name", port=None, backend='Popen', wait=100, backlog = 1024))
  1. port
    • if None, then the input code is hashed and a port is chosen from 50000 to 60000 using the hash
    • if int, then int is chosen
    • otherwise, a random open port is chosen
  2. backend - either 'Popen' or 'multiprocessing'. Popen Should be used in general.
  3. wait - approx max number of seconds to wait for the server to run. No waiting done if set to 0, default 100
  4. backlog - max number of backlogged requests before returning errors, python default is 5, but default in ServerHandler is 1024.

Features

runs http.server.HTTPServer.
ServerHandler and AsyncServerHandler objects can be loaded and unloaded with pickle.
Uses Popen or multiprocessing to run the server.
Uses only a single external dependency (aiohttp), and only for async.
http, not https.
chooses a port based on hash of input. (unless specified otherwise)

Advantages

Minimal code changes.
Should be compatible with almost all functions in almost all CPython envs. (Not sure where it could fail? Please add an issue if you find one.)
Memory leaks or errors (from the server) are extremely unlikely since it is minimal, single threaded, single process and a default component of python stdlib.
Exceptions cause 5xx errors without closing the server.
Even Separate Processes will make requests to 1 instance of the same server unless specified otherwise. (Because it's looking for a server on a specific port.).
Can specify otherwise by set the port to any free port so that a new ServerHandler object starts a new server.
http post requests : lightweight, few ms overhead, reliable.
Async is a good feature.
now with tests.

Disadvatages

Having a string of code as an argument to a class is not pythonic, unless the decorator is used.
Importing inside functions is not ideal, even when the decorator is used.
http post requests : insecure, few ms overhead.
Exceptions inside the server are not sent back.
No batching.
No inbuilt logging. (Could be added). Initialization delay of upto few seconds to start the server. Async functions will not work on the server.

Possible Edge cases

No auto server restart in case server closes.
May leave some resources locked for a while (<1min) if not closed properly.
Problems might occur if Popen or multiprocessing are not available.
Possible nested async errors with jupyter or other? Please look into nest-asyncio and the issues.
Warnings from somewhat hacky (but legit and completely functional) workarounds.
Closing of server process in del and atexit.redister(del) fail for some reason (tested and unlikely).

Installation

Use the package manager pip to install auto_function_serving

pip install auto_function_serving

How does this work?

Code for the server is stored in ServerHandler.base_code and some string formatting is used to fill in the blanks.
The server process is started with Popen (or multiprocessing if specified). The first thing it does is import socket and bind the port - if it's not available the code stops after an exception. Therefore only 1 instance of the server runs at a time on a machine.
We know the function is ready after we can receive a valid get request from the server.
Inputs and outputs are sent as bytes, converted to and from objects using pickle.
If port is None in while initializing (default), a port from 50000 to 60000 is chosen by hashing the input code to make it independent of the source of a function. Collisions of different functions are possible, but unlikely. The collision of the same function in multiple processes is used to make sure only 1 server process runs at a time. The port can be specified if needed.

Performance (On my machine)

overhead for small input and output (few bytes) -
~2ms for requests with urllib.request
~4ms for async requests with aiohttp.ClientSession
overhead for large input and output
~10ms for 0.5 mb input and output (1mb total transfer).
~60ms for 5 mb input and output (10 mb total transfer).
~600ms for 50 mb input and output (100 mb total transfer).

More Usage

It can also be used with the provided decorator for functions with no dependencies outside the function.

from auto_function_serving.ServerHandler import ServerHandler
@ServerHandler.decorator
def someheavyfunction(args,**kwargs):
    for i in range(big_number)
        someexpensivecomputation

imports inside the function will work

from auto_function_serving.ServerHandler import ServerHandler
@ServerHandler.decorator
def someheavyfunction(args,**kwargs):
    import numpy as np
from auto_function_serving.ServerHandler import ServerHandler
@ServerHandler.decorator
def someheavyfunction(args,**kwargs):
    if not hasattr(someheavyfunction,'RunOnce'):
	    global np
        import numpy as np
    setattr(someheavyfunction,'RunOnce',None)
	... etc

When the somemodule does not have any expensive global loading.

from auto_function_serving.ServerHandler import ServerHandler
from somemodule import someheavyfunction
someheavyfunction = ServerHandler.decorator(someheavyfunction)

Ip address can be changed by setting ServerHandler.ip_address (default "127.0.0.1") before creating a new instance.

AsyncServerHandler

AsyncServerHandler is also available which uses aiohttp to make the requests asynchronously, for use with fastapi and other async use cases.
AsyncServerHandler has the same usage as ServerHandler, except calls need to be awaited or used with asyncio.run() or with asyncio.get_event_loop().run_until_complete().
Number of async calls can be limited by setting AsyncServerHandler.TCPConnector_limit which controls the TCPconnector limit (default 100). Using Semaphore is also something to consider.

Other things to look into

Libraries : Celery, Tfserving, Torchserve, Flask
Sending globals and locals to exec
ast trees

Contributing

Pull requests are welcome.

License

Apache License 2.0

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

auto_function_serving-0.1.9.tar.gz (11.7 kB view details)

Uploaded Source

Built Distribution

auto_function_serving-0.1.9-py3-none-any.whl (12.4 kB view details)

Uploaded Python 3

File details

Details for the file auto_function_serving-0.1.9.tar.gz.

File metadata

  • Download URL: auto_function_serving-0.1.9.tar.gz
  • Upload date:
  • Size: 11.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.7.10

File hashes

Hashes for auto_function_serving-0.1.9.tar.gz
Algorithm Hash digest
SHA256 813344c9804b9d1c5f26767ea52b04804695df018da38637b8f6c9bfee17a39d
MD5 db2c2cdfe49d742cee03dd4147cb2fa6
BLAKE2b-256 bd166b72d4be7709fbf7fffcbbaec012ffd0cbd8142e7a669ebbed4b155aca0f

See more details on using hashes here.

File details

Details for the file auto_function_serving-0.1.9-py3-none-any.whl.

File metadata

File hashes

Hashes for auto_function_serving-0.1.9-py3-none-any.whl
Algorithm Hash digest
SHA256 00bf0b7012cf8638db2cecedea22a17f4fe03215ce76b19dab1a9c2765926468
MD5 d9855b8f2d3d490cb13b577f16f56b37
BLAKE2b-256 7becacc9e088194a89f3370ef025fe5adfb753afa3baaa357b65f8394cd272b9

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page