Skip to main content

Python Stream processing.

Project description

faust

Python Stream Processing Fork

python versions version codecov slack Code style: black pre-commit license downloads

Installation

pip install faust-streaming

Documentation

Why the fork

We have decided to fork the original Faust project because there is a critical process of releasing new versions which causes uncertainty in the community. Everybody is welcome to contribute to this fork, and you can be added as a maintainer.

We want to:

  • Ensure continues release
  • Code quality
  • Use of latest versions of kafka drivers (for now only aiokafka)
  • Support kafka transactions
  • Update the documentation

and more...

Usage

# Python Streams
# Forever scalable event processing & in-memory durable K/V store;
# as a library w/ asyncio & static typing.
import faust

Faust is a stream processing library, porting the ideas from Kafka Streams to Python.

It is used at Robinhood to build high performance distributed systems and real-time data pipelines that process billions of events every day.

Faust provides both stream processing and event processing, sharing similarity with tools such as Kafka Streams, Apache Spark, Storm, Samza, Flink,

It does not use a DSL, it's just Python! This means you can use all your favorite Python libraries when stream processing: NumPy, PyTorch, Pandas, NLTK, Django, Flask, SQLAlchemy, ++

Faust requires Python 3.6 or later for the new async/await_ syntax, and variable type annotations.

Here's an example processing a stream of incoming orders:

app = faust.App('myapp', broker='kafka://localhost')

# Models describe how messages are serialized:
# {"account_id": "3fae-...", amount": 3}
class Order(faust.Record):
    account_id: str
    amount: int

@app.agent(value_type=Order)
async def order(orders):
    async for order in orders:
        # process infinite stream of orders.
        print(f'Order for {order.account_id}: {order.amount}')

The Agent decorator defines a "stream processor" that essentially consumes from a Kafka topic and does something for every event it receives.

The agent is an async def function, so can also perform other operations asynchronously, such as web requests.

This system can persist state, acting like a database. Tables are named distributed key/value stores you can use as regular Python dictionaries.

Tables are stored locally on each machine using a super fast embedded database written in C++, called RocksDB.

Tables can also store aggregate counts that are optionally "windowed" so you can keep track of "number of clicks from the last day," or "number of clicks in the last hour." for example. Like Kafka Streams, we support tumbling, hopping and sliding windows of time, and old windows can be expired to stop data from filling up.

For reliability, we use a Kafka topic as "write-ahead-log". Whenever a key is changed we publish to the changelog. Standby nodes consume from this changelog to keep an exact replica of the data and enables instant recovery should any of the nodes fail.

To the user a table is just a dictionary, but data is persisted between restarts and replicated across nodes so on failover other nodes can take over automatically.

You can count page views by URL:

# data sent to 'clicks' topic sharded by URL key.
# e.g. key="http://example.com" value="1"
click_topic = app.topic('clicks', key_type=str, value_type=int)

# default value for missing URL will be 0 with `default=int`
counts = app.Table('click_counts', default=int)

@app.agent(click_topic)
async def count_click(clicks):
    async for url, count in clicks.items():
        counts[url] += count

The data sent to the Kafka topic is partitioned, which means the clicks will be sharded by URL in such a way that every count for the same URL will be delivered to the same Faust worker instance.

Faust supports any type of stream data: bytes, Unicode and serialized structures, but also comes with "Models" that use modern Python syntax to describe how keys and values in streams are serialized:

# Order is a json serialized dictionary,
# having these fields:

class Order(faust.Record):
    account_id: str
    product_id: str
    price: float
    quantity: float = 1.0

orders_topic = app.topic('orders', key_type=str, value_type=Order)

@app.agent(orders_topic)
async def process_order(orders):
    async for order in orders:
        # process each order using regular Python
        total_price = order.price * order.quantity
        await send_order_received_email(order.account_id, order)

Faust is statically typed, using the mypy type checker, so you can take advantage of static types when writing applications.

The Faust source code is small, well organized, and serves as a good resource for learning the implementation of Kafka Streams.

Learn more about Faust in the introduction introduction page to read more about Faust, system requirements, installation instructions, community resources, and more.

or go directly to the quickstart tutorial to see Faust in action by programming a streaming application.

then explore the User Guide for in-depth information organized by topic.

Local development

  1. Clone the project
  2. Create a virtualenv: python3.7 -m venv venv && source venv/bin/activate
  3. Install the requirements: ./scripts/install
  4. Run lint: ./scripts/lint
  5. Run tests: ./scripts/tests

Faust key points

Simple

Faust is extremely easy to use. To get started using other stream processing solutions you have complicated hello-world projects, and infrastructure requirements. Faust only requires Kafka, the rest is just Python, so If you know Python you can already use Faust to do stream processing, and it can integrate with just about anything.

Here's one of the easier applications you can make::

import faust

class Greeting(faust.Record):
    from_name: str
    to_name: str

app = faust.App('hello-app', broker='kafka://localhost')
topic = app.topic('hello-topic', value_type=Greeting)

@app.agent(topic)
async def hello(greetings):
    async for greeting in greetings:
        print(f'Hello from {greeting.from_name} to {greeting.to_name}')

@app.timer(interval=1.0)
async def example_sender(app):
    await hello.send(
        value=Greeting(from_name='Faust', to_name='you'),
    )

if __name__ == '__main__':
    app.main()

You're probably a bit intimidated by the async and await keywords, but you don't have to know how asyncio works to use Faust: just mimic the examples, and you'll be fine.

The example application starts two tasks: one is processing a stream, the other is a background thread sending events to that stream. In a real-life application, your system will publish events to Kafka topics that your processors can consume from, and the background thread is only needed to feed data into our example.

Highly Available

Faust is highly available and can survive network problems and server crashes. In the case of node failure, it can automatically recover, and tables have standby nodes that will take over.

Distributed

Start more instances of your application as needed.

Fast

A single-core Faust worker instance can already process tens of thousands of events every second, and we are reasonably confident that throughput will increase once we can support a more optimized Kafka client.

Flexible

Faust is just Python, and a stream is an infinite asynchronous iterator. If you know how to use Python, you already know how to use Faust, and it works with your favorite Python libraries like Django, Flask, SQLAlchemy, NLTK, NumPy, SciPy, TensorFlow, etc.

Bundles

Faust also defines a group of setuptools extensions that can be used to install Faust and the dependencies for a given feature.

You can specify these in your requirements or on the pip command-line by using brackets. Separate multiple bundles using the comma:

pip install "faust-streaming[rocksdb]"

pip install "faust-streaming[rocksdb,uvloop,fast,redis, aerospike]"

The following bundles are available:

Faust with extras

Stores

pip install faust-streaming[rocksdb] or pip install faust-streaming[rocksdict] for using RocksDB for storing Faust table state. Recommended in production.

pip install faust-streaming[aerospike] for using Aerospike for storing Faust table state. Recommended if supported

Aerospike Configuration

Aerospike can be enabled as the state store by specifying store="aerospike://"

By default, all tables backed by Aerospike use use_partitioner=True and generate changelog topic events similar to a state store backed by RocksDB. The following configuration options should be passed in as keys to the options parameter in Table namespace : aerospike namespace

ttl: TTL for all KV's in the table

username: username to connect to the Aerospike cluster

password: password to connect to the Aerospike cluster

hosts : the hosts parameter as specified in the aerospike client

policies: the different policies for read/write/scans policies

client: a dict of host and policies defined above

Caching

faust-streaming[redis] for using Redis as a simple caching backend (Memcached-style).

Codecs

faust-streaming[yaml] for using YAML and the PyYAML library in streams.

Optimization

faust-streaming[fast] for installing all the available C speedup extensions to Faust core.

Sensors

faust-streaming[datadog] for using the Datadog Faust monitor.

faust-streaming[statsd] for using the Statsd Faust monitor.

faust-streaming[prometheus] for using the Prometheus Faust monitor.

Event Loops

faust-streaming[uvloop] for using Faust with uvloop.

faust-streaming[eventlet] for using Faust with eventlet

Debugging

faust-streaming[debug] for using aiomonitor to connect and debug a running Faust worker.

faust-streaming[setproctitle]when the setproctitle module is installed the Faust worker will use it to set a nicer process name in ps/top listings.vAlso installed with the fast and debug bundles.

Downloading and installing from source

Download the latest version of Faust from https://pypi.org/project/faust-streaming/

You can install it by doing:

$ tar xvfz faust-streaming-0.0.0.tar.gz
$ cd faust-streaming-0.0.0
$ python setup.py build
# python setup.py install

The last command must be executed as a privileged user if you are not currently using a virtualenv.

Using the development version

With pip

You can install the latest snapshot of Faust using the following pip command:

pip install https://github.com/faust-streaming/faust/zipball/master#egg=faust

FAQ

Can I use Faust with Django/Flask/etc

Yes! Use eventlet as a bridge to integrate with asyncio.

Using eventlet

This approach works with any blocking Python library that can work with eventlet

Using eventlet requires you to install the faust-aioeventlet module, and you can install this as a bundle along with Faust:

pip install -U faust-streaming[eventlet]

Then to actually use eventlet as the event loop you have to either use the -L <faust --loop> argument to the faust program:

faust -L eventlet -A myproj worker -l info

or add import mode.loop.eventlet at the top of your entry point script:

#!/usr/bin/env python3
import mode.loop.eventlet  # noqa

It's very important this is at the very top of the module, and that it executes before you import libraries.

Can I use Faust with Tornado

Yes! Use the tornado.platform.asyncio bridge

Can I use Faust with Twisted

Yes! Use the asyncio reactor implementation: https://twistedmatrix.com/documents/current/api/twisted.internet.asyncioreactor.html

Will you support Python 2.7 or Python 3.5

No. Faust requires Python 3.7 or later, since it heavily uses features that were introduced in Python 3.6 (async, await, variable type annotations).

I get a maximum number of open files exceeded error by RocksDB when running a Faust app locally. How can I fix this

You may need to increase the limit for the maximum number of open files. On macOS and Linux you can use:

ulimit -n max_open_files to increase the open files limit to max_open_files.

On docker, you can use the --ulimit flag:

docker run --ulimit nofile=50000:100000 <image-tag> where 50000 is the soft limit, and 100000 is the hard limit See the difference.

What kafka versions faust supports

Faust supports kafka with version >= 0.10.

Getting Help

Slack

For discussions about the usage, development, and future of Faust, please join the fauststream Slack.

Resources

Bug tracker

If you have any suggestions, bug reports, or annoyances please report them to our issue tracker at https://github.com/faust-streaming/faust/issues/

License

This software is licensed under the New BSD License. See the LICENSE file in the top distribution directory for the full license text.

Contributing

Development of Faust happens at GitHub

You're highly encouraged to participate in the development of Faust.

Code of Conduct

Everyone interacting in the project's code bases, issue trackers, chat rooms, and mailing lists is expected to follow the Faust Code of Conduct.

As contributors and maintainers of these projects, and in the interest of fostering an open and welcoming community, we pledge to respect all people who contribute through reporting issues, posting feature requests, updating documentation, submitting pull requests or patches, and other activities.

We are committed to making participation in these projects a harassment-free experience for everyone, regardless of level of experience, gender, gender identity and expression, sexual orientation, disability, personal appearance, body size, race, ethnicity, age, religion, or nationality.

Examples of unacceptable behavior by participants include:

  • The use of sexualized language or imagery
  • Personal attacks
  • Trolling or insulting/derogatory comments
  • Public or private harassment
  • Publishing other's private information, such as physical or electronic addresses, without explicit permission
  • Other unethical or unprofessional conduct.

Project maintainers have the right and responsibility to remove, edit, or reject comments, commits, code, wiki edits, issues, and other contributions that are not aligned to this Code of Conduct. By adopting this Code of Conduct, project maintainers commit themselves to fairly and consistently applying these principles to every aspect of managing this project. Project maintainers who do not follow or enforce the Code of Conduct may be permanently removed from the project team.

This code of conduct applies both within project spaces and in public spaces when an individual is representing the project or its community.

Instances of abusive, harassing, or otherwise unacceptable behavior may be reported by opening an issue or contacting one or more of the project maintainers.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

faust-streaming-0.10.5.tar.gz (754.7 kB view details)

Uploaded Source

Built Distributions

faust_streaming-0.10.5-cp311-cp311-musllinux_1_1_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.11 musllinux: musl 1.1+ x86-64

faust_streaming-0.10.5-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.11 manylinux: glibc 2.17+ x86-64 manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.5-cp311-cp311-macosx_10_9_x86_64.whl (483.4 kB view details)

Uploaded CPython 3.11 macOS 10.9+ x86-64

faust_streaming-0.10.5-cp310-cp310-musllinux_1_1_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.10 musllinux: musl 1.1+ x86-64

faust_streaming-0.10.5-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.0 MB view details)

Uploaded CPython 3.10 manylinux: glibc 2.17+ x86-64 manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.5-cp310-cp310-macosx_10_9_x86_64.whl (488.7 kB view details)

Uploaded CPython 3.10 macOS 10.9+ x86-64

faust_streaming-0.10.5-cp39-cp39-musllinux_1_1_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.9 musllinux: musl 1.1+ x86-64

faust_streaming-0.10.5-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.0 MB view details)

Uploaded CPython 3.9 manylinux: glibc 2.17+ x86-64 manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.5-cp39-cp39-macosx_10_9_x86_64.whl (488.2 kB view details)

Uploaded CPython 3.9 macOS 10.9+ x86-64

faust_streaming-0.10.5-cp38-cp38-musllinux_1_1_x86_64.whl (1.2 MB view details)

Uploaded CPython 3.8 musllinux: musl 1.1+ x86-64

faust_streaming-0.10.5-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.8 manylinux: glibc 2.17+ x86-64 manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.5-cp38-cp38-macosx_10_9_x86_64.whl (486.8 kB view details)

Uploaded CPython 3.8 macOS 10.9+ x86-64

faust_streaming-0.10.5-cp37-cp37m-musllinux_1_1_x86_64.whl (1.0 MB view details)

Uploaded CPython 3.7m musllinux: musl 1.1+ x86-64

faust_streaming-0.10.5-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (965.5 kB view details)

Uploaded CPython 3.7m manylinux: glibc 2.17+ x86-64 manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.5-cp37-cp37m-macosx_10_9_x86_64.whl (484.4 kB view details)

Uploaded CPython 3.7m macOS 10.9+ x86-64

File details

Details for the file faust-streaming-0.10.5.tar.gz.

File metadata

  • Download URL: faust-streaming-0.10.5.tar.gz
  • Upload date:
  • Size: 754.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.11.2

File hashes

Hashes for faust-streaming-0.10.5.tar.gz
Algorithm Hash digest
SHA256 3cfc4c9684ddc38eb61dd91be864a3a8952d9bf4d3ccb6e66c71c0bc65b405ad
MD5 72276e128b11418ad71932a2460991aa
BLAKE2b-256 fe9691fa449204d7d19d072250a93e4b873c9077eefc392da3483318f500d630

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.5-cp311-cp311-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.5-cp311-cp311-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 f2bee20fe7c1b25a4309772b30c081c4c61a45fbecf0762d110b978e7b10e9f0
MD5 9b2d7ad52d4a8f852cf93b1126463824
BLAKE2b-256 df65222bc4a68ed34144a3ac2e332ebcbfebf0679e5004e1d671d4e24f2829ea

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.5-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.5-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 e92253d2adca88cade5007cf88e58f15033affa01d7a5d93a68c73ae0969d62e
MD5 7dbfeac86356996d3bc4cbe52562778a
BLAKE2b-256 188d8a0b85f53a34169c864ea7315672aee81c9c6a6a0fae482bbaf980d2a7dc

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.5-cp311-cp311-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.5-cp311-cp311-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 bec6713ed6c0774804303e120e25a7032b16b8c56ecb4fb9befea40cdaeba55e
MD5 66558761dfe50e1c95fa34d322d07caa
BLAKE2b-256 41a0e7b362d3bf055e6e7fc0c840aeeb23c2b1004c76452fb5cb28322025f353

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.5-cp310-cp310-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.5-cp310-cp310-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 d9cf6075249d3ead46eafb1fd3b5149ebbb84011386ccf58c532b13b4726763b
MD5 5cfbb0262e985f6129a380ae4eba62a9
BLAKE2b-256 9a4d53b884a9d6fda428babec6384667ff7d273b6167de9c70f14daaf5913e91

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.5-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.5-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 2d56bc7a1a21ecf2a927a6c4bc8e66945cd7502e8bf145d96b099fadcace3466
MD5 6c84695f32612a485d14eedd440510d5
BLAKE2b-256 c159ce5a18a55e6e8a2039ac7414045a22b2111a69bd23fe2a94f36fd394bdbf

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.5-cp310-cp310-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.5-cp310-cp310-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 0ef2c6ad5cdd3e8d2bfaba62e4a7d3cf527a7bf487e267bb308f2a79a1e3ca85
MD5 fed4391cb1fd94e1ec20267a63aa56d1
BLAKE2b-256 de341736521f189f2e62b8fd17dbdc05f70a1e66c692d641cb3fb4df7b5f3cca

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.5-cp39-cp39-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.5-cp39-cp39-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 d7de95c6aa9db5694575c8217608226eabe204266f0e5d04b7d6603959ca50bb
MD5 1a74e6880fc496b2ba1529fccdc6fc32
BLAKE2b-256 baf8f4343bb0e39c7965b0dfc886ad5edf88dd64acbaae109b888bc8b6b40a67

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.5-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.5-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 05741d884e288d2b8c706577f832384898349faa23d2e3078388efb01b7ba326
MD5 d16c4b6fd53659c0af275223ab3a002e
BLAKE2b-256 672764bb70bde227d9d8e0ec7110ec3d7b14e9f441eda420e073b30e39e07a3d

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.5-cp39-cp39-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.5-cp39-cp39-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 a658390d2a26d9c0bd9dd75e803e33c4ec81b89946ff0cef5e6eb9583522e6f6
MD5 8e50f9e6c43cc06f2396f7bb31a921f0
BLAKE2b-256 14e7bb6281adaba4190ca35c6422f0a17f08e4a2ea62d1b379c458046c4bb4cc

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.5-cp38-cp38-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.5-cp38-cp38-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 9a34df88624608af77e91f2d324adb2f635dae7d62d43b95539de766583ea6fb
MD5 2ca3e7e3e41be8f06e262099b95ce582
BLAKE2b-256 8405501724790e4c9db562b280adbb7f95bf05cfc302d37194564a58c08ff6f3

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.5-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.5-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 bcc6c18620d379a78fd23e63cc5fd9c70c5748e9b90904bb76a733bde35e7ad4
MD5 7a5ebb2321bf4c7b17cf6cef2c5bc26c
BLAKE2b-256 0a5124d6d9a1db0f0fc40412051de912651bc7035bb1ba95edb6d522c0791a6c

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.5-cp38-cp38-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.5-cp38-cp38-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 f8379bc139dcef4d41b3c4cc200e920afe5adbe6ab6e90453d29606fae3b8905
MD5 ae610604b4fd4ce227ad2f6e8e5bddb6
BLAKE2b-256 7e5565fc88f7f9a478c971103be5b82d82c22256dd3da868c2ea3a9a50cc81e2

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.5-cp37-cp37m-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.5-cp37-cp37m-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 8e911a8f9066111c7d36df84ef8221d43c306e70f4eae9fcc4f89c3fe7541b95
MD5 bc1de9d11b6696bb63854d25c00b0f58
BLAKE2b-256 0daa2573b7b666f7f852ba3b476bfd72e2853ade83cf153c0c758bdf77c59d55

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.5-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.5-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 db0024d5d74f797a19fd1644c1a4f07627a95fd7ba8456ad424f81b988f7bc04
MD5 c889b63c858b27ade15224a68c77d19d
BLAKE2b-256 cf08ca0fea49ab8801a4e1f9590d85e3659dc5adee66748223c275f80d39af43

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.5-cp37-cp37m-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.5-cp37-cp37m-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 86ef43bebdbf05d919f5d4a0fccf8a2705a540b492227f69e23ba12e71bbe17a
MD5 7e3741b34716402849d225565bdd0f15
BLAKE2b-256 f29975e1a987ee42883b34d7eb7b0388f56742842244f76ed7a4955e0d584c6e

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page