Skip to main content

Python Stream processing.

Project description

faust

Python Stream Processing Fork

python versions version codecov slack Code style: black pre-commit license downloads

Installation

pip install faust-streaming

Documentation

Why the fork

We have decided to fork the original Faust project because there is a critical process of releasing new versions which causes uncertainty in the community. Everybody is welcome to contribute to this fork, and you can be added as a maintainer.

We want to:

  • Ensure continues release
  • Code quality
  • Use of latest versions of kafka drivers (for now only aiokafka)
  • Support kafka transactions
  • Update the documentation

and more...

Usage

# Python Streams
# Forever scalable event processing & in-memory durable K/V store;
# as a library w/ asyncio & static typing.
import faust

Faust is a stream processing library, porting the ideas from Kafka Streams to Python.

It is used at Robinhood to build high performance distributed systems and real-time data pipelines that process billions of events every day.

Faust provides both stream processing and event processing, sharing similarity with tools such as Kafka Streams, Apache Spark, Storm, Samza, Flink,

It does not use a DSL, it's just Python! This means you can use all your favorite Python libraries when stream processing: NumPy, PyTorch, Pandas, NLTK, Django, Flask, SQLAlchemy, ++

Faust requires Python 3.6 or later for the new async/await_ syntax, and variable type annotations.

Here's an example processing a stream of incoming orders:

app = faust.App('myapp', broker='kafka://localhost')

# Models describe how messages are serialized:
# {"account_id": "3fae-...", amount": 3}
class Order(faust.Record):
    account_id: str
    amount: int

@app.agent(value_type=Order)
async def order(orders):
    async for order in orders:
        # process infinite stream of orders.
        print(f'Order for {order.account_id}: {order.amount}')

The Agent decorator defines a "stream processor" that essentially consumes from a Kafka topic and does something for every event it receives.

The agent is an async def function, so can also perform other operations asynchronously, such as web requests.

This system can persist state, acting like a database. Tables are named distributed key/value stores you can use as regular Python dictionaries.

Tables are stored locally on each machine using a super fast embedded database written in C++, called RocksDB.

Tables can also store aggregate counts that are optionally "windowed" so you can keep track of "number of clicks from the last day," or "number of clicks in the last hour." for example. Like Kafka Streams, we support tumbling, hopping and sliding windows of time, and old windows can be expired to stop data from filling up.

For reliability, we use a Kafka topic as "write-ahead-log". Whenever a key is changed we publish to the changelog. Standby nodes consume from this changelog to keep an exact replica of the data and enables instant recovery should any of the nodes fail.

To the user a table is just a dictionary, but data is persisted between restarts and replicated across nodes so on failover other nodes can take over automatically.

You can count page views by URL:

# data sent to 'clicks' topic sharded by URL key.
# e.g. key="http://example.com" value="1"
click_topic = app.topic('clicks', key_type=str, value_type=int)

# default value for missing URL will be 0 with `default=int`
counts = app.Table('click_counts', default=int)

@app.agent(click_topic)
async def count_click(clicks):
    async for url, count in clicks.items():
        counts[url] += count

The data sent to the Kafka topic is partitioned, which means the clicks will be sharded by URL in such a way that every count for the same URL will be delivered to the same Faust worker instance.

Faust supports any type of stream data: bytes, Unicode and serialized structures, but also comes with "Models" that use modern Python syntax to describe how keys and values in streams are serialized:

# Order is a json serialized dictionary,
# having these fields:

class Order(faust.Record):
    account_id: str
    product_id: str
    price: float
    quantity: float = 1.0

orders_topic = app.topic('orders', key_type=str, value_type=Order)

@app.agent(orders_topic)
async def process_order(orders):
    async for order in orders:
        # process each order using regular Python
        total_price = order.price * order.quantity
        await send_order_received_email(order.account_id, order)

Faust is statically typed, using the mypy type checker, so you can take advantage of static types when writing applications.

The Faust source code is small, well organized, and serves as a good resource for learning the implementation of Kafka Streams.

Learn more about Faust in the introduction introduction page to read more about Faust, system requirements, installation instructions, community resources, and more.

or go directly to the quickstart tutorial to see Faust in action by programming a streaming application.

then explore the User Guide for in-depth information organized by topic.

Local development

  1. Clone the project
  2. Create a virtualenv: python3.7 -m venv venv && source venv/bin/activate
  3. Install the requirements: ./scripts/install
  4. Run lint: ./scripts/lint
  5. Run tests: ./scripts/tests

Faust key points

Simple

Faust is extremely easy to use. To get started using other stream processing solutions you have complicated hello-world projects, and infrastructure requirements. Faust only requires Kafka, the rest is just Python, so If you know Python you can already use Faust to do stream processing, and it can integrate with just about anything.

Here's one of the easier applications you can make::

import faust

class Greeting(faust.Record):
    from_name: str
    to_name: str

app = faust.App('hello-app', broker='kafka://localhost')
topic = app.topic('hello-topic', value_type=Greeting)

@app.agent(topic)
async def hello(greetings):
    async for greeting in greetings:
        print(f'Hello from {greeting.from_name} to {greeting.to_name}')

@app.timer(interval=1.0)
async def example_sender(app):
    await hello.send(
        value=Greeting(from_name='Faust', to_name='you'),
    )

if __name__ == '__main__':
    app.main()

You're probably a bit intimidated by the async and await keywords, but you don't have to know how asyncio works to use Faust: just mimic the examples, and you'll be fine.

The example application starts two tasks: one is processing a stream, the other is a background thread sending events to that stream. In a real-life application, your system will publish events to Kafka topics that your processors can consume from, and the background thread is only needed to feed data into our example.

Highly Available

Faust is highly available and can survive network problems and server crashes. In the case of node failure, it can automatically recover, and tables have standby nodes that will take over.

Distributed

Start more instances of your application as needed.

Fast

A single-core Faust worker instance can already process tens of thousands of events every second, and we are reasonably confident that throughput will increase once we can support a more optimized Kafka client.

Flexible

Faust is just Python, and a stream is an infinite asynchronous iterator. If you know how to use Python, you already know how to use Faust, and it works with your favorite Python libraries like Django, Flask, SQLAlchemy, NLTK, NumPy, SciPy, TensorFlow, etc.

Bundles

Faust also defines a group of setuptools extensions that can be used to install Faust and the dependencies for a given feature.

You can specify these in your requirements or on the pip command-line by using brackets. Separate multiple bundles using the comma:

pip install "faust-streaming[rocksdb]"

pip install "faust-streaming[rocksdb,uvloop,fast,redis,aerospike]"

The following bundles are available:

Faust with extras

Stores

RocksDB

For using RocksDB for storing Faust table state. Recommended in production.

pip install faust-streaming[rocksdb] (uses RocksDB 6)

pip install faust-streaming[rocksdict] (uses RocksDB 8, not backwards compatible with 6)

Aerospike

pip install faust-streaming[aerospike] for using Aerospike for storing Faust table state. Recommended if supported

Aerospike Configuration

Aerospike can be enabled as the state store by specifying store="aerospike://"

By default, all tables backed by Aerospike use use_partitioner=True and generate changelog topic events similar to a state store backed by RocksDB. The following configuration options should be passed in as keys to the options parameter in Table namespace : aerospike namespace

ttl: TTL for all KV's in the table

username: username to connect to the Aerospike cluster

password: password to connect to the Aerospike cluster

hosts : the hosts parameter as specified in the aerospike client

policies: the different policies for read/write/scans policies

client: a dict of host and policies defined above

Caching

faust-streaming[redis] for using Redis as a simple caching backend (Memcached-style).

Codecs

faust-streaming[yaml] for using YAML and the PyYAML library in streams.

Optimization

faust-streaming[fast] for installing all the available C speedup extensions to Faust core.

Sensors

faust-streaming[datadog] for using the Datadog Faust monitor.

faust-streaming[statsd] for using the Statsd Faust monitor.

faust-streaming[prometheus] for using the Prometheus Faust monitor.

Event Loops

faust-streaming[uvloop] for using Faust with uvloop.

faust-streaming[eventlet] for using Faust with eventlet

Debugging

faust-streaming[debug] for using aiomonitor to connect and debug a running Faust worker.

faust-streaming[setproctitle]when the setproctitle module is installed the Faust worker will use it to set a nicer process name in ps/top listings.vAlso installed with the fast and debug bundles.

Downloading and installing from source

Download the latest version of Faust from https://pypi.org/project/faust-streaming/

You can install it by doing:

$ tar xvfz faust-streaming-0.0.0.tar.gz
$ cd faust-streaming-0.0.0
$ python setup.py build
# python setup.py install

The last command must be executed as a privileged user if you are not currently using a virtualenv.

Using the development version

With pip

You can install the latest snapshot of Faust using the following pip command:

pip install https://github.com/faust-streaming/faust/zipball/master#egg=faust

FAQ

Can I use Faust with Django/Flask/etc

Yes! Use eventlet as a bridge to integrate with asyncio.

Using eventlet

This approach works with any blocking Python library that can work with eventlet

Using eventlet requires you to install the faust-aioeventlet module, and you can install this as a bundle along with Faust:

pip install -U faust-streaming[eventlet]

Then to actually use eventlet as the event loop you have to either use the -L <faust --loop> argument to the faust program:

faust -L eventlet -A myproj worker -l info

or add import mode.loop.eventlet at the top of your entry point script:

#!/usr/bin/env python3
import mode.loop.eventlet  # noqa

It's very important this is at the very top of the module, and that it executes before you import libraries.

Can I use Faust with Tornado

Yes! Use the tornado.platform.asyncio bridge

Can I use Faust with Twisted

Yes! Use the asyncio reactor implementation: https://twistedmatrix.com/documents/current/api/twisted.internet.asyncioreactor.html

Will you support Python 2.7 or Python 3.5

No. Faust requires Python 3.7 or later, since it heavily uses features that were introduced in Python 3.6 (async, await, variable type annotations).

I get a maximum number of open files exceeded error by RocksDB when running a Faust app locally. How can I fix this

You may need to increase the limit for the maximum number of open files. On macOS and Linux you can use:

ulimit -n max_open_files to increase the open files limit to max_open_files.

On docker, you can use the --ulimit flag:

docker run --ulimit nofile=50000:100000 <image-tag> where 50000 is the soft limit, and 100000 is the hard limit See the difference.

What kafka versions faust supports

Faust supports kafka with version >= 0.10.

Getting Help

Slack

For discussions about the usage, development, and future of Faust, please join the fauststream Slack.

Resources

Bug tracker

If you have any suggestions, bug reports, or annoyances please report them to our issue tracker at https://github.com/faust-streaming/faust/issues/

License

This software is licensed under the New BSD License. See the LICENSE file in the top distribution directory for the full license text.

Contributing

Development of Faust happens at GitHub

You're highly encouraged to participate in the development of Faust.

Code of Conduct

Everyone interacting in the project's code bases, issue trackers, chat rooms, and mailing lists is expected to follow the Faust Code of Conduct.

As contributors and maintainers of these projects, and in the interest of fostering an open and welcoming community, we pledge to respect all people who contribute through reporting issues, posting feature requests, updating documentation, submitting pull requests or patches, and other activities.

We are committed to making participation in these projects a harassment-free experience for everyone, regardless of level of experience, gender, gender identity and expression, sexual orientation, disability, personal appearance, body size, race, ethnicity, age, religion, or nationality.

Examples of unacceptable behavior by participants include:

  • The use of sexualized language or imagery
  • Personal attacks
  • Trolling or insulting/derogatory comments
  • Public or private harassment
  • Publishing other's private information, such as physical or electronic addresses, without explicit permission
  • Other unethical or unprofessional conduct.

Project maintainers have the right and responsibility to remove, edit, or reject comments, commits, code, wiki edits, issues, and other contributions that are not aligned to this Code of Conduct. By adopting this Code of Conduct, project maintainers commit themselves to fairly and consistently applying these principles to every aspect of managing this project. Project maintainers who do not follow or enforce the Code of Conduct may be permanently removed from the project team.

This code of conduct applies both within project spaces and in public spaces when an individual is representing the project or its community.

Instances of abusive, harassing, or otherwise unacceptable behavior may be reported by opening an issue or contacting one or more of the project maintainers.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

faust-streaming-0.10.13.tar.gz (755.7 kB view details)

Uploaded Source

Built Distributions

faust_streaming-0.10.13-cp311-cp311-musllinux_1_1_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.11 musllinux: musl 1.1+ x86-64

faust_streaming-0.10.13-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.11 manylinux: glibc 2.17+ x86-64 manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.13-cp311-cp311-macosx_10_9_x86_64.whl (483.6 kB view details)

Uploaded CPython 3.11 macOS 10.9+ x86-64

faust_streaming-0.10.13-cp310-cp310-musllinux_1_1_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.10 musllinux: musl 1.1+ x86-64

faust_streaming-0.10.13-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.0 MB view details)

Uploaded CPython 3.10 manylinux: glibc 2.17+ x86-64 manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.13-cp310-cp310-macosx_10_9_x86_64.whl (488.9 kB view details)

Uploaded CPython 3.10 macOS 10.9+ x86-64

faust_streaming-0.10.13-cp39-cp39-musllinux_1_1_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.9 musllinux: musl 1.1+ x86-64

faust_streaming-0.10.13-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.0 MB view details)

Uploaded CPython 3.9 manylinux: glibc 2.17+ x86-64 manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.13-cp39-cp39-macosx_10_9_x86_64.whl (488.4 kB view details)

Uploaded CPython 3.9 macOS 10.9+ x86-64

faust_streaming-0.10.13-cp38-cp38-musllinux_1_1_x86_64.whl (1.2 MB view details)

Uploaded CPython 3.8 musllinux: musl 1.1+ x86-64

faust_streaming-0.10.13-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.8 manylinux: glibc 2.17+ x86-64 manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.13-cp38-cp38-macosx_10_9_x86_64.whl (487.1 kB view details)

Uploaded CPython 3.8 macOS 10.9+ x86-64

faust_streaming-0.10.13-cp37-cp37m-musllinux_1_1_x86_64.whl (1.0 MB view details)

Uploaded CPython 3.7m musllinux: musl 1.1+ x86-64

faust_streaming-0.10.13-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (965.3 kB view details)

Uploaded CPython 3.7m manylinux: glibc 2.17+ x86-64 manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.13-cp37-cp37m-macosx_10_9_x86_64.whl (484.7 kB view details)

Uploaded CPython 3.7m macOS 10.9+ x86-64

File details

Details for the file faust-streaming-0.10.13.tar.gz.

File metadata

  • Download URL: faust-streaming-0.10.13.tar.gz
  • Upload date:
  • Size: 755.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.11.3

File hashes

Hashes for faust-streaming-0.10.13.tar.gz
Algorithm Hash digest
SHA256 cb36a823377598e25bc91615c1af6c18bbd0744f35e6007d584b90daee1c6afd
MD5 8dc585ebffb44aabf1542c8baab0fd9f
BLAKE2b-256 620f098d40dbb2f97ac5cdbd7ac524bbd123536786bf601bf7f6e94742f45231

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.13-cp311-cp311-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.13-cp311-cp311-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 9058fdd2eb84dd98a38174e1f37e2a77b7ff496a1d44f10c39bc919e9b365cd5
MD5 6efcaf688edfddeaa8ad5ae1a1d1baf0
BLAKE2b-256 bd7cdd74ea39a147f870ee917944498952323b357e4950ca3a9a0b3fd0037cfa

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.13-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.13-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 73e9ae26533bbd39ac51be9b0e06c4c10ba40d0b6786633a3708d917e3caa91b
MD5 11be99d3ef47e9399c76f9611c665fb0
BLAKE2b-256 10e4e4b49c90e948de077c696acbe023d1a8878ccd84d68f159e36f10849f4a9

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.13-cp311-cp311-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.13-cp311-cp311-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 5d586c6fdb3fb4a42290b89027131e0db0c975a30b8845b867befc501fab37a0
MD5 06cf9e609e664edc6587907e13e0dfd9
BLAKE2b-256 c4dc9f8171a7e5434873b08e9ceb932d786ec63bf2039763b5ef945ae9c56c0d

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.13-cp310-cp310-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.13-cp310-cp310-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 135ed1a40cb90e8600970cd59a4c4c0544277602721bef327c94c53fbc1a4c58
MD5 ec7ae0c99197d23799e17c2ddfb94de9
BLAKE2b-256 2e5f3add94f4253f7e59aa61177be2e7998626f31d2744493840e63e2c0b4cc3

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.13-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.13-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 942fdd8756bcc89aca3c68345bcfbedb9a6f211c1127301b1369132dc8c1cee8
MD5 7776b6f82f49dd77445ee57e16307ca2
BLAKE2b-256 a469c8a0d16f5469aeeff19b0e198cfff84e55263d850a49f6b0927eff9a1709

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.13-cp310-cp310-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.13-cp310-cp310-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 cfda185974ed7f04731e23008b2f6d7ffc8674d7d04c446bb5bd6e9f4ef8673a
MD5 a3fab5f52e187c8b1892d7285f9f5a03
BLAKE2b-256 95a96f0845652dea4da47fd5485328f63bcef40980dd6195ccace552fef9723e

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.13-cp39-cp39-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.13-cp39-cp39-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 436c22a707f48fc6cd02e67b03d86b5ca3684ff418f6c7e1c20440270d83ab26
MD5 51b3a013971b0eedeae620146432e283
BLAKE2b-256 65b9b5979d1aec9d14e1a994201eaf2236fc00f640bb9cda16e16362db45a7a4

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.13-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.13-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 32fbe4bce38facaa036b1eee979424410e2d9ff3c78aa0cfa3a92b6dcdb4b1cb
MD5 a8486dbcfdf9a8a4ef939bea0014aa4f
BLAKE2b-256 38890a2a2a49726afdf41af28f4b1f03396be916ae75385b73f5401fdb9f7b99

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.13-cp39-cp39-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.13-cp39-cp39-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 f982e0d69940584ae2db15f36602b2c1add4afae2ceeef5803fa4555a0840366
MD5 769f61143ba87feec18f42ded23c0b19
BLAKE2b-256 198527bd2743907d84520318605b842c4d212e68555b1ce85e5b6240cbe25ab4

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.13-cp38-cp38-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.13-cp38-cp38-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 8f11e2489c33662d7aeff245c0c84aa1c080de2cf1a31d2e9bfcb4725c87e029
MD5 6b7eea3df20ff2e2a4d23157e5ee3de8
BLAKE2b-256 975bbc780a016803c0fa4121dd77fb5fdb3751c114473e3d74f798c6187e427f

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.13-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.13-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 78a591d31b8107fd37edaa30e6b5b36c04d66d909fffb26abf822839bc5c4bc6
MD5 6023ee76c041d768468642c2d643fb5e
BLAKE2b-256 27a66049a3aaf8f27bd33e32aa90ab942674777fd7d9724bada7a51bb00c88d8

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.13-cp38-cp38-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.13-cp38-cp38-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 e65f7cc7028aa0d69952428d7dcbbda9bdf6b5a31a185510883b5eb417f119f2
MD5 c441cb7d0b313d17284a238b97cf2a25
BLAKE2b-256 a860b1fbe62e6fd5ab1de06d60f7dd4a64a3d4cd3218f959633334f3ebf005e9

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.13-cp37-cp37m-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.13-cp37-cp37m-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 a2df35ffe84ae70b0081be6e6051fa2edbfbd210c1460f0f793861190e7fc9a4
MD5 78452174fdc6dc3c5cecb11b91d332d9
BLAKE2b-256 124d358ec46c8d594721bc6de9fa9e38da2c8f44b2d76739d0ff2b9992291a6d

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.13-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.13-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 ffbb340b6643d12080984de1ceef867fd90307df5c69fa1e0c0d5daeecc3052c
MD5 ebac46f7b99eda184349e3c6c12f6af0
BLAKE2b-256 ff0aa843dd5ee987a22056bd382d48c9078c75a1de07f8e8190239f42f81feb7

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.13-cp37-cp37m-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.13-cp37-cp37m-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 7e2a3e9082e8771d72a35f78c94dcd59d632bbd6320ce573c8e44f0f5eadd9db
MD5 3bc7393ede55e7797e3eb841ebc21128
BLAKE2b-256 6c8a44520aebf3abf92faaa37840f928543f6b47522ba385c9ea30036a4cfda2

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page