Skip to main content

Python Stream processing.

Project description

faust

Python Stream Processing Fork

python versions version codecov slack Code style: black pre-commit license downloads

Installation

pip install faust-streaming

Documentation

Why the fork

We have decided to fork the original Faust project because there is a critical process of releasing new versions which causes uncertainty in the community. Everybody is welcome to contribute to this fork, and you can be added as a maintainer.

We want to:

  • Ensure continues release
  • Code quality
  • Use of latest versions of kafka drivers (for now only aiokafka)
  • Support kafka transactions
  • Update the documentation

and more...

Usage

# Python Streams
# Forever scalable event processing & in-memory durable K/V store;
# as a library w/ asyncio & static typing.
import faust

Faust is a stream processing library, porting the ideas from Kafka Streams to Python.

It is used at Robinhood to build high performance distributed systems and real-time data pipelines that process billions of events every day.

Faust provides both stream processing and event processing, sharing similarity with tools such as Kafka Streams, Apache Spark, Storm, Samza, Flink,

It does not use a DSL, it's just Python! This means you can use all your favorite Python libraries when stream processing: NumPy, PyTorch, Pandas, NLTK, Django, Flask, SQLAlchemy, ++

Faust requires Python 3.6 or later for the new async/await_ syntax, and variable type annotations.

Here's an example processing a stream of incoming orders:

app = faust.App('myapp', broker='kafka://localhost')

# Models describe how messages are serialized:
# {"account_id": "3fae-...", amount": 3}
class Order(faust.Record):
    account_id: str
    amount: int

@app.agent(value_type=Order)
async def order(orders):
    async for order in orders:
        # process infinite stream of orders.
        print(f'Order for {order.account_id}: {order.amount}')

The Agent decorator defines a "stream processor" that essentially consumes from a Kafka topic and does something for every event it receives.

The agent is an async def function, so can also perform other operations asynchronously, such as web requests.

This system can persist state, acting like a database. Tables are named distributed key/value stores you can use as regular Python dictionaries.

Tables are stored locally on each machine using a super fast embedded database written in C++, called RocksDB.

Tables can also store aggregate counts that are optionally "windowed" so you can keep track of "number of clicks from the last day," or "number of clicks in the last hour." for example. Like Kafka Streams, we support tumbling, hopping and sliding windows of time, and old windows can be expired to stop data from filling up.

For reliability, we use a Kafka topic as "write-ahead-log". Whenever a key is changed we publish to the changelog. Standby nodes consume from this changelog to keep an exact replica of the data and enables instant recovery should any of the nodes fail.

To the user a table is just a dictionary, but data is persisted between restarts and replicated across nodes so on failover other nodes can take over automatically.

You can count page views by URL:

# data sent to 'clicks' topic sharded by URL key.
# e.g. key="http://example.com" value="1"
click_topic = app.topic('clicks', key_type=str, value_type=int)

# default value for missing URL will be 0 with `default=int`
counts = app.Table('click_counts', default=int)

@app.agent(click_topic)
async def count_click(clicks):
    async for url, count in clicks.items():
        counts[url] += count

The data sent to the Kafka topic is partitioned, which means the clicks will be sharded by URL in such a way that every count for the same URL will be delivered to the same Faust worker instance.

Faust supports any type of stream data: bytes, Unicode and serialized structures, but also comes with "Models" that use modern Python syntax to describe how keys and values in streams are serialized:

# Order is a json serialized dictionary,
# having these fields:

class Order(faust.Record):
    account_id: str
    product_id: str
    price: float
    quantity: float = 1.0

orders_topic = app.topic('orders', key_type=str, value_type=Order)

@app.agent(orders_topic)
async def process_order(orders):
    async for order in orders:
        # process each order using regular Python
        total_price = order.price * order.quantity
        await send_order_received_email(order.account_id, order)

Faust is statically typed, using the mypy type checker, so you can take advantage of static types when writing applications.

The Faust source code is small, well organized, and serves as a good resource for learning the implementation of Kafka Streams.

Learn more about Faust in the introduction introduction page to read more about Faust, system requirements, installation instructions, community resources, and more.

or go directly to the quickstart tutorial to see Faust in action by programming a streaming application.

then explore the User Guide for in-depth information organized by topic.

Local development

  1. Clone the project
  2. Create a virtualenv: python3.7 -m venv venv && source venv/bin/activate
  3. Install the requirements: ./scripts/install
  4. Run lint: ./scripts/lint
  5. Run tests: ./scripts/tests

Faust key points

Simple

Faust is extremely easy to use. To get started using other stream processing solutions you have complicated hello-world projects, and infrastructure requirements. Faust only requires Kafka, the rest is just Python, so If you know Python you can already use Faust to do stream processing, and it can integrate with just about anything.

Here's one of the easier applications you can make::

import faust

class Greeting(faust.Record):
    from_name: str
    to_name: str

app = faust.App('hello-app', broker='kafka://localhost')
topic = app.topic('hello-topic', value_type=Greeting)

@app.agent(topic)
async def hello(greetings):
    async for greeting in greetings:
        print(f'Hello from {greeting.from_name} to {greeting.to_name}')

@app.timer(interval=1.0)
async def example_sender(app):
    await hello.send(
        value=Greeting(from_name='Faust', to_name='you'),
    )

if __name__ == '__main__':
    app.main()

You're probably a bit intimidated by the async and await keywords, but you don't have to know how asyncio works to use Faust: just mimic the examples, and you'll be fine.

The example application starts two tasks: one is processing a stream, the other is a background thread sending events to that stream. In a real-life application, your system will publish events to Kafka topics that your processors can consume from, and the background thread is only needed to feed data into our example.

Highly Available

Faust is highly available and can survive network problems and server crashes. In the case of node failure, it can automatically recover, and tables have standby nodes that will take over.

Distributed

Start more instances of your application as needed.

Fast

A single-core Faust worker instance can already process tens of thousands of events every second, and we are reasonably confident that throughput will increase once we can support a more optimized Kafka client.

Flexible

Faust is just Python, and a stream is an infinite asynchronous iterator. If you know how to use Python, you already know how to use Faust, and it works with your favorite Python libraries like Django, Flask, SQLAlchemy, NLTK, NumPy, SciPy, TensorFlow, etc.

Bundles

Faust also defines a group of setuptools extensions that can be used to install Faust and the dependencies for a given feature.

You can specify these in your requirements or on the pip command-line by using brackets. Separate multiple bundles using the comma:

pip install "faust-streaming[rocksdb]"

pip install "faust-streaming[rocksdb,uvloop,fast,redis, aerospike]"

The following bundles are available:

Faust with extras

Stores

pip install faust-streaming[rocksdb] or pip install faust-streaming[rocksdict] for using RocksDB for storing Faust table state. Recommended in production.

pip install faust-streaming[aerospike] for using Aerospike for storing Faust table state. Recommended if supported

Aerospike Configuration

Aerospike can be enabled as the state store by specifying store="aerospike://"

By default, all tables backed by Aerospike use use_partitioner=True and generate changelog topic events similar to a state store backed by RocksDB. The following configuration options should be passed in as keys to the options parameter in Table namespace : aerospike namespace

ttl: TTL for all KV's in the table

username: username to connect to the Aerospike cluster

password: password to connect to the Aerospike cluster

hosts : the hosts parameter as specified in the aerospike client

policies: the different policies for read/write/scans policies

client: a dict of host and policies defined above

Caching

faust-streaming[redis] for using Redis as a simple caching backend (Memcached-style).

Codecs

faust-streaming[yaml] for using YAML and the PyYAML library in streams.

Optimization

faust-streaming[fast] for installing all the available C speedup extensions to Faust core.

Sensors

faust-streaming[datadog] for using the Datadog Faust monitor.

faust-streaming[statsd] for using the Statsd Faust monitor.

faust-streaming[prometheus] for using the Prometheus Faust monitor.

Event Loops

faust-streaming[uvloop] for using Faust with uvloop.

faust-streaming[eventlet] for using Faust with eventlet

Debugging

faust-streaming[debug] for using aiomonitor to connect and debug a running Faust worker.

faust-streaming[setproctitle]when the setproctitle module is installed the Faust worker will use it to set a nicer process name in ps/top listings.vAlso installed with the fast and debug bundles.

Downloading and installing from source

Download the latest version of Faust from https://pypi.org/project/faust-streaming/

You can install it by doing:

$ tar xvfz faust-streaming-0.0.0.tar.gz
$ cd faust-streaming-0.0.0
$ python setup.py build
# python setup.py install

The last command must be executed as a privileged user if you are not currently using a virtualenv.

Using the development version

With pip

You can install the latest snapshot of Faust using the following pip command:

pip install https://github.com/faust-streaming/faust/zipball/master#egg=faust

FAQ

Can I use Faust with Django/Flask/etc

Yes! Use eventlet as a bridge to integrate with asyncio.

Using eventlet

This approach works with any blocking Python library that can work with eventlet

Using eventlet requires you to install the faust-aioeventlet module, and you can install this as a bundle along with Faust:

pip install -U faust-streaming[eventlet]

Then to actually use eventlet as the event loop you have to either use the -L <faust --loop> argument to the faust program:

faust -L eventlet -A myproj worker -l info

or add import mode.loop.eventlet at the top of your entry point script:

#!/usr/bin/env python3
import mode.loop.eventlet  # noqa

It's very important this is at the very top of the module, and that it executes before you import libraries.

Can I use Faust with Tornado

Yes! Use the tornado.platform.asyncio bridge

Can I use Faust with Twisted

Yes! Use the asyncio reactor implementation: https://twistedmatrix.com/documents/current/api/twisted.internet.asyncioreactor.html

Will you support Python 2.7 or Python 3.5

No. Faust requires Python 3.7 or later, since it heavily uses features that were introduced in Python 3.6 (async, await, variable type annotations).

I get a maximum number of open files exceeded error by RocksDB when running a Faust app locally. How can I fix this

You may need to increase the limit for the maximum number of open files. On macOS and Linux you can use:

ulimit -n max_open_files to increase the open files limit to max_open_files.

On docker, you can use the --ulimit flag:

docker run --ulimit nofile=50000:100000 <image-tag> where 50000 is the soft limit, and 100000 is the hard limit See the difference.

What kafka versions faust supports

Faust supports kafka with version >= 0.10.

Getting Help

Slack

For discussions about the usage, development, and future of Faust, please join the fauststream Slack.

Resources

Bug tracker

If you have any suggestions, bug reports, or annoyances please report them to our issue tracker at https://github.com/faust-streaming/faust/issues/

License

This software is licensed under the New BSD License. See the LICENSE file in the top distribution directory for the full license text.

Contributing

Development of Faust happens at GitHub

You're highly encouraged to participate in the development of Faust.

Code of Conduct

Everyone interacting in the project's code bases, issue trackers, chat rooms, and mailing lists is expected to follow the Faust Code of Conduct.

As contributors and maintainers of these projects, and in the interest of fostering an open and welcoming community, we pledge to respect all people who contribute through reporting issues, posting feature requests, updating documentation, submitting pull requests or patches, and other activities.

We are committed to making participation in these projects a harassment-free experience for everyone, regardless of level of experience, gender, gender identity and expression, sexual orientation, disability, personal appearance, body size, race, ethnicity, age, religion, or nationality.

Examples of unacceptable behavior by participants include:

  • The use of sexualized language or imagery
  • Personal attacks
  • Trolling or insulting/derogatory comments
  • Public or private harassment
  • Publishing other's private information, such as physical or electronic addresses, without explicit permission
  • Other unethical or unprofessional conduct.

Project maintainers have the right and responsibility to remove, edit, or reject comments, commits, code, wiki edits, issues, and other contributions that are not aligned to this Code of Conduct. By adopting this Code of Conduct, project maintainers commit themselves to fairly and consistently applying these principles to every aspect of managing this project. Project maintainers who do not follow or enforce the Code of Conduct may be permanently removed from the project team.

This code of conduct applies both within project spaces and in public spaces when an individual is representing the project or its community.

Instances of abusive, harassing, or otherwise unacceptable behavior may be reported by opening an issue or contacting one or more of the project maintainers.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

faust-streaming-0.10.7.tar.gz (755.1 kB view details)

Uploaded Source

Built Distributions

faust_streaming-0.10.7-cp311-cp311-musllinux_1_1_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.11 musllinux: musl 1.1+ x86-64

faust_streaming-0.10.7-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.11 manylinux: glibc 2.17+ x86-64 manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.7-cp311-cp311-macosx_10_9_x86_64.whl (483.4 kB view details)

Uploaded CPython 3.11 macOS 10.9+ x86-64

faust_streaming-0.10.7-cp310-cp310-musllinux_1_1_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.10 musllinux: musl 1.1+ x86-64

faust_streaming-0.10.7-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.0 MB view details)

Uploaded CPython 3.10 manylinux: glibc 2.17+ x86-64 manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.7-cp310-cp310-macosx_10_9_x86_64.whl (488.6 kB view details)

Uploaded CPython 3.10 macOS 10.9+ x86-64

faust_streaming-0.10.7-cp39-cp39-musllinux_1_1_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.9 musllinux: musl 1.1+ x86-64

faust_streaming-0.10.7-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.0 MB view details)

Uploaded CPython 3.9 manylinux: glibc 2.17+ x86-64 manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.7-cp39-cp39-macosx_10_9_x86_64.whl (488.2 kB view details)

Uploaded CPython 3.9 macOS 10.9+ x86-64

faust_streaming-0.10.7-cp38-cp38-musllinux_1_1_x86_64.whl (1.2 MB view details)

Uploaded CPython 3.8 musllinux: musl 1.1+ x86-64

faust_streaming-0.10.7-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.8 manylinux: glibc 2.17+ x86-64 manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.7-cp38-cp38-macosx_10_9_x86_64.whl (486.8 kB view details)

Uploaded CPython 3.8 macOS 10.9+ x86-64

faust_streaming-0.10.7-cp37-cp37m-musllinux_1_1_x86_64.whl (1.0 MB view details)

Uploaded CPython 3.7m musllinux: musl 1.1+ x86-64

faust_streaming-0.10.7-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (965.4 kB view details)

Uploaded CPython 3.7m manylinux: glibc 2.17+ x86-64 manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.7-cp37-cp37m-macosx_10_9_x86_64.whl (484.4 kB view details)

Uploaded CPython 3.7m macOS 10.9+ x86-64

File details

Details for the file faust-streaming-0.10.7.tar.gz.

File metadata

  • Download URL: faust-streaming-0.10.7.tar.gz
  • Upload date:
  • Size: 755.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.11.2

File hashes

Hashes for faust-streaming-0.10.7.tar.gz
Algorithm Hash digest
SHA256 6ddfbb7dcaca4faf3a4aae462e807684c5b7ecf92eeb5f0f7fd9e4c27764d08d
MD5 2f92b3277a79b9c24f29e632e181c971
BLAKE2b-256 a7a997bedfb2926345c6bc8431b4d3aa293900848efe9fd3dfdd9bd172f0e091

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.7-cp311-cp311-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.7-cp311-cp311-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 7dfa238c8a863e2f509b69b568dbcdfe8d0c2da879e0039a6823d836d8794128
MD5 66cd2c597f8a2b5c328a6781775d0fdf
BLAKE2b-256 b06dc15209b14ad1403c66176da16644bf9fede23d26bdbcff755976c6752b15

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.7-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.7-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 83e99a5a8d41d48c38e05aa3f32df7b7e2d95931f125c4b79f4a00b58e397867
MD5 82ea4d180281899ebd1ddc6964dd8ef4
BLAKE2b-256 0b813edb56a0d1cdf010a8eeabe01317d382c908c7539ca8ebe7dc7c175644f4

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.7-cp311-cp311-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.7-cp311-cp311-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 aff1972d63a5b25e80e18e4513b02891e74b8bd3033c230cff50ae8eb08092b3
MD5 13f6011ab1a6df965a54a78ebb0543ac
BLAKE2b-256 f47aafaf0cedd51c77ccff252503d84a3a76a0d08213957edc2b3a006f31d657

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.7-cp310-cp310-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.7-cp310-cp310-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 fc07f99930512cdf226f0853345c60f8cae2334a909c7c63a91ba0fe9e144343
MD5 e33e3f720142cdfe583c4e6cc524d240
BLAKE2b-256 aa53637565ddeeb04b9178d0d0f5f40fba19ae41dccf6bad14453c1a6863d21d

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.7-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.7-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 e33f879bd0d7533b0a3f513192f217933d3f0c50148304d323ae367a59e29b15
MD5 6cfeaa8c5406a61c39f237ed3afc051c
BLAKE2b-256 b559b4d434c3b6b93814560e2cfba38330a696205413ec21d01a7a5de26e964b

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.7-cp310-cp310-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.7-cp310-cp310-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 15c61638df5dc58f0820f7a7964efe497de8a3cf5de76ff0ffdf554e2b896441
MD5 584ddd83900ffe15f5351eb376e2d290
BLAKE2b-256 fb8cbcdedf1c488eecc0b4b0206846e03f84341cf967300009637496701b3d5a

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.7-cp39-cp39-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.7-cp39-cp39-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 bb6aa3889a76ec97da95db82a22d1e165a41dec0ecf007ceff574515c01c63d3
MD5 f912cb140ff905b571e47077d8f06847
BLAKE2b-256 8a536c90cd066126f029c7d787990b9d5aacdb66fdaf72cfe65126f6f3797cd4

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.7-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.7-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 268f3747744eb12154a4fab551ff8d30692b3230bef5e8a61270d58cdad05f38
MD5 34420deec7ef4591786fa706f558c747
BLAKE2b-256 e686b15ee21d55339eb1d16b36244ba3abd1f175cb21f898531d488aa8878ebd

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.7-cp39-cp39-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.7-cp39-cp39-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 fe7f5f65350adf5d4028c59c13821d7fb3325d949edb65f5fdf6d359ae91a243
MD5 f4eaf7257b7956a9b274a693249dbf0d
BLAKE2b-256 efe64f7f610d95ecd534330117a19eeb73f1048e3f141d8767a767616ba89b46

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.7-cp38-cp38-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.7-cp38-cp38-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 41c44d1eef3562a305ebd22e05ce955033b99b4fc9e9d9e7b4b222b5f36e0c1f
MD5 5ebe6c9c485c7edf8eb3b80d98306270
BLAKE2b-256 83741127f87d9fdbb1657ce7224461dd058eb5af9059b8aca47f1363c9a86526

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.7-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.7-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 ed6bef08d552701b3a9e649a0054d35dd551b736853a6f87505592e3754c8b44
MD5 bdae3d5d344805833bd8004ddffecdcc
BLAKE2b-256 52be378420ab7f8a8370b5f24ab0a398b56799b17e15a211022a08341d2e1fdb

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.7-cp38-cp38-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.7-cp38-cp38-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 71dc109fa92f438ef6e40d8250982750009ce194147b264e6b6797f596b63f06
MD5 6cb866601347bff638b88ac209f1a1ef
BLAKE2b-256 1e3616ea54d414b5f9d0b2470506fab83476615ea41f1700dd5053f989f1f50d

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.7-cp37-cp37m-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.7-cp37-cp37m-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 f2ff090824ed51b4595f71e08c16e8cf6f2d7ce81148b6369d3debfeb932cc92
MD5 b1e182638bd12d669d2b53638e0a7e49
BLAKE2b-256 7884e5e290fb77a7b6449d8b7e7295af167e51de2c5d26bdf16a49a5e905f225

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.7-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.7-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 0d0dda786392e72b2ff779eac52ec1a6b2fb95edd0f8fe972a0f5a9c99f5c57c
MD5 302538d307ddac7205b590e4380585cf
BLAKE2b-256 f7bf42a888b432ee4e6a81402bf3062260fa269fe02eeb3024320f7899ab68fa

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.7-cp37-cp37m-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.7-cp37-cp37m-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 92a8939d2fc1ddeb08b0517913acc66b761147ac2bd7a1b308e2abf50aa694f7
MD5 a4f100e077430528b8a6744dc99b2abd
BLAKE2b-256 80e69820c1e8b3c4671e7893eabbd994fac230b45535a2647f6e42ff315d59ee

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page