IBM Streams Kafka integration
Project description
Overview
Provides functions to read messages from a Kafka broker as a stream and submit tuples to a Kafka broker as messages.
The broker configuration must be done with properties in an application configuration or by using a dictionary variable. The minimum set of properties must contain the bootstrap.servers configuration, which is valid for both consumers and producers, i.e. for the subscribe and publish functions.
It is also possible to use different application configurations for subscribe and publish when special consumer or producer configs must be used.
Sample
A simple hello world example of a Streams application publishing to a topic and the same application consuming the same topic:
from streamsx.topology.topology import Topology
from streamsx.topology.schema import CommonSchema
from streamsx.topology.context import submit, ContextTypes
import streamsx.kafka as kafka
import time
def delay(v):
time.sleep(5.0)
return True
topology = Topology('KafkaHelloWorld')
to_kafka = topology.source(['Hello', 'World!'])
to_kafka = to_kafka.as_string()
# delay tuple by tuple
to_kafka = to_kafka.filter(delay)
# Publish a stream to Kafka using TEST topic, the Kafka servers
# assuming, the broker is running on localhost, port 9092
kafka_props = {}
kafka_props['bootstrap.servers'] = 'localhost:9092'
kafka.publish(to_kafka, 'TEST', kafka_props)
# Subscribe to same topic as a stream
from_kafka = kafka.subscribe(topology, 'TEST', kafka_props, CommonSchema.String)
# You'll find the Hello World! in stdout log file:
from_kafka.print()
submit(ContextTypes.DISTRIBUTED, topology)
# The Streams job is kept running.
Documentation
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file streamsx.kafka-1.2.1.tar.gz.
File metadata
- Download URL: streamsx.kafka-1.2.1.tar.gz
- Upload date:
- Size: 12.1 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: Python-urllib/3.6
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
f5b24873366589fd9658b8a51d9cdf950fbc3620c65e779e1a22d7c6b7701a3b
|
|
| MD5 |
d63913530286c46322b42163823f0448
|
|
| BLAKE2b-256 |
59ff3b3b798f6d3fe83e97eed04390b91a0b17ce0c9b4d5da75cb2f2fa2b1f72
|
File details
Details for the file streamsx.kafka-1.2.1-py2.py3-none-any.whl.
File metadata
- Download URL: streamsx.kafka-1.2.1-py2.py3-none-any.whl
- Upload date:
- Size: 13.9 kB
- Tags: Python 2, Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: Python-urllib/3.6
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
c02bbbb42a7683119cc1a79128a0d8ae69a1a81b57bdac13e172f8f52d129bec
|
|
| MD5 |
2cc2422329697d703d02c215184e02e5
|
|
| BLAKE2b-256 |
967d17d312f89f8a89f432593c8bf51c0d7ef1308d8fd091901c64e0d8291e2d
|