Skip to main content

IBM Streams Kafka integration

Project description

Overview

Provides functions to read messages from a Kafka broker as a stream and submit tuples to a Kafka broker as messages.

The broker configuration must be done with properties in an application configuration or a by using a dictionary variable. The minimum set of properties must contain the bootstrap.servers configuration, which is valid for both consumers and producers, i.e. for the subscribe and publish functions.

It is also possible to use different application configurations for subscribe and publish when special consumer or producer configs must be used.

Sample

A simple hello world example of a Streams application publishing to a topic and the same application consuming the same topic:

from streamsx.topology.topology import Topology
from streamsx.topology.schema import CommonSchema
from streamsx.topology.context import submit
from streamsx.topology.context import ContextTypes
import streamsx.kafka as kafka
import time

def delay (v):
    time.sleep (5.0)
    return True

topo = Topology('KafkaHelloWorld')

to_kafka = topo.source (['Hello', 'World!'])
to_kafka = to_kafka.as_string()
# delay tuple by tuple
to_kafka = to_kafka.filter (delay)

# Publish a stream to Kafka using TEST topic, the Kafka servers
# (bootstrap.servers) are configured in the application configuration 'kafka_props'.
kafka.publish (to_kafka, 'TEST', 'kafka_props')

# Subscribe to same topic as a stream
from_kafka = kafka.subscribe (topo, 'TEST', 'kafka_props', CommonSchema.String)

# You'll find the Hello World! in stdout log file:
from_kafka.print()

submit (ContextTypes.DISTRIBUTED, topo)
# The Streams job is kept running.

Documentation

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

streamsx.kafka-1.0.1.tar.gz (4.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

streamsx.kafka-1.0.1-py2.py3-none-any.whl (9.6 kB view details)

Uploaded Python 2Python 3

File details

Details for the file streamsx.kafka-1.0.1.tar.gz.

File metadata

  • Download URL: streamsx.kafka-1.0.1.tar.gz
  • Upload date:
  • Size: 4.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: Python-urllib/3.5

File hashes

Hashes for streamsx.kafka-1.0.1.tar.gz
Algorithm Hash digest
SHA256 f9d3ca47b8b0788d92ca5b10262f295d0d9989ff1afd5ee7086cc9cb8a4efd1b
MD5 5ab67c8f6346697d4d8706038ea1fabd
BLAKE2b-256 fb416cc9df4c8c07523219682369112143dc4938ad23a40586d9a02ccc874c52

See more details on using hashes here.

File details

Details for the file streamsx.kafka-1.0.1-py2.py3-none-any.whl.

File metadata

File hashes

Hashes for streamsx.kafka-1.0.1-py2.py3-none-any.whl
Algorithm Hash digest
SHA256 67c2c4880f10c5cc90dd8683d19885645c6b0778090dcbeafaa440a587de61f6
MD5 5e0a801ea13c60854ebb47fa900210e9
BLAKE2b-256 8c4d6da9d1cee9f4bb3857112ec76a93f0d09657f9973f8ad66c6a17b88f28cb

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page