| Kafka Python client |
| ------------------------ |
|
|
| .. image:: https://img.shields.io/badge/kafka-2.6%2C%202.5%2C%202.4%2C%202.3%2C%202.2%2C%202.1%2C%202.0%2C%201.1%2C%201.0%2C%200.11%2C%200.10%2C%200.9%2C%200.8-brightgreen.svg |
| :target: https://kafka-python.readthedocs.io/en/master/compatibility.html |
| .. image:: https://img.shields.io/pypi/pyversions/kafka-python.svg |
| :target: https://pypi.python.org/pypi/kafka-python |
| .. image:: https://coveralls.io/repos/dpkp/kafka-python/badge.svg?branch=master&service=github |
| :target: https://coveralls.io/github/dpkp/kafka-python?branch=master |
| .. image:: https://travis-ci.org/dpkp/kafka-python.svg?branch=master |
| :target: https://travis-ci.org/dpkp/kafka-python |
| .. image:: https://img.shields.io/badge/license-Apache%202-blue.svg |
| :target: https://github.com/dpkp/kafka-python/blob/master/LICENSE |
|
|
| Python client for the Apache Kafka distributed stream processing system. |
| kafka-python is designed to function much like the official java client, with a |
| sprinkling of pythonic interfaces (e.g., consumer iterators). |
|
|
| kafka-python is best used with newer brokers (0.9+), but is backwards-compatible with |
| older versions (to 0.8.0). Some features will only be enabled on newer brokers. |
| For example, fully coordinated consumer groups -- i.e., dynamic partition |
| assignment to multiple consumers in the same group -- requires use of 0.9+ kafka |
| brokers. Supporting this feature for earlier broker releases would require |
| writing and maintaining custom leadership election and membership / health |
| check code (perhaps using zookeeper or consul). For older brokers, you can |
| achieve something similar by manually assigning different partitions to each |
| consumer instance with config management tools like chef, ansible, etc. This |
| approach will work fine, though it does not support rebalancing on failures. |
| See <https://kafka-python.readthedocs.io/en/master/compatibility.html> |
| for more details. |
|
|
| Please note that the master branch may contain unreleased features. For release |
| documentation, please see readthedocs and/or python's inline help. |
| |
| >>> pip install kafka-python |
| |
| |
| KafkaConsumer |
| ************* |
| |
| KafkaConsumer is a high-level message consumer, intended to operate as similarly |
| as possible to the official java client. Full support for coordinated |
| consumer groups requires use of kafka brokers that support the Group APIs: kafka v0.9+. |
| |
| See <https://kafka-python.readthedocs.io/en/master/apidoc/KafkaConsumer.html> |
| for API and configuration details. |
| |
| The consumer iterator returns ConsumerRecords, which are simple namedtuples |
| that expose basic message attributes: topic, partition, offset, key, and value: |
| |
| >>> from kafka import KafkaConsumer |
| >>> consumer = KafkaConsumer('my_favorite_topic') |
| >>> for msg in consumer: |
| ... print (msg) |
| |
| >>> # join a consumer group for dynamic partition assignment and offset commits |
| >>> from kafka import KafkaConsumer |
| >>> consumer = KafkaConsumer('my_favorite_topic', group_id='my_favorite_group') |
| >>> for msg in consumer: |
| ... print (msg) |
| |
| >>> # manually assign the partition list for the consumer |
| >>> from kafka import TopicPartition |
| >>> consumer = KafkaConsumer(bootstrap_servers='localhost:1234') |
| >>> consumer.assign([TopicPartition('foobar', 2)]) |
| >>> msg = next(consumer) |
| |
| >>> # Deserialize msgpack-encoded values |
| >>> consumer = KafkaConsumer(value_deserializer=msgpack.loads) |
| >>> consumer.subscribe(['msgpackfoo']) |
| >>> for msg in consumer: |
| ... assert isinstance(msg.value, dict) |
| |
| >>> # Access record headers. The returned value is a list of tuples |
| >>> # with str, bytes for key and value |
| >>> for msg in consumer: |
| ... print (msg.headers) |
| |
| >>> # Get consumer metrics |
| >>> metrics = consumer.metrics() |
| |
| |
| KafkaProducer |
| ************* |
| |
| KafkaProducer is a high-level, asynchronous message producer. The class is |
| intended to operate as similarly as possible to the official java client. |
| See <https://kafka-python.readthedocs.io/en/master/apidoc/KafkaProducer.html> |
| for more details. |
| |
| >>> from kafka import KafkaProducer |
| >>> producer = KafkaProducer(bootstrap_servers='localhost:1234') |
| >>> for _ in range(100): |
| ... producer.send('foobar', b'some_message_bytes') |
| |
| >>> # Block until a single message is sent (or timeout) |
| >>> future = producer.send('foobar', b'another_message') |
| >>> result = future.get(timeout=60) |
| |
| >>> # Block until all pending messages are at least put on the network |
| >>> # NOTE: This does not guarantee delivery or success! It is really |
| >>> # only useful if you configure internal batching using linger_ms |
| >>> producer.flush() |
| |
| >>> # Use a key for hashed-partitioning |
| >>> producer.send('foobar', key=b'foo', value=b'bar') |
| |
| >>> # Serialize json messages |
| >>> import json |
| >>> producer = KafkaProducer(value_serializer=lambda v: json.dumps(v).encode('utf-8')) |
| >>> producer.send('fizzbuzz', {'foo': 'bar'}) |
| |
| >>> # Serialize string keys |
| >>> producer = KafkaProducer(key_serializer=str.encode) |
| >>> producer.send('flipflap', key='ping', value=b'1234') |
| |
| >>> # Compress messages |
| >>> producer = KafkaProducer(compression_type='gzip') |
| >>> for i in range(1000): |
| ... producer.send('foobar', b'msg %d' % i) |
| |
| >>> # Include record headers. The format is list of tuples with string key |
| >>> # and bytes value. |
| >>> producer.send('foobar', value=b'c29tZSB2YWx1ZQ==', headers=[('content-encoding', b'base64')]) |
| |
| >>> # Get producer performance metrics |
| >>> metrics = producer.metrics() |
| |
| |
| Thread safety |
| ************* |
| |
| The KafkaProducer can be used across threads without issue, unlike the |
| KafkaConsumer which cannot. |
| |
| While it is possible to use the KafkaConsumer in a thread-local manner, |
| multiprocessing is recommended. |
| |
| |
| Compression |
| *********** |
| |
| kafka-python supports the following compression formats: |
| |
| - gzip |
| - LZ4 |
| - Snappy |
| - Zstandard (zstd) |
| |
| gzip is supported natively, the others require installing additional libraries. |
| See <https://kafka-python.readthedocs.io/en/master/install.html> for more information. |
| |
| |
| Optimized CRC32 Validation |
| ************************** |
| |
| Kafka uses CRC32 checksums to validate messages. kafka-python includes a pure |
| python implementation for compatibility. To improve performance for high-throughput |
| applications, kafka-python will use `crc32c` for optimized native code if installed. |
| See <https://kafka-python.readthedocs.io/en/master/install.html> for installation instructions. |
| See https://pypi.org/project/crc32c/ for details on the underlying crc32c lib. |
| |
| |
| Protocol |
| ******** |
| |
| A secondary goal of kafka-python is to provide an easy-to-use protocol layer |
| for interacting with kafka brokers via the python repl. This is useful for |
| testing, probing, and general experimentation. The protocol support is |
| leveraged to enable a KafkaClient.check_version() method that |
| probes a kafka broker and attempts to identify which version it is running |
| (0.8.0 to 2.6+). |
| |