From 6cd16d26b8f44265c03b6387b7312a54fcc035a6 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Markus=20M=C3=A4kel=C3=A4?= Date: Mon, 30 Jan 2017 20:14:27 +0200 Subject: [PATCH] Flush the producer after every new record As the cdc_kafka_producer script is an example, it should flush the producer after every new record. This should make it easier to see that events from MaxScale are sent to Kafka. --- server/modules/protocol/examples/cdc_kafka_producer.py | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/server/modules/protocol/examples/cdc_kafka_producer.py b/server/modules/protocol/examples/cdc_kafka_producer.py index 5e09f0734..cddcf1bff 100755 --- a/server/modules/protocol/examples/cdc_kafka_producer.py +++ b/server/modules/protocol/examples/cdc_kafka_producer.py @@ -48,6 +48,7 @@ while True: data = decoder.raw_decode(rbuf.decode('ascii')) rbuf = rbuf[data[1]:] producer.send(topic=opts.kafka_topic, value=json.dumps(data[0]).encode()) + producer.flush() # JSONDecoder will return a ValueError if a partial JSON object is read except ValueError as err: @@ -57,5 +58,3 @@ while True: except Exception as ex: print(ex) break - -producer.flush()