[英]Kafka + Avro + Python
我想由生產者發送消息並由消費者獲取它們。 它必須在avro中,但我不知道該怎么做。 看一看:
schema = { "type":"record", "name":"myrecord", "fields": [{"name":"typ","type":"string"}, {"name":"pred","type":"int"}
producer = KafkaProducer(bootstrap_servers=['xxxx:xxxx'],value_serializer = avro.schema.parse(json.dumps(schema)))
for i in range(100):
message = {"typ":"sth","pred":i}
producer.send("xxxx", value=message)
你能幫我如何正確地做到這一點嗎?
使用kafka-python
, value_serializer
需要是value的 function ,而不是解析的 Avro 模式。
例如
from avro.io import DatumWriter
schema_def = { ... }
schema = avro.schema.parse(json.dumps(schema_def).encode('utf-8'))
def serialize(value):
writer = DatumWriter()
# TODO: add schema to writer
# TODO: write value payload to writer
# TODO: return writer to bytes
producer = KafkaProducer(value_serializer=serialize)
這比您真正需要做的工作更多。 請查看confluent-kafka-python
示例代碼 - https://github.com/confluentinc/confluent-kafka-python/blob/master/examples/avro_producer.py
這樣的事情應該可以解決問題:
from kafka import KafkaProducer
import io
from avro.schema import Parse
from avro.io import DatumWriter, DatumReader, BinaryEncoder, BinaryDecoder
# Create a Kafka client ready to produce messages
producer = KafkaProducer(bootstrap_servers=bootstrap_address,
security_protocol="...", ...)
# Get the schema to use to serialize the message
schema = Parse(open(FILENAME_WHERE_YOU_HAVE_YOUR_AVRO_SCHEMA, "rb").read())
# serialize the message data using the schema
buf = io.BytesIO()
encoder = BinaryEncoder(buf)
writer = DatumWriter(writer_schema=schema)
writer.write(myobject, encoder)
buf.seek(0)
message_data = (buf.read())
# message key if needed
key = None
# headers if needed
headers = []
# Send the serialized message to the Kafka topic
producer.send(topicname,
message_data,
key,
headers)
producer.flush()
聲明:本站的技術帖子網頁,遵循CC BY-SA 4.0協議,如果您需要轉載,請注明本站網址或者原文地址。任何問題請咨詢:yoyou2525@163.com.