I use filebeat to publish log to kafka, but get
"2018/01/25 08:39:54.536086 client.go:203: ERR Kafka (topic=filebeat-test-logmiss30): dropping too large message of size 3463."
We've been reviewing the code and it seems my original diagnosis is wrong. Changing the bulk_max_size may just be mitigating the problem out of sheer luck, but also reducing the throughput. Is not a good fix.
We would like to investigate this problem further, can you provide us with:
Full filebeat.yml
Full Kafka broker configuration
tcpdump of Kafka traffic when this problem happens
2018/01/29 03:53:29.703780 client.go:203: ERR Kafka (topic=filebeat-test-0129): dropping too large message of size 3446.
2018/01/29 03:53:29.703802 client.go:203: ERR Kafka (topic=filebeat-test-0129): dropping too large message of size 3446.
2018/01/29 03:53:29.703811 client.go:203: ERR Kafka (topic=filebeat-test-0129): dropping too large message of size 3446.
===================================================================================== Fix
To ensure that "output.kafka bulk_max_size times max_message_bytes should smaller than Kafka message.max.bytes", I set the topic-level max.message.bytes to 52428800.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.