Does batching actually work? #255
-
I am trying to get batching working with the connect client (on a self-hosted, vanilla Kafka deployment), and no matter what I try my batch size is always one record. The documentation only mentions
but they don't seem to do anything. I've tried setting them in
Does anyone have this working, and if so, can they share a working configuration? TIA |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 1 reply
-
Hi @MarkCWirt ! So it kind of depends on a few quirks of Kafka Connect:
The connector itself doesn't do batching/caching beyond the built-in Connect mechanisms, and those properties are what controls it. (A fuller listing could be found here) Hope that helps! |
Beta Was this translation helpful? Give feedback.
-
Thanks you for you input, @Paultagoras ! I think I made some progress. An issue I was having was knowing what to name the properties. In the So my
and it seems to be working: my records are around 1k, and the connector seems to be batching about 200 records, which would be expected (I'm not really sure how maximum wait time and minimum bytes interact with one another, but now that batching is taking place I should be able to figue it out.) Once again, thanks for your help! |
Beta Was this translation helpful? Give feedback.
Hi @MarkCWirt ! So it kind of depends on a few quirks of Kafka Connect:
max.poll.records
is a property, it might be set asCONNECT_CONSUMER_MAX_POLL_RECORDS
(like in my docker compose file) orconsumer.override.max.poll.records
(like on Confluent). This also applies tofetch.min.bytes
as you can imagine.The connector itself doesn't do batching/caching beyond the built-in Connect mechanisms, and those properties are what controls it. (A fuller listing could be found here)
Hop…