I'm a newbie in Streamsets and Kudu technologies and I'm trying several solutions to reach my goal: I've got a folder containing some Avro files and these files need to be processed and afterward sent to a Kudu schema.
https://i.stack.imgur.com/l5Yf9.jpg
When using an Avro file containing a couple hundreds of records all goes right, but when the number of records increases to 16k this error is shown:
Caused by:
org.apache.kudu.client. NonRecoverableException:
MANUAL_FLUSH is enabled but the buffer is too big.
I've searched in all available configurations both on Streamsets and Kudu and the only solution that I was able to apply consists in editing the Java source code, deleting a single row that switched from the default flush mode to the manual one; this works but it's not the optimal solution because it requires to edit and compile this file each time I want to use it on a new machine.
Anyone knows how to avoid this happens?
Thanks in advance!