I have a Kafka topic that started at about 100GB that I tried to read in to a IMap with Hazelcast-jet. The machine has plenty of memory and I gave it 300 GB of heap. The topic was partitioned into 147 partitions, but when I run the code telling the Pipeline to read from the topic at "earliest" with local parallelism set to 84, the process doesn't seem to use many cores and after running a while doesn't have anywhere near the number of entries there should be in the map (as compared to data ingested into Elastic search at the same time). Now that the topic has grown beyond 500GB I would expect that the process would eventually run out of memory, but it still seems to not use many cores and load only a fraction of the data.
Does anyone have any ideas why this might be?