Github user daroo commented on the issue:
https://github.com/apache/spark/pull/19789
I don't know what are the plans for the structured streaming which @zsxwing
has in mind, but before I created the PR I actually had seen how this problem
is currently solved in kafka-0-10-sql module and I didn't like it. My main
concern is that in certain cases the cache size my grow well beyond configured
spark.sql.kafkaConsumerCache.capacity. Thatâs why Iâve chosen to do it in a
bit different way.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]