I have a quick newbie question.

Spark Structured Streaming creates an unbounded dataframe that keeps
appending rows to it.

So what's the max size of data it can hold? What if the size becomes bigger
than the JVM? Will it spill to disk? I'm using S3 as storage. So will it
write temp data on S3 or on local file system of the cluster?

Nick

Reply via email to