Looks like it's missing, or intended to force custom streaming source
implemented as DSv2.

I'm not sure Spark community wants to expand DSv1 API: I could propose the
change if we get some supports here.

To Spark community: given we bring major changes on DSv2, someone would
want to rely on DSv1 while transition from old DSv2 to new DSv2 happens and
new DSv2 gets stabilized. Would we like to provide necessary changes on
DSv1?

Thanks,
Jungtaek Lim (HeartSaVioR)

On Wed, Oct 2, 2019 at 4:27 AM Jacek Laskowski <ja...@japila.pl> wrote:

> Hi,
>
> I think I've got stuck and without your help I won't move any further.
> Please help.
>
> I'm with Spark 2.4.4 and am developing a streaming Source (DSv1,
> MicroBatch) and in getBatch phase when requested for a DataFrame, there is
> this assert [1] I can't seem to go past with any DataFrame I managed to
> create as it's not streaming.
>
>           assert(batch.isStreaming,
>             s"DataFrame returned by getBatch from $source did not have
> isStreaming=true\n" +
>               s"${batch.queryExecution.logical}")
>
> [1]
> https://github.com/apache/spark/blob/v2.4.4/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/MicroBatchExecution.scala#L439-L441
>
> All I could find is private[sql],
> e.g. SQLContext.internalCreateDataFrame(..., isStreaming = true) [2] or [3]
>
> [2]
> https://github.com/apache/spark/blob/v2.4.4/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala#L422-L428
> [3]
> https://github.com/apache/spark/blob/v2.4.4/sql/core/src/main/scala/org/apache/spark/sql/Dataset.scala#L62-L81
>
> Pozdrawiam,
> Jacek Laskowski
> ----
> https://about.me/JacekLaskowski
> The Internals of Spark SQL https://bit.ly/spark-sql-internals
> The Internals of Spark Structured Streaming
> https://bit.ly/spark-structured-streaming
> The Internals of Apache Kafka https://bit.ly/apache-kafka-internals
> Follow me at https://twitter.com/jaceklaskowski
>
>

Reply via email to