viirya commented on a change in pull request #30885:
URL: https://github.com/apache/spark/pull/30885#discussion_r547420351
##########
File path:
sql/core/src/main/scala/org/apache/spark/sql/streaming/DataStreamWriter.scala
##########
@@ -301,12 +301,24 @@ final class DataStreamWriter[T] private[sql](ds:
Dataset[T]) {
def start(): StreamingQuery = startInternal(None)
/**
+ * :: Experimental ::
+ *
* Starts the execution of the streaming query, which will continually
output results to the given
* table as new data arrives. A new table will be created if the table not
exists. The returned
Review comment:
Maybe we have two small paragraphs for v1 and v2 table separately? E.g.
```
For v1 table, partitioning columns provided by `partitionBy` will be
respected no matter the table exists or not. A new table will be created if the
table not exists.
For v2 table, `partitionBy` will be ignored if the table already exists.
`partitionBy` will be respected only if the v2 table does not exist. Besides,
the v2 table created by this API lacks some functionalities (e.g., customized
properties, options, and serde info). If you need them, please create the v2
table manually before the execution to avoid creating a table with incomplete
information.
```
----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]