viirya commented on a change in pull request #30885:
URL: https://github.com/apache/spark/pull/30885#discussion_r547420351
##########
File path:
sql/core/src/main/scala/org/apache/spark/sql/streaming/DataStreamWriter.scala
##########
@@ -301,12 +301,24 @@ final class DataStreamWriter[T] private[sql](ds:
Dataset[T]) {
def start(): StreamingQuery = startInternal(None)
/**
+ * :: Experimental ::
+ *
* Starts the execution of the streaming query, which will continually
output results to the given
* table as new data arrives. A new table will be created if the table not
exists. The returned
Review comment:
Maybe we have two small paragraphs for v1 and v2 table separately? E.g.
```
For v1 table, partitioning columns provided by `partitionBy` will be
respected
no matter the table exists or not. A new table will be created if the table
not exists.
For v2 table, `partitionBy` will be ignored if the table already exists.
`partitionBy`
will be respected only if the v2 table does not exist. Besides, the v2 table
created
by this API lacks some functionalities (e.g., customized properties,
options, and serde info).
If you need them, please create the v2 table manually before the execution
to avoid
creating a table with incomplete information.
```
----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]