viirya commented on a change in pull request #30885:
URL: https://github.com/apache/spark/pull/30885#discussion_r547420351



##########
File path: 
sql/core/src/main/scala/org/apache/spark/sql/streaming/DataStreamWriter.scala
##########
@@ -301,12 +301,24 @@ final class DataStreamWriter[T] private[sql](ds: 
Dataset[T]) {
   def start(): StreamingQuery = startInternal(None)
 
   /**
+   * :: Experimental ::
+   *
    * Starts the execution of the streaming query, which will continually 
output results to the given
    * table as new data arrives. A new table will be created if the table not 
exists. The returned

Review comment:
       Maybe we have two small paragraphs for v1 and v2 table separately? E.g.
   
   ```
   For v1 table, partitioning columns provided by `partitionBy` will be 
respected no matter the table exists or not. A new table will be created if the 
table not exists.
   
   For v2 table, `partitionBy` will be ignored if the table already exists. 
`partitionBy` will be respected only if the v2 table does not exist. Besides, 
the v2 table created by this API lacks some functionalities (e.g., customized 
properties, options, and serde info). If you need them, please create the v2 
table manually before the execution to avoid creating a table with incomplete 
information.
   ```




----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]



---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to