This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
     new 8aaa9e8  [SPARK-35250][SQL][DOCS] Fix duplicated STOP_AT_DELIMITER to 
SKIP_VALUE at CSV's unescapedQuoteHandling option documentation
8aaa9e8 is described below

commit 8aaa9e890acfa318775627fbe6d12eaf34bb35b1
Author: HyukjinKwon <[email protected]>
AuthorDate: Tue May 4 08:44:18 2021 +0900

    [SPARK-35250][SQL][DOCS] Fix duplicated STOP_AT_DELIMITER to SKIP_VALUE at 
CSV's unescapedQuoteHandling option documentation
    
    ### What changes were proposed in this pull request?
    
    This is rather a followup of https://github.com/apache/spark/pull/30518 
that should be ported back to `branch-3.1` too.
    `STOP_AT_DELIMITER` was mistakenly used twice. The duplicated 
`STOP_AT_DELIMITER` should be `SKIP_VALUE` in the documentation.
    
    ### Why are the changes needed?
    
    To correctly document.
    
    ### Does this PR introduce _any_ user-facing change?
    
    Yes, it fixes the user-facing documentation.
    
    ### How was this patch tested?
    
    I checked them via running linters.
    
    Closes #32423 from HyukjinKwon/SPARK-35250.
    
    Authored-by: HyukjinKwon <[email protected]>
    Signed-off-by: HyukjinKwon <[email protected]>
---
 python/pyspark/sql/readwriter.py                                        | 2 +-
 python/pyspark/sql/streaming.py                                         | 2 +-
 sql/core/src/main/scala/org/apache/spark/sql/DataFrameReader.scala      | 2 +-
 .../main/scala/org/apache/spark/sql/streaming/DataStreamReader.scala    | 2 +-
 4 files changed, 4 insertions(+), 4 deletions(-)

diff --git a/python/pyspark/sql/readwriter.py b/python/pyspark/sql/readwriter.py
index c4134b7..31c1f2f 100644
--- a/python/pyspark/sql/readwriter.py
+++ b/python/pyspark/sql/readwriter.py
@@ -728,7 +728,7 @@ class DataFrameReader(OptionUtils):
             * ``STOP_AT_DELIMITER``: If unescaped quotes are found in the 
input, consider the value
               as an unquoted value. This will make the parser accumulate all 
characters until the
               delimiter or a line ending is found in the input.
-            * ``STOP_AT_DELIMITER``: If unescaped quotes are found in the 
input, the content parsed
+            * ``SKIP_VALUE``: If unescaped quotes are found in the input, the 
content parsed
               for the given value will be skipped and the value set in 
nullValue will be produced
               instead.
             * ``RAISE_ERROR``: If unescaped quotes are found in the input, a 
TextParsingException
diff --git a/python/pyspark/sql/streaming.py b/python/pyspark/sql/streaming.py
index 8b8484d..2c90d7f 100644
--- a/python/pyspark/sql/streaming.py
+++ b/python/pyspark/sql/streaming.py
@@ -941,7 +941,7 @@ class DataStreamReader(OptionUtils):
             * ``STOP_AT_DELIMITER``: If unescaped quotes are found in the 
input, consider the value
               as an unquoted value. This will make the parser accumulate all 
characters until the
               delimiter or a line ending is found in the input.
-            * ``STOP_AT_DELIMITER``: If unescaped quotes are found in the 
input, the content parsed
+            * ``SKIP_VALUE``: If unescaped quotes are found in the input, the 
content parsed
               for the given value will be skipped and the value set in 
nullValue will be produced
               instead.
             * ``RAISE_ERROR``: If unescaped quotes are found in the input, a 
TextParsingException
diff --git a/sql/core/src/main/scala/org/apache/spark/sql/DataFrameReader.scala 
b/sql/core/src/main/scala/org/apache/spark/sql/DataFrameReader.scala
index 9ff37a6..f7e1903 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/DataFrameReader.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/DataFrameReader.scala
@@ -747,7 +747,7 @@ class DataFrameReader private[sql](sparkSession: 
SparkSession) extends Logging {
    *     <li>`STOP_AT_DELIMITER`: If unescaped quotes are found in the input, 
consider the value
    *     as an unquoted value. This will make the parser accumulate all 
characters until the
    *     delimiter or a line ending is found in the input.</li>
-   *     <li>`STOP_AT_DELIMITER`: If unescaped quotes are found in the input, 
the content parsed
+   *     <li>`SKIP_VALUE`: If unescaped quotes are found in the input, the 
content parsed
    *     for the given value will be skipped and the value set in nullValue 
will be produced
    *     instead.</li>
    *     <li>`RAISE_ERROR`: If unescaped quotes are found in the input, a 
TextParsingException
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/streaming/DataStreamReader.scala 
b/sql/core/src/main/scala/org/apache/spark/sql/streaming/DataStreamReader.scala
index a6913fa..f83269e 100644
--- 
a/sql/core/src/main/scala/org/apache/spark/sql/streaming/DataStreamReader.scala
+++ 
b/sql/core/src/main/scala/org/apache/spark/sql/streaming/DataStreamReader.scala
@@ -412,7 +412,7 @@ final class DataStreamReader private[sql](sparkSession: 
SparkSession) extends Lo
    *     <li>`STOP_AT_DELIMITER`: If unescaped quotes are found in the input, 
consider the value
    *     as an unquoted value. This will make the parser accumulate all 
characters until the
    *     delimiter or a line ending is found in the input.</li>
-   *     <li>`STOP_AT_DELIMITER`: If unescaped quotes are found in the input, 
the content parsed
+   *     <li>`SKIP_VALUE`: If unescaped quotes are found in the input, the 
content parsed
    *     for the given value will be skipped and the value set in nullValue 
will be produced
    *     instead.</li>
    *     <li>`RAISE_ERROR`: If unescaped quotes are found in the input, a 
TextParsingException

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to