itholic commented on code in PR #37575:
URL: https://github.com/apache/spark/pull/37575#discussion_r949905055
##########
python/pyspark/sql/functions.py:
##########
@@ -155,34 +182,143 @@ def col(col: str) -> Column:
column = col
-@since(1.3)
def asc(col: "ColumnOrName") -> Column:
"""
Returns a sort expression based on the ascending order of the given column
name.
+
+ .. versionadded:: 1.3.0
+
+ Parameters
+ ----------
+ col : :class:`~pyspark.sql.Column` or str
+ target column to sort by in the ascending order.
+
+ Returns
+ -------
+ :class:`~pyspark.sql.Column`
+ the column specifying the order.
+
+ Examples
+ --------
+ Sort by the column 'id' in the descending order.
+
+ >>> df = spark.range(5)
+ >>> df = df.sort(desc("id"))
+ >>> df.show()
+ +---+
+ | id|
+ +---+
+ | 4|
+ | 3|
+ | 2|
+ | 1|
+ | 0|
+ +---+
+
+ Sort by the column 'id' in the ascending order.
+
+ >>> df.orderBy(asc("id")).show()
+ +---+
+ | id|
+ +---+
+ | 0|
+ | 1|
+ | 2|
+ | 3|
+ | 4|
+ +---+
"""
return col.asc() if isinstance(col, Column) else _invoke_function("asc",
col)
-@since(1.3)
def desc(col: "ColumnOrName") -> Column:
"""
Returns a sort expression based on the descending order of the given
column name.
+
+ .. versionadded:: 1.3.0
+
+ Parameters
+ ----------
+ col : :class:`~pyspark.sql.Column` or str
+ target column to sort by in the descending order.
+
+ Returns
+ -------
+ :class:`~pyspark.sql.Column`
+ the column specifying the order.
+
+ Examples
+ --------
+ Sort by the column 'id' in the descending order.
+
+ >>> spark.range(5).orderBy(desc("id")).show()
+ +---+
+ | id|
+ +---+
+ | 4|
+ | 3|
+ | 2|
+ | 1|
+ | 0|
+ +---+
"""
return col.desc() if isinstance(col, Column) else _invoke_function("desc",
col)
-@since(1.3)
def sqrt(col: "ColumnOrName") -> Column:
"""
Computes the square root of the specified float value.
+
+ .. versionadded:: 1.3.0
+
+ Parameters
+ ----------
+ col : :class:`~pyspark.sql.Column` or str
+ target column to compute on.
+
+ Returns
+ -------
+ :class:`~pyspark.sql.Column`
+ column for compuated results.
+
+ Examples
+ --------
+ >>> df = spark.range(1)
+ >>> df.select(sqrt(lit(4))).show()
+ +-------+
+ |SQRT(4)|
+ +-------+
+ | 2.0|
+ +-------+
"""
return _invoke_function_over_columns("sqrt", col)
-@since(1.3)
def abs(col: "ColumnOrName") -> Column:
"""
Computes the absolute value.
+
+ .. versionadded:: 1.3.0
+
+ Parameters
+ ----------
+ col : :class:`~pyspark.sql.Column` or str
+ target column to compute on.
+
+ Returns
+ -------
+ :class:`~pyspark.sql.Column`
+ column for compuated results.
Review Comment:
ditto ?
##########
python/pyspark/sql/functions.py:
##########
@@ -155,34 +182,143 @@ def col(col: str) -> Column:
column = col
-@since(1.3)
def asc(col: "ColumnOrName") -> Column:
"""
Returns a sort expression based on the ascending order of the given column
name.
+
+ .. versionadded:: 1.3.0
+
+ Parameters
+ ----------
+ col : :class:`~pyspark.sql.Column` or str
+ target column to sort by in the ascending order.
+
+ Returns
+ -------
+ :class:`~pyspark.sql.Column`
+ the column specifying the order.
+
+ Examples
+ --------
+ Sort by the column 'id' in the descending order.
+
+ >>> df = spark.range(5)
+ >>> df = df.sort(desc("id"))
+ >>> df.show()
+ +---+
+ | id|
+ +---+
+ | 4|
+ | 3|
+ | 2|
+ | 1|
+ | 0|
+ +---+
+
+ Sort by the column 'id' in the ascending order.
+
+ >>> df.orderBy(asc("id")).show()
+ +---+
+ | id|
+ +---+
+ | 0|
+ | 1|
+ | 2|
+ | 3|
+ | 4|
+ +---+
"""
return col.asc() if isinstance(col, Column) else _invoke_function("asc",
col)
-@since(1.3)
def desc(col: "ColumnOrName") -> Column:
"""
Returns a sort expression based on the descending order of the given
column name.
+
+ .. versionadded:: 1.3.0
+
+ Parameters
+ ----------
+ col : :class:`~pyspark.sql.Column` or str
+ target column to sort by in the descending order.
+
+ Returns
+ -------
+ :class:`~pyspark.sql.Column`
+ the column specifying the order.
+
+ Examples
+ --------
+ Sort by the column 'id' in the descending order.
+
+ >>> spark.range(5).orderBy(desc("id")).show()
+ +---+
+ | id|
+ +---+
+ | 4|
+ | 3|
+ | 2|
+ | 1|
+ | 0|
+ +---+
"""
return col.desc() if isinstance(col, Column) else _invoke_function("desc",
col)
-@since(1.3)
def sqrt(col: "ColumnOrName") -> Column:
"""
Computes the square root of the specified float value.
+
+ .. versionadded:: 1.3.0
+
+ Parameters
+ ----------
+ col : :class:`~pyspark.sql.Column` or str
+ target column to compute on.
+
+ Returns
+ -------
+ :class:`~pyspark.sql.Column`
+ column for compuated results.
+
+ Examples
+ --------
+ >>> df = spark.range(1)
+ >>> df.select(sqrt(lit(4))).show()
+ +-------+
+ |SQRT(4)|
+ +-------+
+ | 2.0|
+ +-------+
"""
return _invoke_function_over_columns("sqrt", col)
-@since(1.3)
def abs(col: "ColumnOrName") -> Column:
"""
Computes the absolute value.
+
+ .. versionadded:: 1.3.0
+
+ Parameters
+ ----------
+ col : :class:`~pyspark.sql.Column` or str
+ target column to compute on.
+
+ Returns
+ -------
+ :class:`~pyspark.sql.Column`
+ column for compuated results.
Review Comment:
And it seems there are more of the same typos below.
##########
python/pyspark/sql/functions.py:
##########
@@ -155,34 +182,143 @@ def col(col: str) -> Column:
column = col
-@since(1.3)
def asc(col: "ColumnOrName") -> Column:
"""
Returns a sort expression based on the ascending order of the given column
name.
+
+ .. versionadded:: 1.3.0
+
+ Parameters
+ ----------
+ col : :class:`~pyspark.sql.Column` or str
+ target column to sort by in the ascending order.
+
+ Returns
+ -------
+ :class:`~pyspark.sql.Column`
+ the column specifying the order.
+
+ Examples
+ --------
+ Sort by the column 'id' in the descending order.
+
+ >>> df = spark.range(5)
+ >>> df = df.sort(desc("id"))
+ >>> df.show()
+ +---+
+ | id|
+ +---+
+ | 4|
+ | 3|
+ | 2|
+ | 1|
+ | 0|
+ +---+
+
+ Sort by the column 'id' in the ascending order.
+
+ >>> df.orderBy(asc("id")).show()
+ +---+
+ | id|
+ +---+
+ | 0|
+ | 1|
+ | 2|
+ | 3|
+ | 4|
+ +---+
"""
return col.asc() if isinstance(col, Column) else _invoke_function("asc",
col)
-@since(1.3)
def desc(col: "ColumnOrName") -> Column:
"""
Returns a sort expression based on the descending order of the given
column name.
+
+ .. versionadded:: 1.3.0
+
+ Parameters
+ ----------
+ col : :class:`~pyspark.sql.Column` or str
+ target column to sort by in the descending order.
+
+ Returns
+ -------
+ :class:`~pyspark.sql.Column`
+ the column specifying the order.
+
+ Examples
+ --------
+ Sort by the column 'id' in the descending order.
+
+ >>> spark.range(5).orderBy(desc("id")).show()
+ +---+
+ | id|
+ +---+
+ | 4|
+ | 3|
+ | 2|
+ | 1|
+ | 0|
+ +---+
"""
return col.desc() if isinstance(col, Column) else _invoke_function("desc",
col)
-@since(1.3)
def sqrt(col: "ColumnOrName") -> Column:
"""
Computes the square root of the specified float value.
+
+ .. versionadded:: 1.3.0
+
+ Parameters
+ ----------
+ col : :class:`~pyspark.sql.Column` or str
+ target column to compute on.
+
+ Returns
+ -------
+ :class:`~pyspark.sql.Column`
+ column for compuated results.
Review Comment:
nit: Maybe `compuated` -> `computed` ?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]