This is an automated email from the ASF dual-hosted git repository. gurwls223 pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push: new db0da0c0b52 [SPARK-45952][PYTHON][DOCS] Use built-in math constants in math functions db0da0c0b52 is described below commit db0da0c0b52bcbc0d9ac2634773a5e21d45dc691 Author: Ruifeng Zheng <ruife...@apache.org> AuthorDate: Fri Nov 17 09:26:41 2023 +0900 [SPARK-45952][PYTHON][DOCS] Use built-in math constants in math functions ### What changes were proposed in this pull request? Use the newly added built-in math constants (`PI` and `E`) in math functions ### Why are the changes needed? to improve the docstring ### Does this PR introduce _any_ user-facing change? yes, doc change ### How was this patch tested? ci ### Was this patch authored or co-authored using generative AI tooling? no Closes #43837 from zhengruifeng/py_doc_math. Authored-by: Ruifeng Zheng <ruife...@apache.org> Signed-off-by: Hyukjin Kwon <gurwls...@apache.org> --- python/pyspark/sql/functions.py | 107 ++++++++++++++++++++++++++-------------- 1 file changed, 69 insertions(+), 38 deletions(-) diff --git a/python/pyspark/sql/functions.py b/python/pyspark/sql/functions.py index e3b8e4965e4..655806e8377 100644 --- a/python/pyspark/sql/functions.py +++ b/python/pyspark/sql/functions.py @@ -1838,10 +1838,13 @@ def cos(col: "ColumnOrName") -> Column: Examples -------- - >>> import math - >>> df = spark.range(1) - >>> df.select(cos(lit(math.pi))).first() - Row(COS(3.14159...)=-1.0) + >>> from pyspark.sql import functions as sf + >>> spark.range(1).select(sf.cos(sf.pi())).show() + +---------+ + |COS(PI())| + +---------+ + | -1.0| + +---------+ """ return _invoke_function_over_columns("cos", col) @@ -1897,10 +1900,13 @@ def cot(col: "ColumnOrName") -> Column: Examples -------- - >>> import math - >>> df = spark.range(1) - >>> df.select(cot(lit(math.radians(45)))).first() - Row(COT(0.78539...)=1.00000...) + >>> from pyspark.sql import functions as sf + >>> spark.range(1).select(sf.cot(sf.pi() / 4)).show() + +------------------+ + | COT((PI() / 4))| + +------------------+ + |1.0000000000000...| + +------------------+ """ return _invoke_function_over_columns("cot", col) @@ -1927,10 +1933,13 @@ def csc(col: "ColumnOrName") -> Column: Examples -------- - >>> import math - >>> df = spark.range(1) - >>> df.select(csc(lit(math.radians(90)))).first() - Row(CSC(1.57079...)=1.0) + >>> from pyspark.sql import functions as sf + >>> spark.range(1).select(sf.csc(sf.pi() / 2)).show() + +---------------+ + |CSC((PI() / 2))| + +---------------+ + | 1.0| + +---------------+ """ return _invoke_function_over_columns("csc", col) @@ -2091,10 +2100,13 @@ def log(col: "ColumnOrName") -> Column: Examples -------- - >>> import math - >>> df = spark.range(1) - >>> df.select(log(lit(math.e))).first() - Row(ln(2.71828...)=1.0) + >>> from pyspark.sql import functions as sf + >>> spark.range(1).select(sf.log(sf.e())).show() + +-------+ + |ln(E())| + +-------+ + | 1.0| + +-------+ """ return _invoke_function_over_columns("log", col) @@ -2154,15 +2166,22 @@ def log1p(col: "ColumnOrName") -> Column: Examples -------- - >>> import math - >>> df = spark.range(1) - >>> df.select(log1p(lit(math.e))).first() - Row(LOG1P(2.71828...)=1.31326...) + >>> from pyspark.sql import functions as sf + >>> spark.range(1).select(sf.log1p(sf.e())).show() + +------------------+ + | LOG1P(E())| + +------------------+ + |1.3132616875182...| + +------------------+ Same as: - >>> df.select(log(lit(math.e+1))).first() - Row(ln(3.71828...)=1.31326...) + >>> spark.range(1).select(sf.log(sf.e() + 1)).show() + +------------------+ + | ln((E() + 1))| + +------------------+ + |1.3132616875182...| + +------------------+ """ return _invoke_function_over_columns("log1p", col) @@ -2416,10 +2435,13 @@ def sin(col: "ColumnOrName") -> Column: Examples -------- - >>> import math - >>> df = spark.range(1) - >>> df.select(sin(lit(math.radians(90)))).first() - Row(SIN(1.57079...)=1.0) + >>> from pyspark.sql import functions as sf + >>> spark.range(1).select(sf.sin(sf.pi() / 2)).show() + +---------------+ + |SIN((PI() / 2))| + +---------------+ + | 1.0| + +---------------+ """ return _invoke_function_over_columns("sin", col) @@ -2476,10 +2498,13 @@ def tan(col: "ColumnOrName") -> Column: Examples -------- - >>> import math - >>> df = spark.range(1) - >>> df.select(tan(lit(math.radians(45)))).first() - Row(TAN(0.78539...)=0.99999...) + >>> from pyspark.sql import functions as sf + >>> spark.range(1).select(sf.tan(sf.pi() / 4)).show() + +------------------+ + | TAN((PI() / 4))| + +------------------+ + |0.9999999999999...| + +------------------+ """ return _invoke_function_over_columns("tan", col) @@ -2507,10 +2532,13 @@ def tanh(col: "ColumnOrName") -> Column: Examples -------- - >>> import math - >>> df = spark.range(1) - >>> df.select(tanh(lit(math.radians(90)))).first() - Row(TANH(1.57079...)=0.91715...) + >>> from pyspark.sql import functions as sf + >>> spark.range(1).select(sf.tanh(sf.pi() / 2)).show() + +------------------+ + | TANH((PI() / 2))| + +------------------+ + |0.9171523356672744| + +------------------+ """ return _invoke_function_over_columns("tanh", col) @@ -3953,10 +3981,13 @@ def degrees(col: "ColumnOrName") -> Column: Examples -------- - >>> import math - >>> df = spark.range(1) - >>> df.select(degrees(lit(math.pi))).first() - Row(DEGREES(3.14159...)=180.0) + >>> from pyspark.sql import functions as sf + >>> spark.range(1).select(sf.degrees(sf.pi())).show() + +-------------+ + |DEGREES(PI())| + +-------------+ + | 180.0| + +-------------+ """ return _invoke_function_over_columns("degrees", col) --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org