[jira] [Commented] (SPARK-37421) Inline type hints for python/pyspark/mllib/evaluation.py

2021-11-21 Thread Apache Spark (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-37421?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17447223#comment-17447223
 ] 

Apache Spark commented on SPARK-37421:
--

User 'dchvn' has created a pull request for this issue:
https://github.com/apache/spark/pull/34680

> Inline type hints for python/pyspark/mllib/evaluation.py
> 
>
> Key: SPARK-37421
> URL: https://issues.apache.org/jira/browse/SPARK-37421
> Project: Spark
>  Issue Type: Sub-task
>  Components: MLlib, PySpark
>Affects Versions: 3.3.0
>Reporter: Maciej Szymkiewicz
>Priority: Major
>
> Inline type hints from python/pyspark/mlib/evaluation.pyi to 
> python/pyspark/mllib/evaluation.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-37421) Inline type hints for python/pyspark/mllib/evaluation.py

2021-11-21 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37421?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-37421:


Assignee: (was: Apache Spark)

> Inline type hints for python/pyspark/mllib/evaluation.py
> 
>
> Key: SPARK-37421
> URL: https://issues.apache.org/jira/browse/SPARK-37421
> Project: Spark
>  Issue Type: Sub-task
>  Components: MLlib, PySpark
>Affects Versions: 3.3.0
>Reporter: Maciej Szymkiewicz
>Priority: Major
>
> Inline type hints from python/pyspark/mlib/evaluation.pyi to 
> python/pyspark/mllib/evaluation.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-37421) Inline type hints for python/pyspark/mllib/evaluation.py

2021-11-21 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37421?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-37421:


Assignee: Apache Spark

> Inline type hints for python/pyspark/mllib/evaluation.py
> 
>
> Key: SPARK-37421
> URL: https://issues.apache.org/jira/browse/SPARK-37421
> Project: Spark
>  Issue Type: Sub-task
>  Components: MLlib, PySpark
>Affects Versions: 3.3.0
>Reporter: Maciej Szymkiewicz
>Assignee: Apache Spark
>Priority: Major
>
> Inline type hints from python/pyspark/mlib/evaluation.pyi to 
> python/pyspark/mllib/evaluation.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-37421) Inline type hints for python/pyspark/mllib/evaluation.py

2021-11-21 Thread Apache Spark (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-37421?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17447221#comment-17447221
 ] 

Apache Spark commented on SPARK-37421:
--

User 'dchvn' has created a pull request for this issue:
https://github.com/apache/spark/pull/34680

> Inline type hints for python/pyspark/mllib/evaluation.py
> 
>
> Key: SPARK-37421
> URL: https://issues.apache.org/jira/browse/SPARK-37421
> Project: Spark
>  Issue Type: Sub-task
>  Components: MLlib, PySpark
>Affects Versions: 3.3.0
>Reporter: Maciej Szymkiewicz
>Priority: Major
>
> Inline type hints from python/pyspark/mlib/evaluation.pyi to 
> python/pyspark/mllib/evaluation.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-37437) Remove unused profile

2021-11-21 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37437?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-37437:


Assignee: Apache Spark

> Remove unused profile
> -
>
> Key: SPARK-37437
> URL: https://issues.apache.org/jira/browse/SPARK-37437
> Project: Spark
>  Issue Type: Task
>  Components: SQL
>Affects Versions: 3.2.0
>Reporter: angerszhu
>Assignee: Apache Spark
>Priority: Major
>
> remove 
> 
>   hive-2.3
>   
> 



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-37437) Remove unused profile

2021-11-21 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37437?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-37437:


Assignee: (was: Apache Spark)

> Remove unused profile
> -
>
> Key: SPARK-37437
> URL: https://issues.apache.org/jira/browse/SPARK-37437
> Project: Spark
>  Issue Type: Task
>  Components: SQL
>Affects Versions: 3.2.0
>Reporter: angerszhu
>Priority: Major
>
> remove 
> 
>   hive-2.3
>   
> 



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-37437) Remove unused profile

2021-11-21 Thread Apache Spark (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-37437?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17447218#comment-17447218
 ] 

Apache Spark commented on SPARK-37437:
--

User 'AngersZh' has created a pull request for this issue:
https://github.com/apache/spark/pull/34679

> Remove unused profile
> -
>
> Key: SPARK-37437
> URL: https://issues.apache.org/jira/browse/SPARK-37437
> Project: Spark
>  Issue Type: Task
>  Components: SQL
>Affects Versions: 3.2.0
>Reporter: angerszhu
>Priority: Major
>
> remove 
> 
>   hive-2.3
>   
> 



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-37437) Remove unused profile

2021-11-21 Thread Apache Spark (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-37437?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17447219#comment-17447219
 ] 

Apache Spark commented on SPARK-37437:
--

User 'AngersZh' has created a pull request for this issue:
https://github.com/apache/spark/pull/34679

> Remove unused profile
> -
>
> Key: SPARK-37437
> URL: https://issues.apache.org/jira/browse/SPARK-37437
> Project: Spark
>  Issue Type: Task
>  Components: SQL
>Affects Versions: 3.2.0
>Reporter: angerszhu
>Priority: Major
>
> remove 
> 
>   hive-2.3
>   
> 



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-37437) Remove unused profile

2021-11-21 Thread angerszhu (Jira)
angerszhu created SPARK-37437:
-

 Summary: Remove unused profile
 Key: SPARK-37437
 URL: https://issues.apache.org/jira/browse/SPARK-37437
 Project: Spark
  Issue Type: Task
  Components: SQL
Affects Versions: 3.2.0
Reporter: angerszhu


remove 


  hive-2.3
  





--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-37407) Inline type hints for python/pyspark/ml/functions.py

2021-11-21 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-37407:


Assignee: Apache Spark

> Inline type hints for python/pyspark/ml/functions.py
> 
>
> Key: SPARK-37407
> URL: https://issues.apache.org/jira/browse/SPARK-37407
> Project: Spark
>  Issue Type: Sub-task
>  Components: ML, PySpark
>Affects Versions: 3.3.0
>Reporter: Maciej Szymkiewicz
>Assignee: Apache Spark
>Priority: Major
>
> Inline type hints from python/pyspark/ml/functions.pyi to 
> python/pyspark/ml/functions.py.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-37407) Inline type hints for python/pyspark/ml/functions.py

2021-11-21 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-37407:


Assignee: (was: Apache Spark)

> Inline type hints for python/pyspark/ml/functions.py
> 
>
> Key: SPARK-37407
> URL: https://issues.apache.org/jira/browse/SPARK-37407
> Project: Spark
>  Issue Type: Sub-task
>  Components: ML, PySpark
>Affects Versions: 3.3.0
>Reporter: Maciej Szymkiewicz
>Priority: Major
>
> Inline type hints from python/pyspark/ml/functions.pyi to 
> python/pyspark/ml/functions.py.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-37407) Inline type hints for python/pyspark/ml/functions.py

2021-11-21 Thread Apache Spark (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-37407?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17447210#comment-17447210
 ] 

Apache Spark commented on SPARK-37407:
--

User 'dchvn' has created a pull request for this issue:
https://github.com/apache/spark/pull/34678

> Inline type hints for python/pyspark/ml/functions.py
> 
>
> Key: SPARK-37407
> URL: https://issues.apache.org/jira/browse/SPARK-37407
> Project: Spark
>  Issue Type: Sub-task
>  Components: ML, PySpark
>Affects Versions: 3.3.0
>Reporter: Maciej Szymkiewicz
>Priority: Major
>
> Inline type hints from python/pyspark/ml/functions.pyi to 
> python/pyspark/ml/functions.py.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-37436) Uses Python's standard string formatter for SQL API in pandas API on Spark

2021-11-21 Thread Apache Spark (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-37436?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17447208#comment-17447208
 ] 

Apache Spark commented on SPARK-37436:
--

User 'HyukjinKwon' has created a pull request for this issue:
https://github.com/apache/spark/pull/34677

> Uses Python's standard string formatter for SQL API in pandas API on Spark
> --
>
> Key: SPARK-37436
> URL: https://issues.apache.org/jira/browse/SPARK-37436
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Affects Versions: 3.3.0
>Reporter: Hyukjin Kwon
>Priority: Major
>
> Currently {{pyspark.pandas.sql}} uses its own hacky parser: 
> https://github.com/apache/spark/blob/master/python/pyspark/pandas/sql_processor.py
> We should ideally switch it to the standard Python formatter 
> https://docs.python.org/3/library/string.html#custom-string-formatting



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-37436) Uses Python's standard string formatter for SQL API in pandas API on Spark

2021-11-21 Thread Apache Spark (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-37436?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17447207#comment-17447207
 ] 

Apache Spark commented on SPARK-37436:
--

User 'HyukjinKwon' has created a pull request for this issue:
https://github.com/apache/spark/pull/34677

> Uses Python's standard string formatter for SQL API in pandas API on Spark
> --
>
> Key: SPARK-37436
> URL: https://issues.apache.org/jira/browse/SPARK-37436
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Affects Versions: 3.3.0
>Reporter: Hyukjin Kwon
>Priority: Major
>
> Currently {{pyspark.pandas.sql}} uses its own hacky parser: 
> https://github.com/apache/spark/blob/master/python/pyspark/pandas/sql_processor.py
> We should ideally switch it to the standard Python formatter 
> https://docs.python.org/3/library/string.html#custom-string-formatting



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-37436) Uses Python's standard string formatter for SQL API in pandas API on Spark

2021-11-21 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37436?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-37436:


Assignee: Apache Spark

> Uses Python's standard string formatter for SQL API in pandas API on Spark
> --
>
> Key: SPARK-37436
> URL: https://issues.apache.org/jira/browse/SPARK-37436
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Affects Versions: 3.3.0
>Reporter: Hyukjin Kwon
>Assignee: Apache Spark
>Priority: Major
>
> Currently {{pyspark.pandas.sql}} uses its own hacky parser: 
> https://github.com/apache/spark/blob/master/python/pyspark/pandas/sql_processor.py
> We should ideally switch it to the standard Python formatter 
> https://docs.python.org/3/library/string.html#custom-string-formatting



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-37436) Uses Python's standard string formatter for SQL API in pandas API on Spark

2021-11-21 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37436?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-37436:


Assignee: (was: Apache Spark)

> Uses Python's standard string formatter for SQL API in pandas API on Spark
> --
>
> Key: SPARK-37436
> URL: https://issues.apache.org/jira/browse/SPARK-37436
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Affects Versions: 3.3.0
>Reporter: Hyukjin Kwon
>Priority: Major
>
> Currently {{pyspark.pandas.sql}} uses its own hacky parser: 
> https://github.com/apache/spark/blob/master/python/pyspark/pandas/sql_processor.py
> We should ideally switch it to the standard Python formatter 
> https://docs.python.org/3/library/string.html#custom-string-formatting



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37436) Uses Python's standard string formatter for SQL API in pandas API on Spark

2021-11-21 Thread Hyukjin Kwon (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37436?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hyukjin Kwon updated SPARK-37436:
-
Issue Type: Improvement  (was: Test)

> Uses Python's standard string formatter for SQL API in pandas API on Spark
> --
>
> Key: SPARK-37436
> URL: https://issues.apache.org/jira/browse/SPARK-37436
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Affects Versions: 3.3.0
>Reporter: Hyukjin Kwon
>Priority: Major
>
> Currently {{pyspark.pandas.sql}} uses its own hacky parser: 
> https://github.com/apache/spark/blob/master/python/pyspark/pandas/sql_processor.py
> We should ideally switch it to the standard Python formatter 
> https://docs.python.org/3/library/string.html#custom-string-formatting



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-37436) Uses Python's standard string formatter for SQL API in pandas API on Spark

2021-11-21 Thread Hyukjin Kwon (Jira)
Hyukjin Kwon created SPARK-37436:


 Summary: Uses Python's standard string formatter for SQL API in 
pandas API on Spark
 Key: SPARK-37436
 URL: https://issues.apache.org/jira/browse/SPARK-37436
 Project: Spark
  Issue Type: Test
  Components: PySpark
Affects Versions: 3.3.0
Reporter: Hyukjin Kwon


Currently {{pyspark.pandas.sql}} uses its own hacky parser: 
https://github.com/apache/spark/blob/master/python/pyspark/pandas/sql_processor.py

We should ideally switch it to the standard Python formatter 
https://docs.python.org/3/library/string.html#custom-string-formatting



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-36231) Support arithmetic operations of Series containing Decimal(np.nan)

2021-11-21 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-36231?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-36231:


Assignee: Yikun Jiang  (was: Apache Spark)

> Support arithmetic operations of Series containing Decimal(np.nan) 
> ---
>
> Key: SPARK-36231
> URL: https://issues.apache.org/jira/browse/SPARK-36231
> Project: Spark
>  Issue Type: Sub-task
>  Components: PySpark
>Affects Versions: 3.2.0
>Reporter: Xinrong Meng
>Assignee: Yikun Jiang
>Priority: Major
>
> Arithmetic operations of Series containing Decimal(np.nan) raise 
> java.lang.NullPointerException in driver. An example is shown as below:
> {code:java}
> >>> pser = pd.Series([decimal.Decimal(1.0), decimal.Decimal(2.0), 
> >>> decimal.Decimal(np.nan)])
> >>> psser = ps.from_pandas(pser)
> >>> pser + 1
> 0 2
>  1 3
>  2 NaN
> >>> psser + 1
>  Driver stacktrace:
>  at 
> org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2259)
>  at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2208)
>  at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2207)
>  at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
>  at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
>  at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
>  at 
> org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2207)
>  at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1084)
>  at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1084)
>  at scala.Option.foreach(Option.scala:407)
>  at 
> org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1084)
>  at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2446)
>  at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2388)
>  at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2377)
>  at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49)
>  at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:873)
>  at org.apache.spark.SparkContext.runJob(SparkContext.scala:2208)
>  at org.apache.spark.SparkContext.runJob(SparkContext.scala:2303)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$5(Dataset.scala:3648)
>  at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
>  at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1437)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$2(Dataset.scala:3652)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$2$adapted(Dataset.scala:3629)
>  at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3706)
>  at 
> org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:103)
>  at 
> org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:163)
>  at 
> org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:90)
>  at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:774)
>  at 
> org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
>  at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3704)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$1(Dataset.scala:3629)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$1$adapted(Dataset.scala:3628)
>  at 
> org.apache.spark.security.SocketAuthServer$.$anonfun$serveToStream$2(SocketAuthServer.scala:139)
>  at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
>  at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1437)
>  at 
> org.apache.spark.security.SocketAuthServer$.$anonfun$serveToStream$1(SocketAuthServer.scala:141)
>  at 
> org.apache.spark.security.SocketAuthServer$.$anonfun$serveToStream$1$adapted(SocketAuthServer.scala:136)
>  at 
> org.apache.spark.security.SocketFuncServer.handleConnection(SocketAuthServer.scala:113)
>  at 
> org.apache.spark.security.SocketFuncServer.handleConnection(SocketAuthServer.scala:107)
>  at 
> org.apache.spark.security.SocketAuthServer$$anon$1.$anonfun$run$4(SocketAuthServer.scala:68)
>  at scala.util.Try$.apply(Try.scala:213)
>  at 
> org.apache.spark.security.SocketAuthServer$$anon$1.run(SocketAuthServer.scala:68)
>  Caused by: java.lang.NullPointerException
>  at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage1.processNext(Unknown
>  Source)
>  at 
> 

[jira] [Assigned] (SPARK-36231) Support arithmetic operations of Series containing Decimal(np.nan)

2021-11-21 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-36231?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-36231:


Assignee: Apache Spark  (was: Yikun Jiang)

> Support arithmetic operations of Series containing Decimal(np.nan) 
> ---
>
> Key: SPARK-36231
> URL: https://issues.apache.org/jira/browse/SPARK-36231
> Project: Spark
>  Issue Type: Sub-task
>  Components: PySpark
>Affects Versions: 3.2.0
>Reporter: Xinrong Meng
>Assignee: Apache Spark
>Priority: Major
>
> Arithmetic operations of Series containing Decimal(np.nan) raise 
> java.lang.NullPointerException in driver. An example is shown as below:
> {code:java}
> >>> pser = pd.Series([decimal.Decimal(1.0), decimal.Decimal(2.0), 
> >>> decimal.Decimal(np.nan)])
> >>> psser = ps.from_pandas(pser)
> >>> pser + 1
> 0 2
>  1 3
>  2 NaN
> >>> psser + 1
>  Driver stacktrace:
>  at 
> org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2259)
>  at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2208)
>  at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2207)
>  at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
>  at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
>  at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
>  at 
> org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2207)
>  at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1084)
>  at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1084)
>  at scala.Option.foreach(Option.scala:407)
>  at 
> org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1084)
>  at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2446)
>  at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2388)
>  at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2377)
>  at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49)
>  at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:873)
>  at org.apache.spark.SparkContext.runJob(SparkContext.scala:2208)
>  at org.apache.spark.SparkContext.runJob(SparkContext.scala:2303)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$5(Dataset.scala:3648)
>  at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
>  at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1437)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$2(Dataset.scala:3652)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$2$adapted(Dataset.scala:3629)
>  at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3706)
>  at 
> org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:103)
>  at 
> org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:163)
>  at 
> org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:90)
>  at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:774)
>  at 
> org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
>  at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3704)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$1(Dataset.scala:3629)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$1$adapted(Dataset.scala:3628)
>  at 
> org.apache.spark.security.SocketAuthServer$.$anonfun$serveToStream$2(SocketAuthServer.scala:139)
>  at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
>  at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1437)
>  at 
> org.apache.spark.security.SocketAuthServer$.$anonfun$serveToStream$1(SocketAuthServer.scala:141)
>  at 
> org.apache.spark.security.SocketAuthServer$.$anonfun$serveToStream$1$adapted(SocketAuthServer.scala:136)
>  at 
> org.apache.spark.security.SocketFuncServer.handleConnection(SocketAuthServer.scala:113)
>  at 
> org.apache.spark.security.SocketFuncServer.handleConnection(SocketAuthServer.scala:107)
>  at 
> org.apache.spark.security.SocketAuthServer$$anon$1.$anonfun$run$4(SocketAuthServer.scala:68)
>  at scala.util.Try$.apply(Try.scala:213)
>  at 
> org.apache.spark.security.SocketAuthServer$$anon$1.run(SocketAuthServer.scala:68)
>  Caused by: java.lang.NullPointerException
>  at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage1.processNext(Unknown
>  Source)
>  at 
> 

[jira] [Created] (SPARK-37435) Did not find value which can be converted into java.lang.String

2021-11-21 Thread Ziqun Ye (Jira)
Ziqun Ye created SPARK-37435:


 Summary: Did not find value which can be converted into 
java.lang.String
 Key: SPARK-37435
 URL: https://issues.apache.org/jira/browse/SPARK-37435
 Project: Spark
  Issue Type: Bug
  Components: PySpark
Affects Versions: 3.0.2, 2.4.4
Reporter: Ziqun Ye


Got this following error when loading the saved model.

```

ERROR:ADS Exception Traceback (most recent call last): File 
"/home/datascience/conda/pyspark30_p37_cpu_v2/lib/python3.7/site-packages/IPython/core/interactiveshell.py",
 line 3441, in run_code exec(code_obj, self.user_global_ns, self.user_ns) File 
"/tmp/ipykernel_12307/1140552986.py", line 15, in  
LogisticRegressionModel.load(spark, "./lrmodelv2") File 
"/home/datascience/conda/pyspark30_p37_cpu_v2/lib/python3.7/site-packages/pyspark/mllib/classification.py",
 line 249, in load sc._jsc.sc(), path) File 
"/home/datascience/conda/pyspark30_p37_cpu_v2/lib/python3.7/site-packages/py4j/java_gateway.py",
 line 1305, in __call__ answer, self.gateway_client, self.target_id, self.name) 
File 
"/home/datascience/conda/pyspark30_p37_cpu_v2/lib/python3.7/site-packages/pyspark/sql/utils.py",
 line 128, in deco return f(*a, **kw) File 
"/home/datascience/conda/pyspark30_p37_cpu_v2/lib/python3.7/site-packages/py4j/protocol.py",
 line 328, in get_return_value format(target_id, ".", name), value) 
py4j.protocol.Py4JJavaError: An error occurred while calling 
z:org.apache.spark.mllib.classification.LogisticRegressionModel.load. : 
org.json4s.package$MappingException: Did not find value which can be converted 
into java.lang.String at org.json4s.reflect.package$.fail(package.scala:95) at 
org.json4s.Extraction$.$anonfun$convert$2(Extraction.scala:756) at 
scala.Option.getOrElse(Option.scala:189) at 
org.json4s.Extraction$.convert(Extraction.scala:756) at 
org.json4s.Extraction$.$anonfun$extract$10(Extraction.scala:404) at 
org.json4s.Extraction$.$anonfun$customOrElse$1(Extraction.scala:658) at 
scala.PartialFunction.applyOrElse(PartialFunction.scala:127) at 
scala.PartialFunction.applyOrElse$(PartialFunction.scala:126) at 
scala.PartialFunction$$anon$1.applyOrElse(PartialFunction.scala:257) at 
org.json4s.Extraction$.customOrElse(Extraction.scala:658) at 
org.json4s.Extraction$.extract(Extraction.scala:402) at 
org.json4s.Extraction$.extract(Extraction.scala:40) at 
org.json4s.ExtractableJsonAstNode.extract(ExtractableJsonAstNode.scala:21) at 
org.apache.spark.mllib.util.Loader$.loadMetadata(modelSaveLoad.scala:122) at 
org.apache.spark.mllib.classification.LogisticRegressionModel$.load(LogisticRegression.scala:176)
 at 
org.apache.spark.mllib.classification.LogisticRegressionModel.load(LogisticRegression.scala)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:498) at 
py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244) at 
py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357) at 
py4j.Gateway.invoke(Gateway.java:282) at 
py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132) at 
py4j.commands.CallCommand.execute(CallCommand.java:79) at 
py4j.GatewayConnection.run(GatewayConnection.java:238) at 
java.lang.Thread.run(Thread.java:748) Py4JJavaError: An error occurred while 
calling z:org.apache.spark.mllib.classification.LogisticRegressionModel.load. : 
org.json4s.package$MappingException: Did not find value which can be converted 
into java.lang.String at org.json4s.reflect.package$.fail(package.scala:95) at 
org.json4s.Extraction$.$anonfun$convert$2(Extraction.scala:756) at 
scala.Option.getOrElse(Option.scala:189) at 
org.json4s.Extraction$.convert(Extraction.scala:756) at 
org.json4s.Extraction$.$anonfun$extract$10(Extraction.scala:404) at 
org.json4s.Extraction$.$anonfun$customOrElse$1(Extraction.scala:658) at 
scala.PartialFunction.applyOrElse(PartialFunction.scala:127) at 
scala.PartialFunction.applyOrElse$(PartialFunction.scala:126) at 
scala.PartialFunction$$anon$1.applyOrElse(PartialFunction.scala:257) at 
org.json4s.Extraction$.customOrElse(Extraction.scala:658) at 
org.json4s.Extraction$.extract(Extraction.scala:402) at 
org.json4s.Extraction$.extract(Extraction.scala:40) at 
org.json4s.ExtractableJsonAstNode.extract(ExtractableJsonAstNode.scala:21) at 
org.apache.spark.mllib.util.Loader$.loadMetadata(modelSaveLoad.scala:122) at 
org.apache.spark.mllib.classification.LogisticRegressionModel$.load(LogisticRegression.scala:176)
 at 
org.apache.spark.mllib.classification.LogisticRegressionModel.load(LogisticRegression.scala)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
at 

[jira] [Assigned] (SPARK-37434) Add a new profile to auto disable unsupported UTs on Macos using Apple Silicon

2021-11-21 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37434?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-37434:


Assignee: (was: Apache Spark)

> Add a new profile to auto disable unsupported UTs on Macos using Apple Silicon
> --
>
> Key: SPARK-37434
> URL: https://issues.apache.org/jira/browse/SPARK-37434
> Project: Spark
>  Issue Type: Improvement
>  Components: Build
>Affects Versions: 3.3.0
>Reporter: Yang Jie
>Priority: Major
>
> After SPARK-37272  and SPARK-37282,  we can manually add
> {code:java}
> -Dtest.exclude.tags=org.apache.spark.tags.ExtendedLevelDBTest,org.apache.spark.tags.ExtendedRocksDBTest
>  {code}
> when run mvn test or sbt test to disable unsupported UTs on Macos using Apple 
> Silicon.
>  
> We can add a profile to and  activate this property automatically when run 
> UTs on Macos using Apple Silicon.
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-37434) Add a new profile to auto disable unsupported UTs on Macos using Apple Silicon

2021-11-21 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37434?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-37434:


Assignee: Apache Spark

> Add a new profile to auto disable unsupported UTs on Macos using Apple Silicon
> --
>
> Key: SPARK-37434
> URL: https://issues.apache.org/jira/browse/SPARK-37434
> Project: Spark
>  Issue Type: Improvement
>  Components: Build
>Affects Versions: 3.3.0
>Reporter: Yang Jie
>Assignee: Apache Spark
>Priority: Major
>
> After SPARK-37272  and SPARK-37282,  we can manually add
> {code:java}
> -Dtest.exclude.tags=org.apache.spark.tags.ExtendedLevelDBTest,org.apache.spark.tags.ExtendedRocksDBTest
>  {code}
> when run mvn test or sbt test to disable unsupported UTs on Macos using Apple 
> Silicon.
>  
> We can add a profile to and  activate this property automatically when run 
> UTs on Macos using Apple Silicon.
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-37434) Add a new profile to auto disable unsupported UTs on Macos using Apple Silicon

2021-11-21 Thread Apache Spark (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-37434?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17447185#comment-17447185
 ] 

Apache Spark commented on SPARK-37434:
--

User 'LuciferYang' has created a pull request for this issue:
https://github.com/apache/spark/pull/34676

> Add a new profile to auto disable unsupported UTs on Macos using Apple Silicon
> --
>
> Key: SPARK-37434
> URL: https://issues.apache.org/jira/browse/SPARK-37434
> Project: Spark
>  Issue Type: Improvement
>  Components: Build
>Affects Versions: 3.3.0
>Reporter: Yang Jie
>Priority: Major
>
> After SPARK-37272  and SPARK-37282,  we can manually add
> {code:java}
> -Dtest.exclude.tags=org.apache.spark.tags.ExtendedLevelDBTest,org.apache.spark.tags.ExtendedRocksDBTest
>  {code}
> when run mvn test or sbt test to disable unsupported UTs on Macos using Apple 
> Silicon.
>  
> We can add a profile to and  activate this property automatically when run 
> UTs on Macos using Apple Silicon.
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-37434) Add a new profile to auto disable unsupported UTs on Macos using Apple Silicon

2021-11-21 Thread Yang Jie (Jira)
Yang Jie created SPARK-37434:


 Summary: Add a new profile to auto disable unsupported UTs on 
Macos using Apple Silicon
 Key: SPARK-37434
 URL: https://issues.apache.org/jira/browse/SPARK-37434
 Project: Spark
  Issue Type: Improvement
  Components: Build
Affects Versions: 3.3.0
Reporter: Yang Jie


After SPARK-37272  and SPARK-37282,  we can manually add
{code:java}
-Dtest.exclude.tags=org.apache.spark.tags.ExtendedLevelDBTest,org.apache.spark.tags.ExtendedRocksDBTest
 {code}
when run mvn test or sbt test to disable unsupported UTs on Macos using Apple 
Silicon.

 

We can add a profile to and  activate this property automatically when run UTs 
on Macos using Apple Silicon.

 

 

 

 



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-37354) Make the Java version installed on the container image used by the K8s integration tests with SBT configurable

2021-11-21 Thread Kousuke Saruta (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37354?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kousuke Saruta resolved SPARK-37354.

Fix Version/s: 3.3.0
   Resolution: Fixed

Issue resolved in https://github.com/apache/spark/pull/34628

> Make the Java version installed on the container image used by the K8s 
> integration tests with SBT configurable
> --
>
> Key: SPARK-37354
> URL: https://issues.apache.org/jira/browse/SPARK-37354
> Project: Spark
>  Issue Type: Bug
>  Components: Kubernetes, Tests
>Affects Versions: 3.2.0
>Reporter: Kousuke Saruta
>Assignee: Kousuke Saruta
>Priority: Major
> Fix For: 3.3.0
>
>
> I noticed that the default Java version installed on the container image used 
> by the K8s integration tests are different depending on the way to run the 
> tests.
> If the tests are launched by Maven, the Java version is 8 is installed.
> On the other hand, if the tests are launched by SBT, the Java version is 11.
> Further, we have no way to change the version.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-36231) Support arithmetic operations of Series containing Decimal(np.nan)

2021-11-21 Thread Hyukjin Kwon (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-36231?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hyukjin Kwon updated SPARK-36231:
-
Fix Version/s: (was: 3.3.0)

> Support arithmetic operations of Series containing Decimal(np.nan) 
> ---
>
> Key: SPARK-36231
> URL: https://issues.apache.org/jira/browse/SPARK-36231
> Project: Spark
>  Issue Type: Sub-task
>  Components: PySpark
>Affects Versions: 3.2.0
>Reporter: Xinrong Meng
>Assignee: Yikun Jiang
>Priority: Major
>
> Arithmetic operations of Series containing Decimal(np.nan) raise 
> java.lang.NullPointerException in driver. An example is shown as below:
> {code:java}
> >>> pser = pd.Series([decimal.Decimal(1.0), decimal.Decimal(2.0), 
> >>> decimal.Decimal(np.nan)])
> >>> psser = ps.from_pandas(pser)
> >>> pser + 1
> 0 2
>  1 3
>  2 NaN
> >>> psser + 1
>  Driver stacktrace:
>  at 
> org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2259)
>  at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2208)
>  at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2207)
>  at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
>  at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
>  at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
>  at 
> org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2207)
>  at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1084)
>  at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1084)
>  at scala.Option.foreach(Option.scala:407)
>  at 
> org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1084)
>  at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2446)
>  at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2388)
>  at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2377)
>  at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49)
>  at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:873)
>  at org.apache.spark.SparkContext.runJob(SparkContext.scala:2208)
>  at org.apache.spark.SparkContext.runJob(SparkContext.scala:2303)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$5(Dataset.scala:3648)
>  at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
>  at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1437)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$2(Dataset.scala:3652)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$2$adapted(Dataset.scala:3629)
>  at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3706)
>  at 
> org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:103)
>  at 
> org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:163)
>  at 
> org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:90)
>  at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:774)
>  at 
> org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
>  at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3704)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$1(Dataset.scala:3629)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$1$adapted(Dataset.scala:3628)
>  at 
> org.apache.spark.security.SocketAuthServer$.$anonfun$serveToStream$2(SocketAuthServer.scala:139)
>  at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
>  at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1437)
>  at 
> org.apache.spark.security.SocketAuthServer$.$anonfun$serveToStream$1(SocketAuthServer.scala:141)
>  at 
> org.apache.spark.security.SocketAuthServer$.$anonfun$serveToStream$1$adapted(SocketAuthServer.scala:136)
>  at 
> org.apache.spark.security.SocketFuncServer.handleConnection(SocketAuthServer.scala:113)
>  at 
> org.apache.spark.security.SocketFuncServer.handleConnection(SocketAuthServer.scala:107)
>  at 
> org.apache.spark.security.SocketAuthServer$$anon$1.$anonfun$run$4(SocketAuthServer.scala:68)
>  at scala.util.Try$.apply(Try.scala:213)
>  at 
> org.apache.spark.security.SocketAuthServer$$anon$1.run(SocketAuthServer.scala:68)
>  Caused by: java.lang.NullPointerException
>  at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage1.processNext(Unknown
>  Source)
>  at 
> 

[jira] [Reopened] (SPARK-36231) Support arithmetic operations of Series containing Decimal(np.nan)

2021-11-21 Thread Hyukjin Kwon (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-36231?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hyukjin Kwon reopened SPARK-36231:
--

Reverted at 
https://github.com/apache/spark/commit/406455d79f787486f9e6fab1dce0d9a2645b8d14

> Support arithmetic operations of Series containing Decimal(np.nan) 
> ---
>
> Key: SPARK-36231
> URL: https://issues.apache.org/jira/browse/SPARK-36231
> Project: Spark
>  Issue Type: Sub-task
>  Components: PySpark
>Affects Versions: 3.2.0
>Reporter: Xinrong Meng
>Assignee: Yikun Jiang
>Priority: Major
> Fix For: 3.3.0
>
>
> Arithmetic operations of Series containing Decimal(np.nan) raise 
> java.lang.NullPointerException in driver. An example is shown as below:
> {code:java}
> >>> pser = pd.Series([decimal.Decimal(1.0), decimal.Decimal(2.0), 
> >>> decimal.Decimal(np.nan)])
> >>> psser = ps.from_pandas(pser)
> >>> pser + 1
> 0 2
>  1 3
>  2 NaN
> >>> psser + 1
>  Driver stacktrace:
>  at 
> org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2259)
>  at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2208)
>  at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2207)
>  at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
>  at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
>  at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
>  at 
> org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2207)
>  at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1084)
>  at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1084)
>  at scala.Option.foreach(Option.scala:407)
>  at 
> org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1084)
>  at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2446)
>  at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2388)
>  at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2377)
>  at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49)
>  at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:873)
>  at org.apache.spark.SparkContext.runJob(SparkContext.scala:2208)
>  at org.apache.spark.SparkContext.runJob(SparkContext.scala:2303)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$5(Dataset.scala:3648)
>  at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
>  at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1437)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$2(Dataset.scala:3652)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$2$adapted(Dataset.scala:3629)
>  at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3706)
>  at 
> org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:103)
>  at 
> org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:163)
>  at 
> org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:90)
>  at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:774)
>  at 
> org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
>  at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3704)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$1(Dataset.scala:3629)
>  at 
> org.apache.spark.sql.Dataset.$anonfun$collectAsArrowToPython$1$adapted(Dataset.scala:3628)
>  at 
> org.apache.spark.security.SocketAuthServer$.$anonfun$serveToStream$2(SocketAuthServer.scala:139)
>  at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
>  at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1437)
>  at 
> org.apache.spark.security.SocketAuthServer$.$anonfun$serveToStream$1(SocketAuthServer.scala:141)
>  at 
> org.apache.spark.security.SocketAuthServer$.$anonfun$serveToStream$1$adapted(SocketAuthServer.scala:136)
>  at 
> org.apache.spark.security.SocketFuncServer.handleConnection(SocketAuthServer.scala:113)
>  at 
> org.apache.spark.security.SocketFuncServer.handleConnection(SocketAuthServer.scala:107)
>  at 
> org.apache.spark.security.SocketAuthServer$$anon$1.$anonfun$run$4(SocketAuthServer.scala:68)
>  at scala.util.Try$.apply(Try.scala:213)
>  at 
> org.apache.spark.security.SocketAuthServer$$anon$1.run(SocketAuthServer.scala:68)
>  Caused by: java.lang.NullPointerException
>  at 
> 

[jira] [Resolved] (SPARK-37209) YarnShuffleIntegrationSuite and other two similar cases in `resource-managers` test failed

2021-11-21 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37209?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-37209.
--
Fix Version/s: 3.3.0
   3.0.4
   3.2.1
   3.1.3
   Resolution: Fixed

Issue resolved by pull request 34620
[https://github.com/apache/spark/pull/34620]

> YarnShuffleIntegrationSuite  and other two similar cases in 
> `resource-managers` test failed
> ---
>
> Key: SPARK-37209
> URL: https://issues.apache.org/jira/browse/SPARK-37209
> Project: Spark
>  Issue Type: Bug
>  Components: Tests, YARN
>Affects Versions: 3.3.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Minor
> Fix For: 3.3.0, 3.0.4, 3.2.1, 3.1.3
>
> Attachments: failed-unit-tests.log, success-unit-tests.log
>
>
> Execute :
>  # build/mvn clean package -DskipTests -Phadoop-3.2 -Phive-2.3 -Phadoop-cloud 
> -Pmesos -Pyarn -Pkinesis-asl -Phive-thriftserver -Pspark-ganglia-lgpl 
> -Pkubernetes -Phive
>  # build/mvn test -Phadoop-3.2 -Phive-2.3 -Phadoop-cloud -Pmesos -Pyarn 
> -Pkinesis-asl -Phive-thriftserver -Pspark-ganglia-lgpl -Pkubernetes -Phive 
> -Pscala-2.13 -pl resource-managers/yarn
> The test will successful.
>  
> Execute :
>  # build/mvn clean -Phadoop-3.2 -Phive-2.3 -Phadoop-cloud -Pmesos -Pyarn 
> -Pkinesis-asl -Phive-thriftserver -Pspark-ganglia-lgpl -Pkubernetes -Phive
>  # build/mvn clean test -Phadoop-3.2 -Phive-2.3 -Phadoop-cloud -Pmesos -Pyarn 
> -Pkinesis-asl -Phive-thriftserver -Pspark-ganglia-lgpl -Pkubernetes -Phive 
> -Pscala-2.13 -pl resource-managers/yarn 
> The test will failed.
>  
> Execute :
>  # build/mvn clean package -DskipTests -Phadoop-3.2 -Phive-2.3 -Phadoop-cloud 
> -Pmesos -Pyarn -Pkinesis-asl -Phive-thriftserver -Pspark-ganglia-lgpl 
> -Pkubernetes -Phive
>  # Delete assembly/target/scala-2.12/jars manually
>  # build/mvn test -Phadoop-3.2 -Phive-2.3 -Phadoop-cloud -Pmesos -Pyarn 
> -Pkinesis-asl -Phive-thriftserver -Pspark-ganglia-lgpl -Pkubernetes -Phive 
> -Pscala-2.13 -pl resource-managers/yarn 
> The test will failed.
>  
> The error stack is :
> {code:java}
> 21/11/04 19:48:52.159 main ERROR Client: Application diagnostics message: 
> User class threw exception: org.apache.spark.SparkException: Job aborted due 
> to stage failure: Task 0 in stage 0.0 failed 4 times,
>  most recent failure: Lost task 0.3 in stage 0.0 (TID 6) (localhost executor 
> 1): java.lang.NoClassDefFoundError: breeze/linalg/Matrix
> at java.lang.Class.forName0(Native Method)
> at java.lang.Class.forName(Class.java:348)
> at org.apache.spark.util.Utils$.classForName(Utils.scala:216)
> at 
> org.apache.spark.serializer.KryoSerializer$.$anonfun$loadableSparkClasses$1(KryoSerializer.scala:537)
> at scala.collection.immutable.List.flatMap(List.scala:293)
> at scala.collection.immutable.List.flatMap(List.scala:79)
> at 
> org.apache.spark.serializer.KryoSerializer$.loadableSparkClasses$lzycompute(KryoSerializer.scala:535)
> at 
> org.apache.spark.serializer.KryoSerializer$.org$apache$spark$serializer$KryoSerializer$$loadableSparkClasses(KryoSerializer.scala:502)
> at 
> org.apache.spark.serializer.KryoSerializer.newKryo(KryoSerializer.scala:226)
> at 
> org.apache.spark.serializer.KryoSerializer$$anon$1.create(KryoSerializer.scala:102)
> at 
> com.esotericsoftware.kryo.pool.KryoPoolQueueImpl.borrow(KryoPoolQueueImpl.java:48)
> at 
> org.apache.spark.serializer.KryoSerializer$PoolWrapper.borrow(KryoSerializer.scala:109)
> at 
> org.apache.spark.serializer.KryoSerializerInstance.borrowKryo(KryoSerializer.scala:346)
> at 
> org.apache.spark.serializer.KryoSerializationStream.(KryoSerializer.scala:266)
> at 
> org.apache.spark.serializer.KryoSerializerInstance.serializeStream(KryoSerializer.scala:432)
> at 
> org.apache.spark.shuffle.ShufflePartitionPairsWriter.open(ShufflePartitionPairsWriter.scala:76)
> at 
> org.apache.spark.shuffle.ShufflePartitionPairsWriter.write(ShufflePartitionPairsWriter.scala:59)
> at 
> org.apache.spark.util.collection.WritablePartitionedIterator.writeNext(WritablePartitionedPairCollection.scala:83)
> at 
> org.apache.spark.util.collection.ExternalSorter.$anonfun$writePartitionedMapOutput$1(ExternalSorter.scala:772)
> at 
> scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.scala:18)
> at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1468)
> at 
> org.apache.spark.util.collection.ExternalSorter.writePartitionedMapOutput(ExternalSorter.scala:775)
> at 
> org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:70)
> at 
> 

[jira] [Assigned] (SPARK-37209) YarnShuffleIntegrationSuite and other two similar cases in `resource-managers` test failed

2021-11-21 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37209?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-37209:


Assignee: Yang Jie

> YarnShuffleIntegrationSuite  and other two similar cases in 
> `resource-managers` test failed
> ---
>
> Key: SPARK-37209
> URL: https://issues.apache.org/jira/browse/SPARK-37209
> Project: Spark
>  Issue Type: Bug
>  Components: Tests, YARN
>Affects Versions: 3.3.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Minor
> Attachments: failed-unit-tests.log, success-unit-tests.log
>
>
> Execute :
>  # build/mvn clean package -DskipTests -Phadoop-3.2 -Phive-2.3 -Phadoop-cloud 
> -Pmesos -Pyarn -Pkinesis-asl -Phive-thriftserver -Pspark-ganglia-lgpl 
> -Pkubernetes -Phive
>  # build/mvn test -Phadoop-3.2 -Phive-2.3 -Phadoop-cloud -Pmesos -Pyarn 
> -Pkinesis-asl -Phive-thriftserver -Pspark-ganglia-lgpl -Pkubernetes -Phive 
> -Pscala-2.13 -pl resource-managers/yarn
> The test will successful.
>  
> Execute :
>  # build/mvn clean -Phadoop-3.2 -Phive-2.3 -Phadoop-cloud -Pmesos -Pyarn 
> -Pkinesis-asl -Phive-thriftserver -Pspark-ganglia-lgpl -Pkubernetes -Phive
>  # build/mvn clean test -Phadoop-3.2 -Phive-2.3 -Phadoop-cloud -Pmesos -Pyarn 
> -Pkinesis-asl -Phive-thriftserver -Pspark-ganglia-lgpl -Pkubernetes -Phive 
> -Pscala-2.13 -pl resource-managers/yarn 
> The test will failed.
>  
> Execute :
>  # build/mvn clean package -DskipTests -Phadoop-3.2 -Phive-2.3 -Phadoop-cloud 
> -Pmesos -Pyarn -Pkinesis-asl -Phive-thriftserver -Pspark-ganglia-lgpl 
> -Pkubernetes -Phive
>  # Delete assembly/target/scala-2.12/jars manually
>  # build/mvn test -Phadoop-3.2 -Phive-2.3 -Phadoop-cloud -Pmesos -Pyarn 
> -Pkinesis-asl -Phive-thriftserver -Pspark-ganglia-lgpl -Pkubernetes -Phive 
> -Pscala-2.13 -pl resource-managers/yarn 
> The test will failed.
>  
> The error stack is :
> {code:java}
> 21/11/04 19:48:52.159 main ERROR Client: Application diagnostics message: 
> User class threw exception: org.apache.spark.SparkException: Job aborted due 
> to stage failure: Task 0 in stage 0.0 failed 4 times,
>  most recent failure: Lost task 0.3 in stage 0.0 (TID 6) (localhost executor 
> 1): java.lang.NoClassDefFoundError: breeze/linalg/Matrix
> at java.lang.Class.forName0(Native Method)
> at java.lang.Class.forName(Class.java:348)
> at org.apache.spark.util.Utils$.classForName(Utils.scala:216)
> at 
> org.apache.spark.serializer.KryoSerializer$.$anonfun$loadableSparkClasses$1(KryoSerializer.scala:537)
> at scala.collection.immutable.List.flatMap(List.scala:293)
> at scala.collection.immutable.List.flatMap(List.scala:79)
> at 
> org.apache.spark.serializer.KryoSerializer$.loadableSparkClasses$lzycompute(KryoSerializer.scala:535)
> at 
> org.apache.spark.serializer.KryoSerializer$.org$apache$spark$serializer$KryoSerializer$$loadableSparkClasses(KryoSerializer.scala:502)
> at 
> org.apache.spark.serializer.KryoSerializer.newKryo(KryoSerializer.scala:226)
> at 
> org.apache.spark.serializer.KryoSerializer$$anon$1.create(KryoSerializer.scala:102)
> at 
> com.esotericsoftware.kryo.pool.KryoPoolQueueImpl.borrow(KryoPoolQueueImpl.java:48)
> at 
> org.apache.spark.serializer.KryoSerializer$PoolWrapper.borrow(KryoSerializer.scala:109)
> at 
> org.apache.spark.serializer.KryoSerializerInstance.borrowKryo(KryoSerializer.scala:346)
> at 
> org.apache.spark.serializer.KryoSerializationStream.(KryoSerializer.scala:266)
> at 
> org.apache.spark.serializer.KryoSerializerInstance.serializeStream(KryoSerializer.scala:432)
> at 
> org.apache.spark.shuffle.ShufflePartitionPairsWriter.open(ShufflePartitionPairsWriter.scala:76)
> at 
> org.apache.spark.shuffle.ShufflePartitionPairsWriter.write(ShufflePartitionPairsWriter.scala:59)
> at 
> org.apache.spark.util.collection.WritablePartitionedIterator.writeNext(WritablePartitionedPairCollection.scala:83)
> at 
> org.apache.spark.util.collection.ExternalSorter.$anonfun$writePartitionedMapOutput$1(ExternalSorter.scala:772)
> at 
> scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.scala:18)
> at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1468)
> at 
> org.apache.spark.util.collection.ExternalSorter.writePartitionedMapOutput(ExternalSorter.scala:775)
> at 
> org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:70)
> at 
> org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
> at 
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:99)
> at 
> 

[jira] [Commented] (SPARK-37391) SIGNIFICANT bottleneck introduced by fix for SPARK-34497

2021-11-21 Thread Hyukjin Kwon (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-37391?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17447162#comment-17447162
 ] 

Hyukjin Kwon commented on SPARK-37391:
--

[~danny-seismic], it would be great to assess this issue futher with problem 
description and preferably self-contained reproducer

> SIGNIFICANT bottleneck introduced by fix for SPARK-34497
> 
>
> Key: SPARK-37391
> URL: https://issues.apache.org/jira/browse/SPARK-37391
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 3.1.0, 3.1.1, 3.1.2, 3.2.0
> Environment: N/A
>Reporter: Danny Guinther
>Priority: Major
>
> The fix for SPARK-34497 ( [https://github.com/apache/spark/pull/31622] ) does 
> not seem to have consider the reality that some apps may rely on being able 
> to establish many JDBC connections simultaneously for performance reasons.
> The fix forces concurrency to 1 when establishing database connections and 
> that strikes me as a *significant* user impacting change and a *significant* 
> bottleneck.
> Can anyone propose a workaround for this? I have an app that makes 
> connections to thousands of databases and I can't upgrade to any version 
> >3.1.x because of this significant bottleneck.
>  
> Thanks in advance for your help!



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-37433) TimeZoneAwareExpression throws NoSuchElementException: None.get on expr.eval()

2021-11-21 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37433?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-37433:


Assignee: (was: Apache Spark)

> TimeZoneAwareExpression throws NoSuchElementException: None.get on expr.eval()
> --
>
> Key: SPARK-37433
> URL: https://issues.apache.org/jira/browse/SPARK-37433
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 3.2.0
>Reporter: Sathiya Kumar
>Priority: Minor
>
> TimeZoneAwareExpression like hour, date_format etc. throws 
> NoSuchElementException: None.get on expr.eval()
> *hour(current_timestamp).expr.eval()*
> *date_format(current_timestamp, "dd").expr.eval()*
>  
> {code:java}
> java.util.NoSuchElementException: None.get
>   at scala.None$.get(Option.scala:529)
>   at scala.None$.get(Option.scala:527)
>   at 
> org.apache.spark.sql.catalyst.expressions.TimeZoneAwareExpression.zoneId(datetimeExpressions.scala:53)
>   at 
> org.apache.spark.sql.catalyst.expressions.TimeZoneAwareExpression.zoneId$(datetimeExpressions.scala:53)
>   at 
> org.apache.spark.sql.catalyst.expressions.DateFormatClass.zoneId$lzycompute(datetimeExpressions.scala:772)
>   at 
> org.apache.spark.sql.catalyst.expressions.DateFormatClass.zoneId(datetimeExpressions.scala:772)
>   at 
> org.apache.spark.sql.catalyst.expressions.TimestampFormatterHelper.getFormatter(datetimeExpressions.scala:70)
>   at 
> org.apache.spark.sql.catalyst.expressions.TimestampFormatterHelper.getFormatter$(datetimeExpressions.scala:67)
>   at 
> org.apache.spark.sql.catalyst.expressions.DateFormatClass.getFormatter(datetimeExpressions.scala:772)
>   at 
> org.apache.spark.sql.catalyst.expressions.TimestampFormatterHelper.$anonfun$formatterOption$1(datetimeExpressions.scala:64)
> {code}
>  



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-37433) TimeZoneAwareExpression throws NoSuchElementException: None.get on expr.eval()

2021-11-21 Thread Apache Spark (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-37433?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17447159#comment-17447159
 ] 

Apache Spark commented on SPARK-37433:
--

User 'sathiyapk' has created a pull request for this issue:
https://github.com/apache/spark/pull/34675

> TimeZoneAwareExpression throws NoSuchElementException: None.get on expr.eval()
> --
>
> Key: SPARK-37433
> URL: https://issues.apache.org/jira/browse/SPARK-37433
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 3.2.0
>Reporter: Sathiya Kumar
>Priority: Minor
>
> TimeZoneAwareExpression like hour, date_format etc. throws 
> NoSuchElementException: None.get on expr.eval()
> *hour(current_timestamp).expr.eval()*
> *date_format(current_timestamp, "dd").expr.eval()*
>  
> {code:java}
> java.util.NoSuchElementException: None.get
>   at scala.None$.get(Option.scala:529)
>   at scala.None$.get(Option.scala:527)
>   at 
> org.apache.spark.sql.catalyst.expressions.TimeZoneAwareExpression.zoneId(datetimeExpressions.scala:53)
>   at 
> org.apache.spark.sql.catalyst.expressions.TimeZoneAwareExpression.zoneId$(datetimeExpressions.scala:53)
>   at 
> org.apache.spark.sql.catalyst.expressions.DateFormatClass.zoneId$lzycompute(datetimeExpressions.scala:772)
>   at 
> org.apache.spark.sql.catalyst.expressions.DateFormatClass.zoneId(datetimeExpressions.scala:772)
>   at 
> org.apache.spark.sql.catalyst.expressions.TimestampFormatterHelper.getFormatter(datetimeExpressions.scala:70)
>   at 
> org.apache.spark.sql.catalyst.expressions.TimestampFormatterHelper.getFormatter$(datetimeExpressions.scala:67)
>   at 
> org.apache.spark.sql.catalyst.expressions.DateFormatClass.getFormatter(datetimeExpressions.scala:772)
>   at 
> org.apache.spark.sql.catalyst.expressions.TimestampFormatterHelper.$anonfun$formatterOption$1(datetimeExpressions.scala:64)
> {code}
>  



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-37433) TimeZoneAwareExpression throws NoSuchElementException: None.get on expr.eval()

2021-11-21 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37433?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-37433:


Assignee: Apache Spark

> TimeZoneAwareExpression throws NoSuchElementException: None.get on expr.eval()
> --
>
> Key: SPARK-37433
> URL: https://issues.apache.org/jira/browse/SPARK-37433
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 3.2.0
>Reporter: Sathiya Kumar
>Assignee: Apache Spark
>Priority: Minor
>
> TimeZoneAwareExpression like hour, date_format etc. throws 
> NoSuchElementException: None.get on expr.eval()
> *hour(current_timestamp).expr.eval()*
> *date_format(current_timestamp, "dd").expr.eval()*
>  
> {code:java}
> java.util.NoSuchElementException: None.get
>   at scala.None$.get(Option.scala:529)
>   at scala.None$.get(Option.scala:527)
>   at 
> org.apache.spark.sql.catalyst.expressions.TimeZoneAwareExpression.zoneId(datetimeExpressions.scala:53)
>   at 
> org.apache.spark.sql.catalyst.expressions.TimeZoneAwareExpression.zoneId$(datetimeExpressions.scala:53)
>   at 
> org.apache.spark.sql.catalyst.expressions.DateFormatClass.zoneId$lzycompute(datetimeExpressions.scala:772)
>   at 
> org.apache.spark.sql.catalyst.expressions.DateFormatClass.zoneId(datetimeExpressions.scala:772)
>   at 
> org.apache.spark.sql.catalyst.expressions.TimestampFormatterHelper.getFormatter(datetimeExpressions.scala:70)
>   at 
> org.apache.spark.sql.catalyst.expressions.TimestampFormatterHelper.getFormatter$(datetimeExpressions.scala:67)
>   at 
> org.apache.spark.sql.catalyst.expressions.DateFormatClass.getFormatter(datetimeExpressions.scala:772)
>   at 
> org.apache.spark.sql.catalyst.expressions.TimestampFormatterHelper.$anonfun$formatterOption$1(datetimeExpressions.scala:64)
> {code}
>  



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-37433) TimeZoneAwareExpression throws NoSuchElementException: None.get on expr.eval()

2021-11-21 Thread Sathiya Kumar (Jira)
Sathiya Kumar created SPARK-37433:
-

 Summary: TimeZoneAwareExpression throws NoSuchElementException: 
None.get on expr.eval()
 Key: SPARK-37433
 URL: https://issues.apache.org/jira/browse/SPARK-37433
 Project: Spark
  Issue Type: Bug
  Components: SQL
Affects Versions: 3.2.0
Reporter: Sathiya Kumar


TimeZoneAwareExpression like hour, date_format etc. throws 

NoSuchElementException: None.get on expr.eval()

*hour(current_timestamp).expr.eval()*

*date_format(current_timestamp, "dd").expr.eval()*

 
{code:java}
java.util.NoSuchElementException: None.get
  at scala.None$.get(Option.scala:529)
  at scala.None$.get(Option.scala:527)
  at 
org.apache.spark.sql.catalyst.expressions.TimeZoneAwareExpression.zoneId(datetimeExpressions.scala:53)
  at 
org.apache.spark.sql.catalyst.expressions.TimeZoneAwareExpression.zoneId$(datetimeExpressions.scala:53)
  at 
org.apache.spark.sql.catalyst.expressions.DateFormatClass.zoneId$lzycompute(datetimeExpressions.scala:772)
  at 
org.apache.spark.sql.catalyst.expressions.DateFormatClass.zoneId(datetimeExpressions.scala:772)
  at 
org.apache.spark.sql.catalyst.expressions.TimestampFormatterHelper.getFormatter(datetimeExpressions.scala:70)
  at 
org.apache.spark.sql.catalyst.expressions.TimestampFormatterHelper.getFormatter$(datetimeExpressions.scala:67)
  at 
org.apache.spark.sql.catalyst.expressions.DateFormatClass.getFormatter(datetimeExpressions.scala:772)
  at 
org.apache.spark.sql.catalyst.expressions.TimestampFormatterHelper.$anonfun$formatterOption$1(datetimeExpressions.scala:64)
{code}
 



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37432) Driver keep a record of decommission executor

2021-11-21 Thread Hoa Le (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37432?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoa Le updated SPARK-37432:
---
Attachment: master_ui_executor_tab.png

> Driver keep a record of decommission executor
> -
>
> Key: SPARK-37432
> URL: https://issues.apache.org/jira/browse/SPARK-37432
> Project: Spark
>  Issue Type: Bug
>  Components: Kubernetes
>Affects Versions: 3.1.1
>Reporter: Hoa Le
>Priority: Minor
> Attachments: master_ui_executor_tab.png
>
>
> Hi,
> We are running spark on k8s with version 3.1.1. After the spark application 
> running for a while, we are getting the exception below:
> On driver: 
>  
> {code:java}
> 2021-11-21 18:25:21,859 ERROR Failed to send RPC RPC 6827167497981418905 to 
> /10.1.201.113:58354: java.nio.channels.ClosedChannelException 
> (org.apache.spark.network.client.TransportClient) [rpc-server-4-1]
> java.nio.channels.ClosedChannelException
>   at 
> io.netty.channel.AbstractChannel$AbstractUnsafe.newClosedChannelException(AbstractChannel.java:957)
>   at 
> io.netty.channel.AbstractChannel$AbstractUnsafe.write(AbstractChannel.java:865)
>   at 
> io.netty.channel.DefaultChannelPipeline$HeadContext.write(DefaultChannelPipeline.java:1367)
>   at 
> io.netty.channel.AbstractChannelHandlerContext.invokeWrite0(AbstractChannelHandlerContext.java:717)
>   at 
> io.netty.channel.AbstractChannelHandlerContext.invokeWriteAndFlush(AbstractChannelHandlerContext.java:764)
>   at 
> io.netty.channel.AbstractChannelHandlerContext$WriteTask.run(AbstractChannelHandlerContext.java:1071)
>   at 
> io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:164)
>   at 
> io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:472)
>   at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:500)
>   at 
> io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)
>   at 
> io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
>   at 
> io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
>   at java.base/java.lang.Thread.run(Unknown Source)
> 2021-11-21 18:25:21,864 ERROR Failed to send RPC RPC 7618635518207296341 to 
> /10.1.201.113:58354: java.nio.channels.ClosedChannelException 
> (org.apache.spark.network.client.TransportClient) [rpc-server-4-1]
> java.nio.channels.ClosedChannelException
>   at 
> io.netty.channel.AbstractChannel$AbstractUnsafe.newClosedChannelException(AbstractChannel.java:957)
>   at 
> io.netty.channel.AbstractChannel$AbstractUnsafe.write(AbstractChannel.java:865)
>   at 
> io.netty.channel.DefaultChannelPipeline$HeadContext.write(DefaultChannelPipeline.java:1367)
>   at 
> io.netty.channel.AbstractChannelHandlerContext.invokeWrite0(AbstractChannelHandlerContext.java:717)
>   at 
> io.netty.channel.AbstractChannelHandlerContext.invokeWriteAndFlush(AbstractChannelHandlerContext.java:764)
>   at 
> io.netty.channel.AbstractChannelHandlerContext$WriteTask.run(AbstractChannelHandlerContext.java:1071)
>   at 
> io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:164)
>   at 
> io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:472)
>   at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:500)
>   at 
> io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)
>   at 
> io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
>   at 
> io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
>   at java.base/java.lang.Thread.run(Unknown Source)
> 2021-11-21 18:25:21,868 ERROR Failed to send RPC RPC 5040314884474308699 to 
> /10.1.201.113:58354: java.nio.channels.ClosedChannelException 
> (org.apache.spark.network.client.TransportClient) [rpc-server-4-1]
> java.nio.channels.ClosedChannelException
>   at 
> io.netty.channel.AbstractChannel$AbstractUnsafe.newClosedChannelException(AbstractChannel.java:957)
>   at 
> io.netty.channel.AbstractChannel$AbstractUnsafe.write(AbstractChannel.java:865)
>   at 
> io.netty.channel.DefaultChannelPipeline$HeadContext.write(DefaultChannelPipeline.java:1367)
>   at 
> io.netty.channel.AbstractChannelHandlerContext.invokeWrite0(AbstractChannelHandlerContext.java:717)
>   at 
> io.netty.channel.AbstractChannelHandlerContext.invokeWriteAndFlush(AbstractChannelHandlerContext.java:764)
>   at 
> io.netty.channel.AbstractChannelHandlerContext$WriteTask.run(AbstractChannelHandlerContext.java:1071)
>   at 
> io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:164)

[jira] [Created] (SPARK-37432) Driver keep a record of decommission executor

2021-11-21 Thread Hoa Le (Jira)
Hoa Le created SPARK-37432:
--

 Summary: Driver keep a record of decommission executor
 Key: SPARK-37432
 URL: https://issues.apache.org/jira/browse/SPARK-37432
 Project: Spark
  Issue Type: Bug
  Components: Kubernetes
Affects Versions: 3.1.1
Reporter: Hoa Le


Hi,

We are running spark on k8s with version 3.1.1. After the spark application 
running for a while, we are getting the exception below:

On driver: 

 
{code:java}
2021-11-21 18:25:21,859 ERROR Failed to send RPC RPC 6827167497981418905 to 
/10.1.201.113:58354: java.nio.channels.ClosedChannelException 
(org.apache.spark.network.client.TransportClient) [rpc-server-4-1]
java.nio.channels.ClosedChannelException
at 
io.netty.channel.AbstractChannel$AbstractUnsafe.newClosedChannelException(AbstractChannel.java:957)
at 
io.netty.channel.AbstractChannel$AbstractUnsafe.write(AbstractChannel.java:865)
at 
io.netty.channel.DefaultChannelPipeline$HeadContext.write(DefaultChannelPipeline.java:1367)
at 
io.netty.channel.AbstractChannelHandlerContext.invokeWrite0(AbstractChannelHandlerContext.java:717)
at 
io.netty.channel.AbstractChannelHandlerContext.invokeWriteAndFlush(AbstractChannelHandlerContext.java:764)
at 
io.netty.channel.AbstractChannelHandlerContext$WriteTask.run(AbstractChannelHandlerContext.java:1071)
at 
io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:164)
at 
io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:472)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:500)
at 
io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)
at 
io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at 
io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base/java.lang.Thread.run(Unknown Source)
2021-11-21 18:25:21,864 ERROR Failed to send RPC RPC 7618635518207296341 to 
/10.1.201.113:58354: java.nio.channels.ClosedChannelException 
(org.apache.spark.network.client.TransportClient) [rpc-server-4-1]
java.nio.channels.ClosedChannelException
at 
io.netty.channel.AbstractChannel$AbstractUnsafe.newClosedChannelException(AbstractChannel.java:957)
at 
io.netty.channel.AbstractChannel$AbstractUnsafe.write(AbstractChannel.java:865)
at 
io.netty.channel.DefaultChannelPipeline$HeadContext.write(DefaultChannelPipeline.java:1367)
at 
io.netty.channel.AbstractChannelHandlerContext.invokeWrite0(AbstractChannelHandlerContext.java:717)
at 
io.netty.channel.AbstractChannelHandlerContext.invokeWriteAndFlush(AbstractChannelHandlerContext.java:764)
at 
io.netty.channel.AbstractChannelHandlerContext$WriteTask.run(AbstractChannelHandlerContext.java:1071)
at 
io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:164)
at 
io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:472)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:500)
at 
io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)
at 
io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at 
io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base/java.lang.Thread.run(Unknown Source)
2021-11-21 18:25:21,868 ERROR Failed to send RPC RPC 5040314884474308699 to 
/10.1.201.113:58354: java.nio.channels.ClosedChannelException 
(org.apache.spark.network.client.TransportClient) [rpc-server-4-1]
java.nio.channels.ClosedChannelException
at 
io.netty.channel.AbstractChannel$AbstractUnsafe.newClosedChannelException(AbstractChannel.java:957)
at 
io.netty.channel.AbstractChannel$AbstractUnsafe.write(AbstractChannel.java:865)
at 
io.netty.channel.DefaultChannelPipeline$HeadContext.write(DefaultChannelPipeline.java:1367)
at 
io.netty.channel.AbstractChannelHandlerContext.invokeWrite0(AbstractChannelHandlerContext.java:717)
at 
io.netty.channel.AbstractChannelHandlerContext.invokeWriteAndFlush(AbstractChannelHandlerContext.java:764)
at 
io.netty.channel.AbstractChannelHandlerContext$WriteTask.run(AbstractChannelHandlerContext.java:1071)
at 
io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:164)
at 
io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:472)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:500)
at 
io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)
at 
io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
  

[jira] [Assigned] (SPARK-37104) RDD and DStream should be covariant

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37104?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz reassigned SPARK-37104:
--

Assignee: Maciej Szymkiewicz

> RDD and DStream should be covariant
> ---
>
> Key: SPARK-37104
> URL: https://issues.apache.org/jira/browse/SPARK-37104
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Affects Versions: 3.1.0, 3.2.0, 3.3.0
>Reporter: Maciej Szymkiewicz
>Assignee: Maciej Szymkiewicz
>Priority: Major
>
> At the moment {{RDD}} and {{DStream}} are defined as invariant.
>  
> However, there are immutable and could be marked as covariant.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-37104) RDD and DStream should be covariant

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37104?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz resolved SPARK-37104.

Fix Version/s: 3.3.0
   Resolution: Fixed

Issue resolved by pull request 34374
[https://github.com/apache/spark/pull/34374]

> RDD and DStream should be covariant
> ---
>
> Key: SPARK-37104
> URL: https://issues.apache.org/jira/browse/SPARK-37104
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Affects Versions: 3.1.0, 3.2.0, 3.3.0
>Reporter: Maciej Szymkiewicz
>Assignee: Maciej Szymkiewicz
>Priority: Major
> Fix For: 3.3.0
>
>
> At the moment {{RDD}} and {{DStream}} are defined as invariant.
>  
> However, there are immutable and could be marked as covariant.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-37431) pyspark.{ml, mllib}.linalg.Vector API

2021-11-21 Thread Maciej Szymkiewicz (Jira)
Maciej Szymkiewicz created SPARK-37431:
--

 Summary: pyspark.{ml, mllib}.linalg.Vector API
 Key: SPARK-37431
 URL: https://issues.apache.org/jira/browse/SPARK-37431
 Project: Spark
  Issue Type: Improvement
  Components: ML, MLlib, PySpark
Affects Versions: 3.3.0
Reporter: Maciej Szymkiewicz


At the moment both {{Vector}} implementations have minimal API. {{mllib}}

{code:python}
class Vector(object):
__UDT__ = 
def toArray(self): ...
def asML(self): ...
{code}

{{ml}}

{code:python}
class Vector(object):
__UDT__ = ...
def toArray(self): ...

{code}

which doesn't cover actual API being used.

This causes typing issues, when we expect any {{Vector}}.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-37430) Inline type hints for python/pyspark/mllib/linalg/distributed.py

2021-11-21 Thread Maciej Szymkiewicz (Jira)
Maciej Szymkiewicz created SPARK-37430:
--

 Summary: Inline type hints for 
python/pyspark/mllib/linalg/distributed.py
 Key: SPARK-37430
 URL: https://issues.apache.org/jira/browse/SPARK-37430
 Project: Spark
  Issue Type: Sub-task
  Components: MLlib, PySpark
Affects Versions: 3.3.0
Reporter: Maciej Szymkiewicz


Inline type hints from python/pyspark/mlib/linalg/distributed.pyi to 
python/pyspark/mllib/linalg/distributed.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-37429) Inline type hints for python/pyspark/mllib/linalg/__init__.py

2021-11-21 Thread Maciej Szymkiewicz (Jira)
Maciej Szymkiewicz created SPARK-37429:
--

 Summary: Inline type hints for 
python/pyspark/mllib/linalg/__init__.py
 Key: SPARK-37429
 URL: https://issues.apache.org/jira/browse/SPARK-37429
 Project: Spark
  Issue Type: Sub-task
  Components: MLlib, PySpark
Affects Versions: 3.3.0
Reporter: Maciej Szymkiewicz


Inline type hints from python/pyspark/mlib/linalg/__init__.pyi to 
python/pyspark/mllib/linalg/__init__.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37423) Inline type hints for python/pyspark/mllib/fpm.py

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37423?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz updated SPARK-37423:
---
Summary: Inline type hints for python/pyspark/mllib/fpm.py  (was: Inline 
type hints for python/pyspark/mllib/fpm.pypy)

> Inline type hints for python/pyspark/mllib/fpm.py
> -
>
> Key: SPARK-37423
> URL: https://issues.apache.org/jira/browse/SPARK-37423
> Project: Spark
>  Issue Type: Sub-task
>  Components: MLlib, PySpark
>Affects Versions: 3.3.0
>Reporter: Maciej Szymkiewicz
>Priority: Major
>
> Inline type hints from python/pyspark/mlib/fpm.pyi to 
> python/pyspark/mllib/fpm.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37427) Inline type hints for python/pyspark/mllib/tree.py

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37427?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz updated SPARK-37427:
---
Summary: Inline type hints for python/pyspark/mllib/tree.py  (was: Inline 
type hints for python/pyspark/mllib/tree.pypy)

> Inline type hints for python/pyspark/mllib/tree.py
> --
>
> Key: SPARK-37427
> URL: https://issues.apache.org/jira/browse/SPARK-37427
> Project: Spark
>  Issue Type: Sub-task
>  Components: MLlib, PySpark
>Affects Versions: 3.3.0
>Reporter: Maciej Szymkiewicz
>Priority: Major
>
> Inline type hints from python/pyspark/mlib/tree.pyi to 
> python/pyspark/mllib/tree.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37428) Inline type hints for python/pyspark/mllib/util.py

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37428?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz updated SPARK-37428:
---
Summary: Inline type hints for python/pyspark/mllib/util.py  (was: Inline 
type hints for python/pyspark/mllib/util.pypy)

> Inline type hints for python/pyspark/mllib/util.py
> --
>
> Key: SPARK-37428
> URL: https://issues.apache.org/jira/browse/SPARK-37428
> Project: Spark
>  Issue Type: Sub-task
>  Components: PySpark
>Affects Versions: 3.3.0
>Reporter: Maciej Szymkiewicz
>Priority: Major
>
> Inline type hints from python/pyspark/mlib/util.pyi to 
> python/pyspark/mllib/util.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37424) Inline type hints for python/pyspark/mllib/random.py

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37424?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz updated SPARK-37424:
---
Summary: Inline type hints for python/pyspark/mllib/random.py  (was: Inline 
type hints for python/pyspark/mllib/random.pypy)

> Inline type hints for python/pyspark/mllib/random.py
> 
>
> Key: SPARK-37424
> URL: https://issues.apache.org/jira/browse/SPARK-37424
> Project: Spark
>  Issue Type: Sub-task
>  Components: MLlib, PySpark
>Affects Versions: 3.3.0
>Reporter: Maciej Szymkiewicz
>Priority: Major
>
> Inline type hints from python/pyspark/mlib/random.pyi to 
> python/pyspark/mllib/random.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37426) Inline type hints for python/pyspark/mllib/regression.py

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37426?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz updated SPARK-37426:
---
Summary: Inline type hints for python/pyspark/mllib/regression.py  (was: 
Inline type hints for python/pyspark/mllib/regression.pypy)

> Inline type hints for python/pyspark/mllib/regression.py
> 
>
> Key: SPARK-37426
> URL: https://issues.apache.org/jira/browse/SPARK-37426
> Project: Spark
>  Issue Type: Sub-task
>  Components: MLlib, PySpark
>Affects Versions: 3.3.0
>Reporter: Maciej Szymkiewicz
>Priority: Major
>
> Inline type hints from python/pyspark/mlib/regression.pyi to 
> python/pyspark/mllib/regression.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37425) Inline type hints for python/pyspark/mllib/recommendation.py

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37425?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz updated SPARK-37425:
---
Summary: Inline type hints for python/pyspark/mllib/recommendation.py  
(was: Inline type hints for python/pyspark/mllib/recommendation.pypy)

> Inline type hints for python/pyspark/mllib/recommendation.py
> 
>
> Key: SPARK-37425
> URL: https://issues.apache.org/jira/browse/SPARK-37425
> Project: Spark
>  Issue Type: Sub-task
>  Components: MLlib, PySpark
>Affects Versions: 3.3.0
>Reporter: Maciej Szymkiewicz
>Priority: Major
>
> Inline type hints from python/pyspark/mlib/recommendation.pyi to 
> python/pyspark/mllib/recommendation.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37422) Inline type hints for python/pyspark/mllib/feature.py

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37422?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz updated SPARK-37422:
---
Summary: Inline type hints for python/pyspark/mllib/feature.py  (was: 
Inline type hints for python/pyspark/mllib/feature.pypy)

> Inline type hints for python/pyspark/mllib/feature.py
> -
>
> Key: SPARK-37422
> URL: https://issues.apache.org/jira/browse/SPARK-37422
> Project: Spark
>  Issue Type: Sub-task
>  Components: MLlib, PySpark
>Affects Versions: 3.3.0
>Reporter: Maciej Szymkiewicz
>Priority: Major
>
> Inline type hints from python/pyspark/mlib/feature.pyi to 
> python/pyspark/mllib/feature.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37421) Inline type hints for python/pyspark/mllib/evaluation.py

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37421?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz updated SPARK-37421:
---
Summary: Inline type hints for python/pyspark/mllib/evaluation.py  (was: 
Inline type hints for python/pyspark/mllib/evaluation.pypy)

> Inline type hints for python/pyspark/mllib/evaluation.py
> 
>
> Key: SPARK-37421
> URL: https://issues.apache.org/jira/browse/SPARK-37421
> Project: Spark
>  Issue Type: Sub-task
>  Components: MLlib, PySpark
>Affects Versions: 3.3.0
>Reporter: Maciej Szymkiewicz
>Priority: Major
>
> Inline type hints from python/pyspark/mlib/evaluation.pyi to 
> python/pyspark/mllib/evaluation.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37236) Inline type hints for KernelDensity.pyi, test.py in python/pyspark/mllib/stat/

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37236?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz updated SPARK-37236:
---
Parent: (was: SPARK-37233)
Issue Type: Improvement  (was: Sub-task)

> Inline type hints for KernelDensity.pyi, test.py in python/pyspark/mllib/stat/
> --
>
> Key: SPARK-37236
> URL: https://issues.apache.org/jira/browse/SPARK-37236
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Affects Versions: 3.3.0
>Reporter: dch nguyen
>Assignee: dch nguyen
>Priority: Major
> Fix For: 3.3.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37236) Inline type hints for KernelDensity.pyi, test.py in python/pyspark/mllib/stat/

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37236?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz updated SPARK-37236:
---
Parent: SPARK-37396
Issue Type: Sub-task  (was: Improvement)

> Inline type hints for KernelDensity.pyi, test.py in python/pyspark/mllib/stat/
> --
>
> Key: SPARK-37236
> URL: https://issues.apache.org/jira/browse/SPARK-37236
> Project: Spark
>  Issue Type: Sub-task
>  Components: PySpark
>Affects Versions: 3.3.0
>Reporter: dch nguyen
>Assignee: dch nguyen
>Priority: Major
> Fix For: 3.3.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37235) Inline type hints for distribution.py and __init__.py in python/pyspark/mllib/stat

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37235?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz updated SPARK-37235:
---
Parent: SPARK-37396
Issue Type: Sub-task  (was: Improvement)

> Inline type hints for distribution.py and __init__.py in 
> python/pyspark/mllib/stat
> --
>
> Key: SPARK-37235
> URL: https://issues.apache.org/jira/browse/SPARK-37235
> Project: Spark
>  Issue Type: Sub-task
>  Components: PySpark
>Affects Versions: 3.3.0
>Reporter: dch nguyen
>Assignee: dch nguyen
>Priority: Major
> Fix For: 3.3.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37235) Inline type hints for distribution.py and __init__.py in python/pyspark/mllib/stat

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37235?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz updated SPARK-37235:
---
Parent: (was: SPARK-37396)
Issue Type: Improvement  (was: Sub-task)

> Inline type hints for distribution.py and __init__.py in 
> python/pyspark/mllib/stat
> --
>
> Key: SPARK-37235
> URL: https://issues.apache.org/jira/browse/SPARK-37235
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Affects Versions: 3.3.0
>Reporter: dch nguyen
>Assignee: dch nguyen
>Priority: Major
> Fix For: 3.3.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37235) Inline type hints for distribution.py and __init__.py in python/pyspark/mllib/stat

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37235?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz updated SPARK-37235:
---
Parent: SPARK-37396
Issue Type: Sub-task  (was: Improvement)

> Inline type hints for distribution.py and __init__.py in 
> python/pyspark/mllib/stat
> --
>
> Key: SPARK-37235
> URL: https://issues.apache.org/jira/browse/SPARK-37235
> Project: Spark
>  Issue Type: Sub-task
>  Components: PySpark
>Affects Versions: 3.3.0
>Reporter: dch nguyen
>Assignee: dch nguyen
>Priority: Major
> Fix For: 3.3.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37235) Inline type hints for distribution.py and __init__.py in python/pyspark/mllib/stat

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37235?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz updated SPARK-37235:
---
Parent: (was: SPARK-37233)
Issue Type: Improvement  (was: Sub-task)

> Inline type hints for distribution.py and __init__.py in 
> python/pyspark/mllib/stat
> --
>
> Key: SPARK-37235
> URL: https://issues.apache.org/jira/browse/SPARK-37235
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Affects Versions: 3.3.0
>Reporter: dch nguyen
>Assignee: dch nguyen
>Priority: Major
> Fix For: 3.3.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37234) Inline type hints for python/pyspark/mllib/stat/_statistics.py

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37234?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz updated SPARK-37234:
---
Parent: SPARK-37396
Issue Type: Sub-task  (was: Improvement)

> Inline type hints for python/pyspark/mllib/stat/_statistics.py
> --
>
> Key: SPARK-37234
> URL: https://issues.apache.org/jira/browse/SPARK-37234
> Project: Spark
>  Issue Type: Sub-task
>  Components: PySpark
>Affects Versions: 3.3.0
>Reporter: dch nguyen
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-37234) Inline type hints for python/pyspark/mllib/stat/_statistics.py

2021-11-21 Thread Maciej Szymkiewicz (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-37234?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Maciej Szymkiewicz updated SPARK-37234:
---
Parent: (was: SPARK-37233)
Issue Type: Improvement  (was: Sub-task)

> Inline type hints for python/pyspark/mllib/stat/_statistics.py
> --
>
> Key: SPARK-37234
> URL: https://issues.apache.org/jira/browse/SPARK-37234
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Affects Versions: 3.3.0
>Reporter: dch nguyen
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-37428) Inline type hints for python/pyspark/mllib/util.pypy

2021-11-21 Thread Maciej Szymkiewicz (Jira)
Maciej Szymkiewicz created SPARK-37428:
--

 Summary: Inline type hints for python/pyspark/mllib/util.pypy
 Key: SPARK-37428
 URL: https://issues.apache.org/jira/browse/SPARK-37428
 Project: Spark
  Issue Type: Sub-task
  Components: PySpark
Affects Versions: 3.3.0
Reporter: Maciej Szymkiewicz


Inline type hints from python/pyspark/mlib/util.pyi to 
python/pyspark/mllib/util.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-37427) Inline type hints for python/pyspark/mllib/tree.pypy

2021-11-21 Thread Maciej Szymkiewicz (Jira)
Maciej Szymkiewicz created SPARK-37427:
--

 Summary: Inline type hints for python/pyspark/mllib/tree.pypy
 Key: SPARK-37427
 URL: https://issues.apache.org/jira/browse/SPARK-37427
 Project: Spark
  Issue Type: Sub-task
  Components: MLlib, PySpark
Affects Versions: 3.3.0
Reporter: Maciej Szymkiewicz


Inline type hints from python/pyspark/mlib/tree.pyi to 
python/pyspark/mllib/tree.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-37425) Inline type hints for python/pyspark/mllib/recommendation.pypy

2021-11-21 Thread Maciej Szymkiewicz (Jira)
Maciej Szymkiewicz created SPARK-37425:
--

 Summary: Inline type hints for 
python/pyspark/mllib/recommendation.pypy
 Key: SPARK-37425
 URL: https://issues.apache.org/jira/browse/SPARK-37425
 Project: Spark
  Issue Type: Sub-task
  Components: MLlib, PySpark
Affects Versions: 3.3.0
Reporter: Maciej Szymkiewicz


Inline type hints from python/pyspark/mlib/recommendation.pyi to 
python/pyspark/mllib/recommendation.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-37426) Inline type hints for python/pyspark/mllib/regression.pypy

2021-11-21 Thread Maciej Szymkiewicz (Jira)
Maciej Szymkiewicz created SPARK-37426:
--

 Summary: Inline type hints for python/pyspark/mllib/regression.pypy
 Key: SPARK-37426
 URL: https://issues.apache.org/jira/browse/SPARK-37426
 Project: Spark
  Issue Type: Sub-task
  Components: MLlib, PySpark
Affects Versions: 3.3.0
Reporter: Maciej Szymkiewicz


Inline type hints from python/pyspark/mlib/regression.pyi to 
python/pyspark/mllib/regression.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-37423) Inline type hints for python/pyspark/mllib/fpm.pypy

2021-11-21 Thread Maciej Szymkiewicz (Jira)
Maciej Szymkiewicz created SPARK-37423:
--

 Summary: Inline type hints for python/pyspark/mllib/fpm.pypy
 Key: SPARK-37423
 URL: https://issues.apache.org/jira/browse/SPARK-37423
 Project: Spark
  Issue Type: Sub-task
  Components: MLlib, PySpark
Affects Versions: 3.3.0
Reporter: Maciej Szymkiewicz


Inline type hints from python/pyspark/mlib/fpm.pyi to 
python/pyspark/mllib/fpm.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-37424) Inline type hints for python/pyspark/mllib/random.pypy

2021-11-21 Thread Maciej Szymkiewicz (Jira)
Maciej Szymkiewicz created SPARK-37424:
--

 Summary: Inline type hints for python/pyspark/mllib/random.pypy
 Key: SPARK-37424
 URL: https://issues.apache.org/jira/browse/SPARK-37424
 Project: Spark
  Issue Type: Sub-task
  Components: MLlib, PySpark
Affects Versions: 3.3.0
Reporter: Maciej Szymkiewicz


Inline type hints from python/pyspark/mlib/random.pyi to 
python/pyspark/mllib/random.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-37422) Inline type hints for python/pyspark/mllib/feature.pypy

2021-11-21 Thread Maciej Szymkiewicz (Jira)
Maciej Szymkiewicz created SPARK-37422:
--

 Summary: Inline type hints for python/pyspark/mllib/feature.pypy
 Key: SPARK-37422
 URL: https://issues.apache.org/jira/browse/SPARK-37422
 Project: Spark
  Issue Type: Sub-task
  Components: MLlib, PySpark
Affects Versions: 3.3.0
Reporter: Maciej Szymkiewicz


Inline type hints from python/pyspark/mlib/feature.pyi to 
python/pyspark/mllib/feature.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-37421) Inline type hints for python/pyspark/mllib/evaluation.pypy

2021-11-21 Thread Maciej Szymkiewicz (Jira)
Maciej Szymkiewicz created SPARK-37421:
--

 Summary: Inline type hints for python/pyspark/mllib/evaluation.pypy
 Key: SPARK-37421
 URL: https://issues.apache.org/jira/browse/SPARK-37421
 Project: Spark
  Issue Type: Sub-task
  Components: MLlib, PySpark
Affects Versions: 3.3.0
Reporter: Maciej Szymkiewicz


Inline type hints from python/pyspark/mlib/evaluation.pyi to 
python/pyspark/mllib/evaluation.py



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-37395) Inline type hint files for files in python/pyspark/ml

2021-11-21 Thread Maciej Szymkiewicz (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-37395?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17446996#comment-17446996
 ] 

Maciej Szymkiewicz commented on SPARK-37395:


cc [~hyukjin.kwon] [~ueshin] [~XinrongM] FYI. Also SPARK-37396

> Inline type hint files for files in python/pyspark/ml
> -
>
> Key: SPARK-37395
> URL: https://issues.apache.org/jira/browse/SPARK-37395
> Project: Spark
>  Issue Type: Umbrella
>  Components: ML, PySpark
>Affects Versions: 3.3.0
>Reporter: Maciej Szymkiewicz
>Priority: Major
>
> Currently there are type hint stub files ({{*.pyi}}) to show the expected 
> types for functions, but we can also take advantage of static type checking 
> within the functions by inlining the type hints.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org