[ 
https://issues.apache.org/jira/browse/SPARK-39317?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hyukjin Kwon reassigned SPARK-39317:
------------------------------------

    Assignee: Yikun Jiang

> groupby.apply doc test failed when SPARK_CONF_ARROW_ENABLED disable
> -------------------------------------------------------------------
>
>                 Key: SPARK-39317
>                 URL: https://issues.apache.org/jira/browse/SPARK-39317
>             Project: Spark
>          Issue Type: Sub-task
>          Components: Pandas API on Spark
>    Affects Versions: 3.4.0
>            Reporter: Yikun Jiang
>            Assignee: Yikun Jiang
>            Priority: Major
>
> {code:java}
> Traceback (most recent call last):
>   File 
> "/Users/yikun/venv/lib/python3.9/site-packages/IPython/core/interactiveshell.py",
>  line 3361, in run_code
>     exec(code_obj, self.user_global_ns, self.user_ns)
>   File "<ipython-input-1-3cad40f83c12>", line 1, in <cell line: 1>
>     g.apply(plus_min).sort_index()
>   File "/Users/yikun/spark/python/pyspark/pandas/groupby.py", line 1444, in 
> apply
>     psser_or_psdf = ps.from_pandas(pser_or_pdf)
>   File "/Users/yikun/spark/python/pyspark/pandas/namespace.py", line 153, in 
> from_pandas
>     return DataFrame(pobj)
>   File "/Users/yikun/spark/python/pyspark/pandas/frame.py", line 457, in 
> __init__
>     internal = InternalFrame.from_pandas(pdf)
>   File "/Users/yikun/spark/python/pyspark/pandas/internal.py", line 1473, in 
> from_pandas
>     sdf = default_session().createDataFrame(pdf, schema=schema)
>   File "/Users/yikun/spark/python/pyspark/sql/session.py", line 961, in 
> createDataFrame
>     return super(SparkSession, self).createDataFrame(  # type: 
> ignore[call-overload]
>   File "/Users/yikun/spark/python/pyspark/sql/pandas/conversion.py", line 
> 437, in createDataFrame
>     return self._create_dataframe(converted_data, schema, samplingRatio, 
> verifySchema)
>   File "/Users/yikun/spark/python/pyspark/sql/session.py", line 1006, in 
> _create_dataframe
>     rdd, struct = self._createFromLocal(map(prepare, data), schema)
>   File "/Users/yikun/spark/python/pyspark/sql/session.py", line 698, in 
> _createFromLocal
>     data = list(data)
>   File "/Users/yikun/spark/python/pyspark/sql/session.py", line 980, in 
> prepare
>     verify_func(obj)
>   File "/Users/yikun/spark/python/pyspark/sql/types.py", line 1763, in verify
>     verify_value(obj)
>   File "/Users/yikun/spark/python/pyspark/sql/types.py", line 1741, in 
> verify_struct
>     verifier(v)
>   File "/Users/yikun/spark/python/pyspark/sql/types.py", line 1763, in verify
>     verify_value(obj)
>   File "/Users/yikun/spark/python/pyspark/sql/types.py", line 1686, in 
> verify_long
>     verify_acceptable_types(obj)
>   File "/Users/yikun/spark/python/pyspark/sql/types.py", line 1633, in 
> verify_acceptable_types
>     raise TypeError(
> TypeError: field B: LongType() can not accept object 2 in type <class 
> 'numpy.int64'> {code}



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to