[
https://issues.apache.org/jira/browse/SPARK-39317?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Hyukjin Kwon reassigned SPARK-39317:
------------------------------------
Assignee: Yikun Jiang
> groupby.apply doc test failed when SPARK_CONF_ARROW_ENABLED disable
> -------------------------------------------------------------------
>
> Key: SPARK-39317
> URL: https://issues.apache.org/jira/browse/SPARK-39317
> Project: Spark
> Issue Type: Sub-task
> Components: Pandas API on Spark
> Affects Versions: 3.4.0
> Reporter: Yikun Jiang
> Assignee: Yikun Jiang
> Priority: Major
>
> {code:java}
> Traceback (most recent call last):
> File
> "/Users/yikun/venv/lib/python3.9/site-packages/IPython/core/interactiveshell.py",
> line 3361, in run_code
> exec(code_obj, self.user_global_ns, self.user_ns)
> File "<ipython-input-1-3cad40f83c12>", line 1, in <cell line: 1>
> g.apply(plus_min).sort_index()
> File "/Users/yikun/spark/python/pyspark/pandas/groupby.py", line 1444, in
> apply
> psser_or_psdf = ps.from_pandas(pser_or_pdf)
> File "/Users/yikun/spark/python/pyspark/pandas/namespace.py", line 153, in
> from_pandas
> return DataFrame(pobj)
> File "/Users/yikun/spark/python/pyspark/pandas/frame.py", line 457, in
> __init__
> internal = InternalFrame.from_pandas(pdf)
> File "/Users/yikun/spark/python/pyspark/pandas/internal.py", line 1473, in
> from_pandas
> sdf = default_session().createDataFrame(pdf, schema=schema)
> File "/Users/yikun/spark/python/pyspark/sql/session.py", line 961, in
> createDataFrame
> return super(SparkSession, self).createDataFrame( # type:
> ignore[call-overload]
> File "/Users/yikun/spark/python/pyspark/sql/pandas/conversion.py", line
> 437, in createDataFrame
> return self._create_dataframe(converted_data, schema, samplingRatio,
> verifySchema)
> File "/Users/yikun/spark/python/pyspark/sql/session.py", line 1006, in
> _create_dataframe
> rdd, struct = self._createFromLocal(map(prepare, data), schema)
> File "/Users/yikun/spark/python/pyspark/sql/session.py", line 698, in
> _createFromLocal
> data = list(data)
> File "/Users/yikun/spark/python/pyspark/sql/session.py", line 980, in
> prepare
> verify_func(obj)
> File "/Users/yikun/spark/python/pyspark/sql/types.py", line 1763, in verify
> verify_value(obj)
> File "/Users/yikun/spark/python/pyspark/sql/types.py", line 1741, in
> verify_struct
> verifier(v)
> File "/Users/yikun/spark/python/pyspark/sql/types.py", line 1763, in verify
> verify_value(obj)
> File "/Users/yikun/spark/python/pyspark/sql/types.py", line 1686, in
> verify_long
> verify_acceptable_types(obj)
> File "/Users/yikun/spark/python/pyspark/sql/types.py", line 1633, in
> verify_acceptable_types
> raise TypeError(
> TypeError: field B: LongType() can not accept object 2 in type <class
> 'numpy.int64'> {code}
--
This message was sent by Atlassian Jira
(v8.20.7#820007)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]