Sandeep Singh created SPARK-41870:
-------------------------------------

             Summary: Handle duplicate columns in `createDataFrame`
                 Key: SPARK-41870
                 URL: https://issues.apache.org/jira/browse/SPARK-41870
             Project: Spark
          Issue Type: Sub-task
          Components: Connect
    Affects Versions: 3.4.0
            Reporter: Sandeep Singh


{code:java}
import array

data = [Row(longarray=array.array("l", [-9223372036854775808, 0, 
9223372036854775807]))]
df = self.spark.createDataFrame(data) {code}
Error:
{code:java}
Traceback (most recent call last):
  File 
"/Users/s.singh/personal/spark-oss/python/pyspark/sql/tests/test_dataframe.py", 
line 1220, in test_create_dataframe_from_array_of_long
    df = self.spark.createDataFrame(data)
  File 
"/Users/s.singh/personal/spark-oss/python/pyspark/sql/connect/session.py", line 
260, in createDataFrame
    table = pa.Table.from_pylist([row.asDict(recursive=True) for row in _data])
  File "pyarrow/table.pxi", line 3700, in pyarrow.lib.Table.from_pylist
  File "pyarrow/table.pxi", line 5221, in pyarrow.lib._from_pylist
  File "pyarrow/table.pxi", line 3575, in pyarrow.lib.Table.from_arrays
  File "pyarrow/table.pxi", line 1383, in pyarrow.lib._sanitize_arrays
  File "pyarrow/table.pxi", line 1364, in pyarrow.lib._schema_from_arrays
  File "pyarrow/array.pxi", line 320, in pyarrow.lib.array
  File "pyarrow/array.pxi", line 39, in pyarrow.lib._sequence_to_array
  File "pyarrow/error.pxi", line 144, in 
pyarrow.lib.pyarrow_internal_check_status
  File "pyarrow/error.pxi", line 100, in pyarrow.lib.check_status
pyarrow.lib.ArrowInvalid: Could not convert array('l', [-9223372036854775808, 
0, 9223372036854775807]) with type array.array: did not recognize Python value 
type when inferring an Arrow data type{code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to