Bryan Cutler commented on ARROW-6429:

The failure seems to be caused from the removal of pyarrow.Column in favor of 
pyarrow.ChunkedArray. Spark iterates over columns of a pyarrow.Table, calls 
{{to_pandas()}} on each column, and assumes the result is a pd.Series. If the 
column is actually a pyarrow.ChunkedArray, then {{to_pandas()}} can be a 
numpy.array. [~wesmckinn] [~pitrou] I know in the pydoc it says the returned 
value can either be a pandas.Series or numpy.array, but is there anyway to 
ensure it is the former or is that the job of the caller?

> [CI][Crossbow] Nightly spark integration job fails
> --------------------------------------------------
>                 Key: ARROW-6429
>                 URL: https://issues.apache.org/jira/browse/ARROW-6429
>             Project: Apache Arrow
>          Issue Type: Bug
>          Components: Continuous Integration
>            Reporter: Neal Richardson
>            Assignee: Bryan Cutler
>            Priority: Blocker
>              Labels: nightly
>             Fix For: 0.15.0
> See https://circleci.com/gh/ursa-labs/crossbow/2310. Either fix, skip job and 
> create followup Jira to unskip, or delete job.

This message was sent by Atlassian Jira

Reply via email to