[
https://issues.apache.org/jira/browse/ARROW-6520?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Furkan Tektas updated ARROW-6520:
---------------------------------
Environment:
python(3.7.3), pyarrow(0.14.1), arrow-cpp(0.14.1), parquet-cpp(1.5.1), Arch
Linux x86_64
was:
Arch Linux x86_64
arrow-cpp 0.14.1 py37h6b969ab_1 conda-forge
parquet-cpp 1.5.1 2 conda-forge
pyarrow 0.14.1 py37h8b68381_0 conda-forge
python 3.7.3 h33d41f4_1 conda-forge
> Segmentation fault on writing tables with fixed size binary fields
> -------------------------------------------------------------------
>
> Key: ARROW-6520
> URL: https://issues.apache.org/jira/browse/ARROW-6520
> Project: Apache Arrow
> Issue Type: Bug
> Components: Python
> Affects Versions: 0.14.1
> Environment: python(3.7.3), pyarrow(0.14.1), arrow-cpp(0.14.1),
> parquet-cpp(1.5.1), Arch Linux x86_64
> Reporter: Furkan Tektas
> Priority: Critical
> Labels: newbie
>
> I'm not sure if this should be reported to Parquet or here.
> When I tried to serialize a pyarrow table with a fixed size binary field
> (holds 16 byte UUID4 information) to a parquet file, segmentation fault
> occurs.
> Here is the minimal example to reproduce:
> {{import pyarrow as pa}}
> {{from pyarrow import parquet as pq}}
> {{data = \{"col": pa.array([b"1234" for _ in range(10)])}}}
> {{fields = [("col", pa.binary(4))]}}
> {{schema = pa.schema(fields)}}
> {{table = pa.table(data, schema)}}
> {{pq.write_table(table, "test.parquet")}}
> {{segmentation fault (core dumped) ipython}}
>
> Yet, it works if I don't specify the size of the binary field.
> {{import pyarrow as pa}}
> {{from pyarrow import parquet as pq}}
> {{data = \{"col": pa.array([b"1234" for _ in range(10)])}}}
> {{fields = [("col", pa.binary())]}}
> {{schema = pa.schema(fields)}}
> {{table = pa.table(data, schema)}}
> {{pq.write_table(table, "test.parquet")}}
> Thanks,
--
This message was sent by Atlassian Jira
(v8.3.2#803003)