[
https://issues.apache.org/jira/browse/ARROW-6520?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Furkan Tektas updated ARROW-6520:
---------------------------------
Description:
I'm not sure if this should be reported to Parquet or here.
When I tried to serialize a pyarrow table with a fixed size binary field (holds
16 byte UUID4 information) to a parquet file, segmentation fault occurs.
Here is the minimal example to reproduce:
{{import pyarrow as pa}}
{{ from pyarrow import parquet as pq}}
{{ data = \{"col": pa.array([b"1234" for _ in range(10)])}}}
{{ fields = [("col", pa.binary(4))]}}
{{ schema = pa.schema(fields)}}
{{ table = pa.table(data, schema)}}
{{ pq.write_table(table, "test.parquet")}}
{{ segmentation fault (core dumped) ipython}}
Yet, it works if I don't specify the size of the binary field.
{{import pyarrow as pa}}
{{from pyarrow import parquet as pq}}
{{data = \{"col": pa.array([b"1234" for _ in range(10)])}}}
{{fields = [("col", pa.binary())]}}
{{schema = pa.schema(fields)}}
{{table = pa.table(data, schema)}}
{{pq.write_table(table, "test.parquet")}}
Thanks,
was:
I'm not sure if this should be reported to Parquet or here.
When I tried to serialize a pyarrow table with a fixed size binary field (holds
16 byte UUID4 information) to a parquet file, segmentation fault occurs.
Here is the minimal example to reproduce:
{{ }}
{{ {color:#de350b}import pyarrow as pa{color}}}
{{{color:#de350b} from pyarrow import parquet as pqdata = \{"col":
pa.array([b"1234" for _ in range(10)])}{color}}}
{{{color:#de350b} fields = [("col", pa.binary(4))]{color}}}
{{{color:#de350b} schema = pa.schema(fields){color}}}
{{{color:#de350b} table = pa.table(data, schema){color}}}
{{{color:#de350b} pq.write_table(table, "test.parquet"){color}}}
{{ *{color:#de350b}segmentation fault (core dumped) ipython{color}*}}
{{{{Yet, it works if I don't specify the size of the binary field.}}}}
{{ {color:#00875a}import pyarrow as pa{color}}}
{{{color:#00875a} from pyarrow import parquet as
pq{color}}}{{{color:#00875a}data = \{"col": pa.array([b"1234" for _ in
range(10)])}{color}}}
{{{color:#00875a} fields = [("col", pa.binary())]{color}}}
{{{color:#00875a} schema = pa.schema(fields){color}}}
{{{color:#00875a} table = pa.table(data, schema){color}}}
{{{color:#00875a} pq.write_table(table, "test.parquet"){color}}}
{{ }}
Thanks,
> Segmentation fault on writing tables with fixed size binary fields
> -------------------------------------------------------------------
>
> Key: ARROW-6520
> URL: https://issues.apache.org/jira/browse/ARROW-6520
> Project: Apache Arrow
> Issue Type: Bug
> Components: Python
> Affects Versions: 0.14.1
> Environment: Arch Linux x86_64
> arrow-cpp 0.14.1 py37h6b969ab_1 conda-forge
> parquet-cpp 1.5.1 2 conda-forge
> pyarrow 0.14.1 py37h8b68381_0 conda-forge
> python 3.7.3 h33d41f4_1 conda-forge
> Reporter: Furkan Tektas
> Priority: Critical
> Labels: newbie
>
> I'm not sure if this should be reported to Parquet or here.
> When I tried to serialize a pyarrow table with a fixed size binary field
> (holds 16 byte UUID4 information) to a parquet file, segmentation fault
> occurs.
> Here is the minimal example to reproduce:
> {{import pyarrow as pa}}
> {{ from pyarrow import parquet as pq}}
> {{ data = \{"col": pa.array([b"1234" for _ in range(10)])}}}
> {{ fields = [("col", pa.binary(4))]}}
> {{ schema = pa.schema(fields)}}
> {{ table = pa.table(data, schema)}}
> {{ pq.write_table(table, "test.parquet")}}
> {{ segmentation fault (core dumped) ipython}}
>
> Yet, it works if I don't specify the size of the binary field.
> {{import pyarrow as pa}}
> {{from pyarrow import parquet as pq}}
> {{data = \{"col": pa.array([b"1234" for _ in range(10)])}}}
> {{fields = [("col", pa.binary())]}}
> {{schema = pa.schema(fields)}}
> {{table = pa.table(data, schema)}}
> {{pq.write_table(table, "test.parquet")}}
> Thanks,
--
This message was sent by Atlassian Jira
(v8.3.2#803003)