[ 
https://issues.apache.org/jira/browse/IMPALA-13966?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17944692#comment-17944692
 ] 

Quanlong Huang commented on IMPALA-13966:
-----------------------------------------

IIUC, we saw some similar tcmalloc issues in KrpcDataStreamRecvr: IMPALA-5528, 
IMPALA-6359, IMPALA-5518. They were resolved by allocating KrpcDataStreamRecvr 
RowBatch tuples from BufferPool, as [~kwho] summarized:
{quote}The fix for IMPALA-5518 allocates row batches for KrpcDataStreamRecvr 
from the bufferpool which maintains a per-core arena so allocations will be 
recycled, alleviating the pressure on the TCMalloc thread cache.
{quote}
I think we can do the same for MemPools used in the scanner. I have a WIP patch 
but haven't got time to finish it yet: [https://gerrit.cloudera.org/c/22123/]
Need to resolve some reservation issues.

Note that decompressing the parquet data page is just one place that could 
allocate large memory. There are other places like allocating tuple memory for 
long arrays in 
[CollectionValueBuilder::GetFreeMemory()|https://github.com/apache/impala/blob/50a98dce46fd337461f67b98979fece95c7bf738/be/src/runtime/collection-value-builder.h#L72]
 that could also allocate a large memory space (shown in the off-cpu flame 
graph of IMPALA-13487). If BATCH_SIZE is set too high, 
[RowBatch::ResizeAndAllocateTupleBuffer()|https://github.com/apache/impala/blob/50a98dce46fd337461f67b98979fece95c7bf738/be/src/runtime/row-batch.cc#L490]
 could also hit this issue. The above patch aims to solve these allocations.

> Heavy scan concurrency on Parquet tables with large page size is slow
> ---------------------------------------------------------------------
>
>                 Key: IMPALA-13966
>                 URL: https://issues.apache.org/jira/browse/IMPALA-13966
>             Project: IMPALA
>          Issue Type: Bug
>          Components: Backend
>    Affects Versions: Impala 4.5.0
>            Reporter: Michael Smith
>            Priority: Major
>
> When reading Parquet tables with large average page size under heavy scan 
> concurrency, we see performance significantly slow down.
> Impala writes Iceberg tables with its default page size of 64KB, unless 
> {{write.parquet.page-size-bytes}} is explicitly set. The Iceberg library 
> itself defaults to 1MB, and other tools - such as Spark - may use that 
> default when writing tables.
> I was able to distill an example that demonstrates a substantial difference 
> in memory allocation performance for parquet reads when using 1MB page sizes, 
> that is not present for 64KB pages.
> # Get a machine with at least 32 real cores (not hyperthreaded) and an SSD.
> # Create an Iceberg table with millions of rows containing a moderately long 
> string (hundreds of characters) with a large page size; it's also helpful to 
> create a version with the smaller page size. I used the following with and 
> without {{write.parquet.page-size-bytes}} (iceberg_small_page) specified
> {code:java}
> create table iceberg_large_page stored by iceberg 
> tblproperties('write.parquet.page-size-bytes'='1048576') as select *, 
> repeat(l_comment, 10) from tpch.lineitem;
> insert into iceberg_large_page select *, repeat(l_comment, 10) from 
> tpch.lineitem;
> insert into iceberg_large_page select *, repeat(l_comment, 10) from 
> tpch.lineitem;
> insert into iceberg_large_page select *, repeat(l_comment, 10) from 
> tpch.lineitem;
> insert into iceberg_large_page select *, repeat(l_comment, 10) from 
> tpch.lineitem;{code}
> # Restart Impala with {{-num_io_threads_per_solid_state_disk=32}} to increase 
> read parallelism. The SSD should be able to handle it. The goal is to ensure 
> we have as many scanners attempting to load and decompress data at the same 
> time, with ideally concurrent memory allocation on every thread.
> # Run a query that doesn't process much data outside the scan, and forces 
> Impala to read every entry in the long string column
> {code}
> select _c1 from (select _c1, l_shipdate from iceberg_small_page where _c1 
> like "%toad%" UNION ALL select _c1, l_shipdate from iceberg_small_page where 
> _c1 like "%frog%") x ORDER BY l_shipdate LIMIT 10
> {code}
> I also added IMPALA-13487 to display ParquetDataPagePoolAllocDuration to 
> simplify identifying slow allocation performance. One query was sufficient to 
> show some difference in performance, with sufficient scanner threads to fully 
> utilize all DiskIoMgr threads. The small page query had entries like
> {code}
> ParquetDataPagePoolAllocDuration: (Avg: 20.075us ; Min: 0.000ns ; Max: 
> 65.999ms ; Sum: 2s802ms ; Number of samples: 139620)
> ParquetUncompressedPageSize: (Avg: 65.72 KB (67296) ; Min: 1.37 KB (1406) ; 
> Max: 87.44 KB (89539) ; Sum: 6.14 GB (6590048444) ; Number of samples: 97926)
> {code}
> while the large page query had
> {code}
> ParquetDataPagePoolAllocDuration: (Avg: 2.753ms ; Min: 0.000ns ; Max: 
> 64.999ms ; Sum: 30s346ms ; Number of samples: 11022)
> ParquetUncompressedPageSize: (Avg: 901.89 KB (923535) ; Min: 360.00 B (360) ; 
> Max: 1.00 MB (1048583) ; Sum: 6.14 GB (6597738570) ; Number of samples: 7144)
> {code}
> ParquetUncompressedPageSize shows the difference in page sizes.
> Our theory is that this represents thread contention attempting to access the 
> global pool in tcmalloc. TCMalloc maintains per-thread pools for small 
> amounts of memory - up to 256KB - but for larger chunks malloc goes to a 
> global pool. If that's right, some possible options that could help are
> 1. Try to re-use buffers more across parquet reads, so we don't need to 
> allocate memory as frequently.
> 2. Consider a different memory allocator for larger allocations.
> This likely only impacts very high parallelism read-heavy queries. If each 
> buffer is used in more processing, the cost of allocation should become a 
> smaller part of the query time.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to