[
https://issues.apache.org/jira/browse/PARQUET-2059?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Gabor Szadovszky resolved PARQUET-2059.
---------------------------------------
Resolution: Fixed
> Tests require too much memory
> -----------------------------
>
> Key: PARQUET-2059
> URL: https://issues.apache.org/jira/browse/PARQUET-2059
> Project: Parquet
> Issue Type: Test
> Reporter: Gabor Szadovszky
> Assignee: Gabor Szadovszky
> Priority: Major
>
> For testing the solution of PARQUET-1633 we require ~3GB memory that is not
> always available. To solve this issue we temporarily disabled the implemented
> unit test.
> We need to ensure somehow that [this
> test|https://github.com/apache/parquet-mr/blob/master/parquet-hadoop/src/test/java/org/apache/parquet/hadoop/TestLargeColumnChunk.java]
> (and maybe some other similar ones) are executed regularly. Some options we
> might have:
> * Execute this test separately with a maven profile. I am not sure if the CI
> allows allocating such large memory but with Xmx options we might give a try
> and create a separate check for this test only.
> * Similar to the previous with the profile but not executing in the CI ever.
> Instead, we add some comments to the release doc so this test will be
> executed at least once per release.
> * Configuring the CI profile to skip this test but have it in the normal
> scenario meaning the devs will execute it locally. There are a couple of cons
> though. There is no guarantee that devs executes all the tests including this
> one. It also can cause issues if the dev doesn't have enough memory and don't
> know that the test failure is not related to the current change.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)