Maybe there are differences in default
values on different clusters running different hadoop versions for
USE_STATS_FOR_PARALLELIZATION.
With hadoop 3.1.x and phoenix 5.0 useStatsForParallelization is true by default and the number of splits = guidepost count + number of regions. I changed GUIDE_POST_WIDTH to another value: ALTER TABLE <tablename> SET GUIDE_POSTS_WIDTH = 10240000 UPDATE STATISTICS <tablename> ALL Unfortunately this didn't change the guidepost count and also not the split count. Am I missing something here? Am 30.01.2019 um 19:38 schrieb Thomas
D'Silva:
|
- split count for mapreduce jobs with PhoenixInputFormat Edwin Litterst
- Re: split count for mapreduce jobs with PhoenixIn... Josh Elser
- Re: split count for mapreduce jobs with Phoen... Thomas D'Silva
- Re: split count for mapreduce jobs with P... venkata subbarayudu
- Re: split count for mapreduce jobs wi... Ankit Singhal
- Re: split count for mapreduce jobs wi... Josh Elser
- Re: split count for mapreduce jobs with P... Eddie