Hi Herman,

I am not sure what the exact problem here is but can you check to see if
you are not hitting the problem described here:

http://mail-archives.apache.org/mod_mbox/drill-user/201805.mbox/%3ccacwrgnexlxop2vcyugsa4gwd1jgs8f+rcpzq8rhuatfw5fm...@mail.gmail.com%3E

Thanks

Karthik

On Wed, Sep 19, 2018 at 7:02 PM Herman Tan <[email protected]> wrote:

> Hi,
>
> I encountered the following error.
> The Steps I did are as follows:
> 1. Create a view to fix the data type of fields with cast
> 2. Create table (parquet) using the view
> 3. Query select * from table (query a field also does not work)
>
> The error:
> SQL Error: RESOURCE ERROR: Waited for 30000 ms, but only 10 tasks for
> 'Fetch parquet metadata' are complete. Total number of tasks 29,
> parallelism 16.
>
> When I re-run this, the number of tasks will vary.
>
> What could be the problem?
>
> Regards,
> Herman Tan
>
> More info below:
>
> This is the folders of the files
> Total # of lines, 50 million
> ----------
> show files from
> dfs.`D:\retail_sandbox\pos\sales_pos_detail\pos_details_20180825`
> ;
> show files from
> dfs.`D:\retail_sandbox\pos\sales_pos_detail\pos_details_20180825\2011`
> ;
> -----
> sales_pos_detail
>   \pos_details_20180825
>     \2007
>     \2008
>     \2009
>     \2010
>     \2011
>   \pos_details_0.csv
>   \pos_details_1.csv
>   \pos_details_2.csv
>   \pos_details_3.csv
>   \pos_details_4.csv
>   \pos_details_5.csv
>   \pos_details_6.csv
>   \pos_details_7.csv
>   \pos_details_8.csv
>     \2012
>     \2013
>     \2014
>     \2015
>     \2016
>     \2017
>     \2018
>     \others
> -----
>
> create or replace view dfs.tmp.load_pos_sales_detail as
> SELECT
> -- dimension keys
>  cast(dim_date_key as int) dim_date_key
> ,cast(dim_site_key as int) dim_site_key
> ,cast(dim_pos_header_key as bigint) dim_pos_header_key
> ,cast(dim_pos_cashier_key as int) dim_pos_cashier_key
> ,cast(dim_card_number_key as int) dim_card_number_key
> ,cast(dim_hour_minute_key as int) dim_hour_minute_key
> ,cast(dim_pos_clerk_key as int) dim_pos_clerk_key
> ,cast(dim_product_key as int) dim_product_key
> ,cast(dim_pos_employee_purchase_key as int) dim_pos_employee_purchase_key
> ,cast(dim_pos_terminal_key as int) dim_pos_terminal_key
> ,cast(dim_campaign_key as int) dim_campaign_key
> ,cast(dim_promo_key as int) dim_promo_key
> ,cast( case when dim_site_lfl_key = '' then 0 else dim_site_lfl_key end as
> int) dim_site_lfl_key
> -- derived from keys
> ,dim_date_str
> ,`year` as `trx_year`
> -- Measures
> ,Product_Sales_Qty
> ,Product_Sales_Price
> ,Product_Cost_Price
> ,Product_Cost_Amt
> ,Product_Sales_Gross_Amt
> ,Product_Sales_Promo_Disc_Amt
> ,Product_Sales_Add_Promo_Disc_Amt
> ,Product_Sales_Total_Promo_Disc_Amt
> ,Product_Sales_Retail_Promo_Amt
> ,Product_Sales_Retail_Amt
> ,Product_Sales_VAT_Amt
> ,Product_Sales_Product_Margin_Amt
> ,Product_Sales_Initial_Margin_Amt
> from dfs.`D:\retail_sandbox\pos\sales_pos_detail\pos_details_20180825`
> ;
>
> drop table if exists dfs.tmp.load_pos_sales_detail_tbl
> ;
>
> create table dfs.tmp.load_pos_sales_detail_tbl AS
> SELECT
> -- dimension keys
>  dim_date_key
> ,dim_site_key
> ,dim_pos_header_key
> ,dim_pos_cashier_key
> ,dim_card_number_key
> ,dim_hour_minute_key
> ,dim_pos_clerk_key
> ,dim_product_key
> ,dim_pos_employee_purchase_key
> ,dim_pos_terminal_key
> ,dim_campaign_key
> ,dim_promo_key
> ,dim_site_lfl_key
> -- derived from keys
> ,dim_date_str
> ,`trx_year`
> -- Measures
> ,Product_Sales_Qty Sales_Qty
> ,Product_Sales_Price Sales_Price
> ,Product_Cost_Price Cost_Price
> ,Product_Cost_Amt Cost_Amt
> ,Product_Sales_Gross_Amt Sales_Gross_Amt
> ,Product_Sales_Promo_Disc_Amt Sales_Promo_Disc_Amt
> ,Product_Sales_Add_Promo_Disc_Amt Add_Promo_Disc_Amt
> ,Product_Sales_Total_Promo_Disc_Amt Total_Promo_Disc_Amt
> ,Product_Sales_Retail_Promo_Amt Retail_Promo_Amt
> ,Product_Sales_Retail_Amt Retail_Amt
> ,Product_Sales_VAT_Amt VAT_Amt
> ,Product_Sales_Product_Margin_Amt Product_Margin_Amt
> ,Product_Sales_Initial_Margin_Amt Initial_Margin_Amt
> from dfs.tmp.load_pos_sales_detail
> ;
>
> select *
> from dfs.tmp.load_pos_sales_detail_tbl
> ;
>
> ----- ERROR ----------------------------
>
> SQL Error: RESOURCE ERROR: Waited for 30000 ms, but only 10 tasks for
> 'Fetch parquet metadata' are complete. Total number of tasks 29,
> parallelism 16.
>
>
> [Error Id: 3b079174-f5d0-4313-8097-25a0b3070854 on
> IORA-G9KY9P2.stf.nus.edu.sg:31010]
>   RESOURCE ERROR: Waited for 30000 ms, but only 10 tasks for 'Fetch parquet
> metadata' are complete. Total number of tasks 29, parallelism 16.
>
>
> [Error Id: 3b079174-f5d0-4313-8097-25a0b3070854 on
> IORA-G9KY9P2.stf.nus.edu.sg:31010]
>     RESOURCE ERROR: Waited for 30000 ms, but only 10 tasks for 'Fetch
> parquet metadata' are complete. Total number of tasks 29, parallelism 16.
>
>
> [Error Id: 3b079174-f5d0-4313-8097-25a0b3070854 on
> IORA-G9KY9P2.stf.nus.edu.sg:31010]
>     RESOURCE ERROR: Waited for 30000 ms, but only 10 tasks for 'Fetch
> parquet metadata' are complete. Total number of tasks 29, parallelism 16.
>
>
> [Error Id: 3b079174-f5d0-4313-8097-25a0b3070854 on
> IORA-G9KY9P2.stf.nus.edu.sg:31010]
>
> ----------------------------------------
> From Drill log:
>
> 2018-09-20 08:58:12,035 [245d0f5a-ae5f-bfa2-ff04-40f7bdd1c2bf:foreman]
> INFO  o.a.drill.exec.work.foreman.Foreman - Query text for query id
> 245d0f5a-ae5f-bfa2-ff04-40f7bdd1c2bf: select *
> from dfs.tmp.load_pos_sales_detail_tbl
>
> 2018-09-20 08:58:53,068 [245d0f5a-ae5f-bfa2-ff04-40f7bdd1c2bf:foreman]
> ERROR o.a.d.e.s.parquet.metadata.Metadata - Waited for 30000 ms, but only
> 10 tasks for 'Fetch parquet metadata' are complete. Total number of tasks
> 29, parallelism 16.
> java.util.concurrent.CancellationException: null
> at
>
> org.apache.drill.exec.store.TimedCallable$FutureMapper.apply(TimedCallable.java:86)
> ~[drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.TimedCallable$FutureMapper.apply(TimedCallable.java:57)
> ~[drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.common.collections.Collectors.lambda$toList$2(Collectors.java:97)
> ~[drill-common-1.14.0.jar:1.14.0]
> at java.util.ArrayList.forEach(ArrayList.java:1257) ~[na:1.8.0_172]
> at
> org.apache.drill.common.collections.Collectors.toList(Collectors.java:97)
> ~[drill-common-1.14.0.jar:1.14.0]
> at org.apache.drill.exec.store.TimedCallable.run(TimedCallable.java:214)
> ~[drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.metadata.Metadata.getParquetFileMetadata_v3(Metadata.java:340)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.metadata.Metadata.getParquetTableMetadata(Metadata.java:324)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.metadata.Metadata.getParquetTableMetadata(Metadata.java:305)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.metadata.Metadata.getParquetTableMetadata(Metadata.java:124)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetGroupScan.initInternal(ParquetGroupScan.java:254)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.AbstractParquetGroupScan.init(AbstractParquetGroupScan.java:380)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetGroupScan.<init>(ParquetGroupScan.java:132)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetGroupScan.<init>(ParquetGroupScan.java:102)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(ParquetFormatPlugin.java:180)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(ParquetFormatPlugin.java:70)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.dfs.FileSystemPlugin.getPhysicalScan(FileSystemPlugin.java:136)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(AbstractStoragePlugin.java:116)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(AbstractStoragePlugin.java:111)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillTable.getGroupScan(DrillTable.java:99)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillScanRel.<init>(DrillScanRel.java:89)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillScanRel.<init>(DrillScanRel.java:69)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillScanRel.<init>(DrillScanRel.java:62)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillScanRule.onMatch(DrillScanRule.java:38)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:212)
> [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> at
>
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:652)
> [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> at org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:368)
> [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform(DefaultSqlHandler.java:429)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform(DefaultSqlHandler.java:369)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToRawDrel(DefaultSqlHandler.java:255)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToDrel(DefaultSqlHandler.java:318)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan(DefaultSqlHandler.java:180)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan(DrillSqlWorker.java:145)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:83)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:567)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:266)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> [na:1.8.0_172]
> at
>
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> [na:1.8.0_172]
> at java.lang.Thread.run(Thread.java:748) [na:1.8.0_172]
> 2018-09-20 08:58:53,080 [245d0f5a-ae5f-bfa2-ff04-40f7bdd1c2bf:foreman]
> INFO  o.a.d.e.s.parquet.metadata.Metadata - User Error Occurred: Waited for
> 30000 ms, but only 10 tasks for 'Fetch parquet metadata' are complete.
> Total number of tasks 29, parallelism 16. (null)
> org.apache.drill.common.exceptions.UserException: RESOURCE ERROR: Waited
> for 30000 ms, but only 10 tasks for 'Fetch parquet metadata' are complete.
> Total number of tasks 29, parallelism 16.
>
>
> [Error Id: f887dcae-9f55-469c-be52-b6ce2a37eeb0 ]
> at
>
> org.apache.drill.common.exceptions.UserException$Builder.build(UserException.java:633)
> ~[drill-common-1.14.0.jar:1.14.0]
> at org.apache.drill.exec.store.TimedCallable.run(TimedCallable.java:253)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.metadata.Metadata.getParquetFileMetadata_v3(Metadata.java:340)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.metadata.Metadata.getParquetTableMetadata(Metadata.java:324)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.metadata.Metadata.getParquetTableMetadata(Metadata.java:305)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.metadata.Metadata.getParquetTableMetadata(Metadata.java:124)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetGroupScan.initInternal(ParquetGroupScan.java:254)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.AbstractParquetGroupScan.init(AbstractParquetGroupScan.java:380)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetGroupScan.<init>(ParquetGroupScan.java:132)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetGroupScan.<init>(ParquetGroupScan.java:102)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(ParquetFormatPlugin.java:180)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(ParquetFormatPlugin.java:70)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.dfs.FileSystemPlugin.getPhysicalScan(FileSystemPlugin.java:136)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(AbstractStoragePlugin.java:116)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(AbstractStoragePlugin.java:111)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillTable.getGroupScan(DrillTable.java:99)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillScanRel.<init>(DrillScanRel.java:89)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillScanRel.<init>(DrillScanRel.java:69)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillScanRel.<init>(DrillScanRel.java:62)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillScanRule.onMatch(DrillScanRule.java:38)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:212)
> [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> at
>
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:652)
> [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> at org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:368)
> [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform(DefaultSqlHandler.java:429)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform(DefaultSqlHandler.java:369)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToRawDrel(DefaultSqlHandler.java:255)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToDrel(DefaultSqlHandler.java:318)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan(DefaultSqlHandler.java:180)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan(DrillSqlWorker.java:145)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:83)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:567)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:266)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> [na:1.8.0_172]
> at
>
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> [na:1.8.0_172]
> at java.lang.Thread.run(Thread.java:748) [na:1.8.0_172]
> Caused by: java.util.concurrent.CancellationException: null
> at
>
> org.apache.drill.exec.store.TimedCallable$FutureMapper.apply(TimedCallable.java:86)
> ~[drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.TimedCallable$FutureMapper.apply(TimedCallable.java:57)
> ~[drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.common.collections.Collectors.lambda$toList$2(Collectors.java:97)
> ~[drill-common-1.14.0.jar:1.14.0]
> at java.util.ArrayList.forEach(ArrayList.java:1257) ~[na:1.8.0_172]
> at
> org.apache.drill.common.collections.Collectors.toList(Collectors.java:97)
> ~[drill-common-1.14.0.jar:1.14.0]
> at org.apache.drill.exec.store.TimedCallable.run(TimedCallable.java:214)
> [drill-java-exec-1.14.0.jar:1.14.0]
> ... 33 common frames omitted
> 2018-09-20 09:02:10,608 [UserServer-1] WARN
> o.a.drill.exec.rpc.user.UserServer - Message of mode REQUEST of rpc type 3
> took longer than 500ms.  Actual duration was 2042ms.
> 2018-09-20 09:02:10,608 [245d0e6f-0dc1-2a4b-12a4-b9aaad4182fc:foreman]
> INFO  o.a.drill.exec.work.foreman.Foreman - Query text for query id
> 245d0e6f-0dc1-2a4b-12a4-b9aaad4182fc: select *
> from dfs.tmp.load_pos_sales_detail_tbl
>
> 2018-09-20 09:02:42,615 [245d0e6f-0dc1-2a4b-12a4-b9aaad4182fc:foreman]
> ERROR o.a.d.e.s.parquet.metadata.Metadata - Waited for 30000 ms, but only
> 10 tasks for 'Fetch parquet metadata' are complete. Total number of tasks
> 29, parallelism 16.
> java.util.concurrent.CancellationException: null
> at
>
> org.apache.drill.exec.store.TimedCallable$FutureMapper.apply(TimedCallable.java:86)
> ~[drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.TimedCallable$FutureMapper.apply(TimedCallable.java:57)
> ~[drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.common.collections.Collectors.lambda$toList$2(Collectors.java:97)
> ~[drill-common-1.14.0.jar:1.14.0]
> at java.util.ArrayList.forEach(ArrayList.java:1257) ~[na:1.8.0_172]
> at
> org.apache.drill.common.collections.Collectors.toList(Collectors.java:97)
> ~[drill-common-1.14.0.jar:1.14.0]
> at org.apache.drill.exec.store.TimedCallable.run(TimedCallable.java:214)
> ~[drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.metadata.Metadata.getParquetFileMetadata_v3(Metadata.java:340)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.metadata.Metadata.getParquetTableMetadata(Metadata.java:324)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.metadata.Metadata.getParquetTableMetadata(Metadata.java:305)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.metadata.Metadata.getParquetTableMetadata(Metadata.java:124)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetGroupScan.initInternal(ParquetGroupScan.java:254)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.AbstractParquetGroupScan.init(AbstractParquetGroupScan.java:380)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetGroupScan.<init>(ParquetGroupScan.java:132)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetGroupScan.<init>(ParquetGroupScan.java:102)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(ParquetFormatPlugin.java:180)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(ParquetFormatPlugin.java:70)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.dfs.FileSystemPlugin.getPhysicalScan(FileSystemPlugin.java:136)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(AbstractStoragePlugin.java:116)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(AbstractStoragePlugin.java:111)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillTable.getGroupScan(DrillTable.java:99)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillScanRel.<init>(DrillScanRel.java:89)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillScanRel.<init>(DrillScanRel.java:69)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillScanRel.<init>(DrillScanRel.java:62)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillScanRule.onMatch(DrillScanRule.java:38)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:212)
> [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> at
>
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:652)
> [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> at org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:368)
> [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform(DefaultSqlHandler.java:429)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform(DefaultSqlHandler.java:369)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToRawDrel(DefaultSqlHandler.java:255)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToDrel(DefaultSqlHandler.java:318)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan(DefaultSqlHandler.java:180)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan(DrillSqlWorker.java:145)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:83)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:567)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:266)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> [na:1.8.0_172]
> at
>
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> [na:1.8.0_172]
> at java.lang.Thread.run(Thread.java:748) [na:1.8.0_172]
> 2018-09-20 09:02:42,625 [245d0e6f-0dc1-2a4b-12a4-b9aaad4182fc:foreman]
> INFO  o.a.d.e.s.parquet.metadata.Metadata - User Error Occurred: Waited for
> 30000 ms, but only 10 tasks for 'Fetch parquet metadata' are complete.
> Total number of tasks 29, parallelism 16. (null)
> org.apache.drill.common.exceptions.UserException: RESOURCE ERROR: Waited
> for 30000 ms, but only 10 tasks for 'Fetch parquet metadata' are complete.
> Total number of tasks 29, parallelism 16.
>
>
> [Error Id: 3b079174-f5d0-4313-8097-25a0b3070854 ]
> at
>
> org.apache.drill.common.exceptions.UserException$Builder.build(UserException.java:633)
> ~[drill-common-1.14.0.jar:1.14.0]
> at org.apache.drill.exec.store.TimedCallable.run(TimedCallable.java:253)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.metadata.Metadata.getParquetFileMetadata_v3(Metadata.java:340)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.metadata.Metadata.getParquetTableMetadata(Metadata.java:324)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.metadata.Metadata.getParquetTableMetadata(Metadata.java:305)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.metadata.Metadata.getParquetTableMetadata(Metadata.java:124)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetGroupScan.initInternal(ParquetGroupScan.java:254)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.AbstractParquetGroupScan.init(AbstractParquetGroupScan.java:380)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetGroupScan.<init>(ParquetGroupScan.java:132)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetGroupScan.<init>(ParquetGroupScan.java:102)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(ParquetFormatPlugin.java:180)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(ParquetFormatPlugin.java:70)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.dfs.FileSystemPlugin.getPhysicalScan(FileSystemPlugin.java:136)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(AbstractStoragePlugin.java:116)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(AbstractStoragePlugin.java:111)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillTable.getGroupScan(DrillTable.java:99)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillScanRel.<init>(DrillScanRel.java:89)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillScanRel.<init>(DrillScanRel.java:69)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillScanRel.<init>(DrillScanRel.java:62)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.logical.DrillScanRule.onMatch(DrillScanRule.java:38)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:212)
> [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> at
>
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:652)
> [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> at org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:368)
> [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform(DefaultSqlHandler.java:429)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform(DefaultSqlHandler.java:369)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToRawDrel(DefaultSqlHandler.java:255)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToDrel(DefaultSqlHandler.java:318)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan(DefaultSqlHandler.java:180)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan(DrillSqlWorker.java:145)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:83)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:567)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:266)
> [drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> [na:1.8.0_172]
> at
>
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> [na:1.8.0_172]
> at java.lang.Thread.run(Thread.java:748) [na:1.8.0_172]
> Caused by: java.util.concurrent.CancellationException: null
> at
>
> org.apache.drill.exec.store.TimedCallable$FutureMapper.apply(TimedCallable.java:86)
> ~[drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.exec.store.TimedCallable$FutureMapper.apply(TimedCallable.java:57)
> ~[drill-java-exec-1.14.0.jar:1.14.0]
> at
>
> org.apache.drill.common.collections.Collectors.lambda$toList$2(Collectors.java:97)
> ~[drill-common-1.14.0.jar:1.14.0]
> at java.util.ArrayList.forEach(ArrayList.java:1257) ~[na:1.8.0_172]
> at
> org.apache.drill.common.collections.Collectors.toList(Collectors.java:97)
> ~[drill-common-1.14.0.jar:1.14.0]
> at org.apache.drill.exec.store.TimedCallable.run(TimedCallable.java:214)
> [drill-java-exec-1.14.0.jar:1.14.0]
> ... 33 common frames omitted
>
>
> ----------------------------------------
> ----------
>

Reply via email to