[ 
https://issues.apache.org/jira/browse/DRILL-6201?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16382691#comment-16382691
 ] 

Willian Mattos Ribeiro commented on DRILL-6201:
-----------------------------------------------

But Drill is not in cluster. I installed it in a azure virtual machine,
then i tried to configure drill to connect to hive in a cluster that is in
the same network.




> Failed to create input splits: No FileSystem for scheme: maprfs
> ---------------------------------------------------------------
>
>                 Key: DRILL-6201
>                 URL: https://issues.apache.org/jira/browse/DRILL-6201
>             Project: Apache Drill
>          Issue Type: Bug
>          Components: Storage - Hive, Storage - MapRDB
>         Environment: Mapr cluster - CentOS
> Apache Drill installed in other VM (Isn't a cluster node)
>            Reporter: Willian Mattos Ribeiro
>            Priority: Major
>
> 2018-03-01 14:03:28 ERROR HiveMetadataProvider:294 - Failed to create input 
> splits: No FileSystem for scheme: maprfs
> java.io.IOException: No FileSystem for scheme: maprfs
>  at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2644) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2651) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:92) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2687) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2669) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:371) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295) 
> ~[hadoop-common-2.7.1.jar:?]
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider$1.run(HiveMetadataProvider.java:269)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider$1.run(HiveMetadataProvider.java:262)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at java.security.AccessController.doPrivileged(Native Method) ~[?:1.7.0_161]
>  at javax.security.auth.Subject.doAs(Subject.java:421) ~[?:1.7.0_161]
>  at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
>  ~[hadoop-common-2.7.1.jar:?]
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider.splitInputWithUGI(HiveMetadataProvider.java:262)
>  [drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider.getPartitionInputSplits(HiveMetadataProvider.java:154)
>  [drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider.getInputSplits(HiveMetadataProvider.java:176)
>  [drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveScan.getInputSplits(HiveScan.java:122) 
> [drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveScan.getMaxParallelizationWidth(HiveScan.java:171)
>  [drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.physical.ScanPrule.onMatch(ScanPrule.java:41) 
> [drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:228)
>  [calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at 
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:811)
>  [calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:310) 
> [calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform(DefaultSqlHandler.java:400)
>  [drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel(DefaultSqlHandler.java:429)
>  [drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan(DefaultSqlHandler.java:169)
>  [drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan(DrillSqlWorker.java:131)
>  [drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:79)
>  [drill-java-exec-1.12.0.jar:1.12.0]
>  at org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:1017) 
> [drill-java-exec-1.12.0.jar:1.12.0]
>  at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:289) 
> [drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152)
>  [?:1.7.0_161]
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622)
>  [?:1.7.0_161]
>  at java.lang.Thread.run(Thread.java:748) [?:1.7.0_161]
> 2018-03-01 14:03:28 ERROR HiveMetadataProvider:180 - Failed to get InputSplits
> org.apache.drill.common.exceptions.DrillRuntimeException: Failed to create 
> input splits: No FileSystem for scheme: maprfs
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider.splitInputWithUGI(HiveMetadataProvider.java:295)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider.getPartitionInputSplits(HiveMetadataProvider.java:154)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider.getInputSplits(HiveMetadataProvider.java:176)
>  [drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveScan.getInputSplits(HiveScan.java:122) 
> [drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveScan.getMaxParallelizationWidth(HiveScan.java:171)
>  [drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.physical.ScanPrule.onMatch(ScanPrule.java:41) 
> [drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:228)
>  [calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at 
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:811)
>  [calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:310) 
> [calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform(DefaultSqlHandler.java:400)
>  [drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel(DefaultSqlHandler.java:429)
>  [drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan(DefaultSqlHandler.java:169)
>  [drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan(DrillSqlWorker.java:131)
>  [drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:79)
>  [drill-java-exec-1.12.0.jar:1.12.0]
>  at org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:1017) 
> [drill-java-exec-1.12.0.jar:1.12.0]
>  at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:289) 
> [drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152)
>  [?:1.7.0_161]
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622)
>  [?:1.7.0_161]
>  at java.lang.Thread.run(Thread.java:748) [?:1.7.0_161]
> Caused by: java.io.IOException: No FileSystem for scheme: maprfs
>  at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2644) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2651) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:92) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2687) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2669) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:371) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295) 
> ~[hadoop-common-2.7.1.jar:?]
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider$1.run(HiveMetadataProvider.java:269)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider$1.run(HiveMetadataProvider.java:262)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at java.security.AccessController.doPrivileged(Native Method) ~[?:1.7.0_161]
>  at javax.security.auth.Subject.doAs(Subject.java:421) ~[?:1.7.0_161]
>  at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
>  ~[hadoop-common-2.7.1.jar:?]
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider.splitInputWithUGI(HiveMetadataProvider.java:262)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  ... 18 more
> 2018-03-01 14:03:28 ERROR Foreman:593 - SYSTEM ERROR: IOException: No 
> FileSystem for scheme: maprfs
> [Error Id: dc5bcb9a-a350-427d-b443-c60ca9cef294 on 172.28.32.7:31010]
> org.apache.drill.common.exceptions.UserException: SYSTEM ERROR: IOException: 
> No FileSystem for scheme: maprfs
> [Error Id: dc5bcb9a-a350-427d-b443-c60ca9cef294 on 172.28.32.7:31010]
>  at 
> org.apache.drill.common.exceptions.UserException$Builder.build(UserException.java:586)
>  [drill-common-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.work.foreman.Foreman$ForemanResult.close(Foreman.java:801)
>  [drill-java-exec-1.12.0.jar:1.12.0]
>  at org.apache.drill.exec.work.foreman.Foreman.moveToState(Foreman.java:896) 
> [drill-java-exec-1.12.0.jar:1.12.0]
>  at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:306) 
> [drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152)
>  [?:1.7.0_161]
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622)
>  [?:1.7.0_161]
>  at java.lang.Thread.run(Thread.java:748) [?:1.7.0_161]
> Caused by: org.apache.drill.exec.work.foreman.ForemanException: Unexpected 
> exception during fragment initialization: Internal error: Error while 
> applying rule Prel.ScanPrule, args 
> [rel#1208:DrillScanRel.LOGICAL.ANY([]).[](table=[hivenormal, dw, 
> dim_order],groupscan=HiveScan [table=Table(dbName:dw, tableName:dim_order), 
> columns=[`id_order`, `code_delivery`, `code_order`, `id_country`, 
> `id_channel`, `id_store`, `id_campaign`, `id_order_status`, 
> `flag_marketplace`, `order_date`, `update_date`, `nm_country`], 
> numPartitions=1, partitions= [Partition(values:[BR])], 
> inputDirectories=[maprfs:/mapr/az-prd-mapr-cluster/NS_GovernedData/dim_order/nm_country=BR]])]
>  ... 4 more
> Caused by: java.lang.AssertionError: Internal error: Error while applying 
> rule Prel.ScanPrule, args 
> [rel#1208:DrillScanRel.LOGICAL.ANY([]).[](table=[hivenormal, dw, 
> dim_order],groupscan=HiveScan [table=Table(dbName:dw, tableName:dim_order), 
> columns=[`id_order`, `code_delivery`, `code_order`, `id_country`, 
> `id_channel`, `id_store`, `id_campaign`, `id_order_status`, 
> `flag_marketplace`, `order_date`, `update_date`, `nm_country`], 
> numPartitions=1, partitions= [Partition(values:[BR])], 
> inputDirectories=[maprfs:/mapr/az-prd-mapr-cluster/NS_GovernedData/dim_order/nm_country=BR]])]
>  at org.apache.calcite.util.Util.newInternal(Util.java:792) 
> ~[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at 
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:251)
>  ~[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at 
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:811)
>  ~[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:310) 
> ~[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform(DefaultSqlHandler.java:400)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel(DefaultSqlHandler.java:429)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan(DefaultSqlHandler.java:169)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan(DrillSqlWorker.java:131)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:79)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:1017) 
> ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:289) 
> ~[drill-java-exec-1.12.0.jar:1.12.0]
>  ... 3 more
> Caused by: org.apache.drill.common.exceptions.DrillRuntimeException: Failed 
> to get InputSplits
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider.getInputSplits(HiveMetadataProvider.java:181)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveScan.getInputSplits(HiveScan.java:122) 
> ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveScan.getMaxParallelizationWidth(HiveScan.java:171)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.physical.ScanPrule.onMatch(ScanPrule.java:41) 
> ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:228)
>  ~[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at 
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:811)
>  ~[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:310) 
> ~[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform(DefaultSqlHandler.java:400)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel(DefaultSqlHandler.java:429)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan(DefaultSqlHandler.java:169)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan(DrillSqlWorker.java:131)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:79)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:1017) 
> ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:289) 
> ~[drill-java-exec-1.12.0.jar:1.12.0]
>  ... 3 more
> Caused by: org.apache.drill.common.exceptions.DrillRuntimeException: Failed 
> to create input splits: No FileSystem for scheme: maprfs
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider.splitInputWithUGI(HiveMetadataProvider.java:295)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider.getPartitionInputSplits(HiveMetadataProvider.java:154)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider.getInputSplits(HiveMetadataProvider.java:176)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveScan.getInputSplits(HiveScan.java:122) 
> ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveScan.getMaxParallelizationWidth(HiveScan.java:171)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.physical.ScanPrule.onMatch(ScanPrule.java:41) 
> ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:228)
>  ~[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at 
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:811)
>  ~[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:310) 
> ~[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform(DefaultSqlHandler.java:400)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel(DefaultSqlHandler.java:429)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan(DefaultSqlHandler.java:169)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan(DrillSqlWorker.java:131)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:79)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:1017) 
> ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:289) 
> ~[drill-java-exec-1.12.0.jar:1.12.0]
>  ... 3 more
> Caused by: java.io.IOException: No FileSystem for scheme: maprfs
>  at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2644) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2651) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:92) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2687) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2669) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:371) 
> ~[hadoop-common-2.7.1.jar:?]
>  at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295) 
> ~[hadoop-common-2.7.1.jar:?]
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider$1.run(HiveMetadataProvider.java:269)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider$1.run(HiveMetadataProvider.java:262)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at java.security.AccessController.doPrivileged(Native Method) ~[?:1.7.0_161]
>  at javax.security.auth.Subject.doAs(Subject.java:421) ~[?:1.7.0_161]
>  at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
>  ~[hadoop-common-2.7.1.jar:?]
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider.splitInputWithUGI(HiveMetadataProvider.java:262)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider.getPartitionInputSplits(HiveMetadataProvider.java:154)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveMetadataProvider.getInputSplits(HiveMetadataProvider.java:176)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveScan.getInputSplits(HiveScan.java:122) 
> ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.store.hive.HiveScan.getMaxParallelizationWidth(HiveScan.java:171)
>  ~[drill-storage-hive-core-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.physical.ScanPrule.onMatch(ScanPrule.java:41) 
> ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:228)
>  ~[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at 
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:811)
>  ~[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:310) 
> ~[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform(DefaultSqlHandler.java:400)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel(DefaultSqlHandler.java:429)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan(DefaultSqlHandler.java:169)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan(DrillSqlWorker.java:131)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:79)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:1017) 
> ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:289) 
> ~[drill-java-exec-1.12.0.jar:1.12.0]
>  ... 3 more
> 2018-03-01 14:03:28 ERROR QueryResources:91 - Query from Web UI Failed
> org.apache.drill.common.exceptions.UserRemoteException: SYSTEM ERROR: 
> IOException: No FileSystem for scheme: maprfs
> [Error Id: dc5bcb9a-a350-427d-b443-c60ca9cef294 on 172.28.32.7:31010]
> (org.apache.drill.exec.work.foreman.ForemanException) Unexpected exception 
> during fragment initialization: Internal error: Error while applying rule 
> Prel.ScanPrule, args 
> [rel#1208:DrillScanRel.LOGICAL.ANY([]).[](table=[hivenormal, dw, 
> dim_order],groupscan=HiveScan [table=Table(dbName:dw, tableName:dim_order), 
> columns=[`id_order`, `code_delivery`, `code_order`, `id_country`, 
> `id_channel`, `id_store`, `id_campaign`, `id_order_status`, 
> `flag_marketplace`, `order_date`, `update_date`, `nm_country`], 
> numPartitions=1, partitions= [Partition(values:[BR])], 
> inputDirectories=[maprfs:/mapr/az-prd-mapr-cluster/NS_GovernedData/dim_order/nm_country=BR]])]
>  org.apache.drill.exec.work.foreman.Foreman.run():306
>  java.util.concurrent.ThreadPoolExecutor.runWorker():1152
>  java.util.concurrent.ThreadPoolExecutor$Worker.run():622
>  java.lang.Thread.run():748
>  Caused By (java.lang.AssertionError) Internal error: Error while applying 
> rule Prel.ScanPrule, args 
> [rel#1208:DrillScanRel.LOGICAL.ANY([]).[](table=[hivenormal, dw, 
> dim_order],groupscan=HiveScan [table=Table(dbName:dw, tableName:dim_order), 
> columns=[`id_order`, `code_delivery`, `code_order`, `id_country`, 
> `id_channel`, `id_store`, `id_campaign`, `id_order_status`, 
> `flag_marketplace`, `order_date`, `update_date`, `nm_country`], 
> numPartitions=1, partitions= [Partition(values:[BR])], 
> inputDirectories=[maprfs:/mapr/az-prd-mapr-cluster/NS_GovernedData/dim_order/nm_country=BR]])]
>  org.apache.calcite.util.Util.newInternal():792
>  org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch():251
>  org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp():811
>  org.apache.calcite.tools.Programs$RuleSetProgram.run():310
>  org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform():400
>  
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel():429
>  org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan():169
>  org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan():131
>  org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan():79
>  org.apache.drill.exec.work.foreman.Foreman.runSQL():1017
>  org.apache.drill.exec.work.foreman.Foreman.run():289
>  java.util.concurrent.ThreadPoolExecutor.runWorker():1152
>  java.util.concurrent.ThreadPoolExecutor$Worker.run():622
>  java.lang.Thread.run():748
>  Caused By (org.apache.drill.common.exceptions.DrillRuntimeException) Failed 
> to get InputSplits
>  org.apache.drill.exec.store.hive.HiveMetadataProvider.getInputSplits():181
>  org.apache.drill.exec.store.hive.HiveScan.getInputSplits():122
>  org.apache.drill.exec.store.hive.HiveScan.getMaxParallelizationWidth():171
>  org.apache.drill.exec.planner.physical.ScanPrule.onMatch():41
>  org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch():228
>  org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp():811
>  org.apache.calcite.tools.Programs$RuleSetProgram.run():310
>  org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform():400
>  
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel():429
>  org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan():169
>  org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan():131
>  org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan():79
>  org.apache.drill.exec.work.foreman.Foreman.runSQL():1017
>  org.apache.drill.exec.work.foreman.Foreman.run():289
>  java.util.concurrent.ThreadPoolExecutor.runWorker():1152
>  java.util.concurrent.ThreadPoolExecutor$Worker.run():622
>  java.lang.Thread.run():748
>  Caused By (org.apache.drill.common.exceptions.DrillRuntimeException) Failed 
> to create input splits: No FileSystem for scheme: maprfs
>  org.apache.drill.exec.store.hive.HiveMetadataProvider.splitInputWithUGI():295
>  
> org.apache.drill.exec.store.hive.HiveMetadataProvider.getPartitionInputSplits():154
>  org.apache.drill.exec.store.hive.HiveMetadataProvider.getInputSplits():176
>  org.apache.drill.exec.store.hive.HiveScan.getInputSplits():122
>  org.apache.drill.exec.store.hive.HiveScan.getMaxParallelizationWidth():171
>  org.apache.drill.exec.planner.physical.ScanPrule.onMatch():41
>  org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch():228
>  org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp():811
>  org.apache.calcite.tools.Programs$RuleSetProgram.run():310
>  org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform():400
>  
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel():429
>  org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan():169
>  org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan():131
>  org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan():79
>  org.apache.drill.exec.work.foreman.Foreman.runSQL():1017
>  org.apache.drill.exec.work.foreman.Foreman.run():289
>  java.util.concurrent.ThreadPoolExecutor.runWorker():1152
>  java.util.concurrent.ThreadPoolExecutor$Worker.run():622
>  java.lang.Thread.run():748
>  Caused By (java.io.IOException) No FileSystem for scheme: maprfs
>  org.apache.hadoop.fs.FileSystem.getFileSystemClass():2644
>  org.apache.hadoop.fs.FileSystem.createFileSystem():2651
>  org.apache.hadoop.fs.FileSystem.access$200():92
>  org.apache.hadoop.fs.FileSystem$Cache.getInternal():2687
>  org.apache.hadoop.fs.FileSystem$Cache.get():2669
>  org.apache.hadoop.fs.FileSystem.get():371
>  org.apache.hadoop.fs.Path.getFileSystem():295
>  org.apache.drill.exec.store.hive.HiveMetadataProvider$1.run():269
>  org.apache.drill.exec.store.hive.HiveMetadataProvider$1.run():262
>  java.security.AccessController.doPrivileged():-2
>  javax.security.auth.Subject.doAs():421
>  org.apache.hadoop.security.UserGroupInformation.doAs():1657
>  org.apache.drill.exec.store.hive.HiveMetadataProvider.splitInputWithUGI():262
>  
> org.apache.drill.exec.store.hive.HiveMetadataProvider.getPartitionInputSplits():154
>  org.apache.drill.exec.store.hive.HiveMetadataProvider.getInputSplits():176
>  org.apache.drill.exec.store.hive.HiveScan.getInputSplits():122
>  org.apache.drill.exec.store.hive.HiveScan.getMaxParallelizationWidth():171
>  org.apache.drill.exec.planner.physical.ScanPrule.onMatch():41
>  org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch():228
>  org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp():811
>  org.apache.calcite.tools.Programs$RuleSetProgram.run():310
>  org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform():400
>  
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel():429
>  org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan():169
>  org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan():131
>  org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan():79
>  org.apache.drill.exec.work.foreman.Foreman.runSQL():1017
>  org.apache.drill.exec.work.foreman.Foreman.run():289
>  java.util.concurrent.ThreadPoolExecutor.runWorker():1152
>  java.util.concurrent.ThreadPoolExecutor$Worker.run():622
>  java.lang.Thread.run():748
> at 
> org.apache.drill.exec.rpc.AbstractDisposableUserClientConnection.sendResult(AbstractDisposableUserClientConnection.java:85)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> org.apache.drill.exec.work.foreman.Foreman$ForemanResult.close(Foreman.java:822)
>  ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at org.apache.drill.exec.work.foreman.Foreman.moveToState(Foreman.java:896) 
> ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:306) 
> ~[drill-java-exec-1.12.0.jar:1.12.0]
>  at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152)
>  ~[?:1.7.0_161]
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622)
>  ~[?:1.7.0_161]
>  at java.lang.Thread.run(Thread.java:748) [?:1.7.0_161]
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to