Hi,

I am trying to run demo of Drill 1.0 according to following instructions:
https://cwiki.apache.org/confluence/display/DRILL/Demo+HowTo (file version: Sep 
02, 2013)
All is clear till the moment of 'Run a Query', when I'm pasting query:

"select * from "sample-data/region.parquet";"

from content of HowTo and then query fails with "java.io.FileNotFoundException: 
File sample-data/region.parquet does not exist.".
It seams to be lack of parquet file. Should I get it from somewhere?
Info from console after run query:

0: jdbc:drill:schema=parquet-local> select * from "sample-data/region.parquet";
16:19:04.185 [main] DEBUG o.a.drill.optiq.EnumerableDrillRel - implementing 
enumerable
16:19:04.213 [main] DEBUG o.a.drill.optiq.DrillImplementor - Optiq Generated 
Logical Plan: {
  "head" : {
    "type" : "APACHE_DRILL_LOGICAL",
    "version" : "1",
    "generator" : {
      "type" : "optiq",
      "info" : "na"
    }
  },
  "storage" : {
    "parquet-local" : {
      "type" : "parquet",
      "dfsName" : "file:///"
    }
  },
  "query" : [ {
    "op" : "scan",
    "memo" : "initial_scan",
    "ref" : "_MAP",
    "storageengine" : "parquet-local",
    "selection" : [ {
      "path" : "sample-data/region.parquet"
    } ],
    "@id" : 1
  }, {
    "op" : "store",
    "input" : 1,
    "storageengine" : "queue",
    "memo" : "output sink",
    "target" : {
      "number" : 0
    },
    "@id" : 2
  } ]
}
16:19:04.273 [main] DEBUG o.a.drill.common.config.DrillConfig - Loading configs 
at the following URLs 
[jar:file:/home/user/.m2/repository/org/apache/drill/exec/ref/1.0-SNAPSHOT/ref-1.0-SNAPSHOT.jar!/drill-module.conf,
 
jar:file:/home/user/.m2/repository/org/apache/drill/exec/java-exec/1.0-SNAPSHOT/java-exec-1.0-SNAPSHOT.jar!/drill-module.conf,
 
jar:file:/home/user/.m2/repository/org/apache/drill/common/1.0-SNAPSHOT/common-1.0-SNAPSHOT.jar!/drill-module.conf]
16:19:04.280 [main] DEBUG o.a.d.c.l.data.LogicalOperatorBase - Adding Logical 
Operator sub types: [class org.apache.drill.common.logical.data.Sequence, class 
org.apache.drill.common.logical.data.Transform, class 
org.apache.drill.common.logical.data.Flatten, class 
org.apache.drill.common.logical.data.Store, class 
org.apache.drill.common.logical.data.Filter, class 
org.apache.drill.common.logical.data.Order, class 
org.apache.drill.common.logical.data.WindowFrame, class 
org.apache.drill.common.logical.data.Segment, class 
org.apache.drill.common.logical.data.Join, class 
org.apache.drill.common.logical.data.Union, class 
org.apache.drill.common.logical.data.Limit, class 
org.apache.drill.common.logical.data.Constant, class 
org.apache.drill.common.logical.data.Project, class 
org.apache.drill.common.logical.data.Scan, class 
org.apache.drill.common.logical.data.RunningAggregate, class 
org.apache.drill.common.logical.data.CollapsingAggregate]
16:19:04.281 [main] DEBUG o.a.d.c.l.StorageEngineConfigBase - Adding Storage 
Engine Configs including [class 
org.apache.drill.exec.ref.rse.FileSystemRSE$FileSystemRSEConfig, class 
org.apache.drill.exec.ref.rse.ConsoleRSE$ConsoleRSEConfig, class 
org.apache.drill.exec.ref.rse.QueueRSE$QueueRSEConfig, class 
org.apache.drill.exec.store.json.JSONStorageEngineConfig, class 
org.apache.drill.exec.store.parquet.ParquetStorageEngineConfig, class 
org.apache.drill.exec.store.mock.MockStorageEngineConfig, class 
org.apache.drill.exec.ref.rse.ClasspathRSE$ClasspathRSEConfig]
16:19:04.283 [main] DEBUG o.a.d.sql.client.full.BatchLoaderMap - Starting next 
query result batch.
16:19:04.288 [WorkManager Event Thread] DEBUG 
o.apache.drill.exec.work.WorkManager - Starting pending task 
org.apache.drill.exec.work.foreman.Foreman@297e4571
16:19:04.289 [Client-1] DEBUG o.a.d.e.rpc.user.QueryResultHandler - Received 
QueryId part1: -8933707500775258587
part2: -7304718200146300837
succesfully.  Adding listener 
org.apache.drill.sql.client.full.BatchListener@25d2076d
16:19:04.289 [WorkManager-4] DEBUG o.a.d.e.planner.PhysicalPlanReader - Reading 
logical plan 
{"head":{"type":"APACHE_DRILL_LOGICAL","version":"1","generator":{"type":"optiq","info":"na"}},"storage":{"parquet-local":{"type":"parquet","dfsName":"file:///"}},"query":[{"op":"scan","memo":"initial_scan","ref":"_MAP","storageengine":"parquet-local","selection":[{"path":"sample-data/region.parquet"}],"@id":1},{"op":"store","input":1,"storageengine":"queue","memo":"output
 sink","target":{"number":0},"@id":2}]}
16:19:04.290 [WorkManager-4] DEBUG o.a.drill.exec.work.foreman.Foreman - 
Logical {
  head : {
    type : "APACHE_DRILL_LOGICAL",
    version : 1,
    generator : {
      type : "optiq",
      info : "na"
    }
  },
  storage : {
    parquet-local : {
      type : "parquet",
      dfsName : "file:///"
    }
  },
  query : [ {
    op : "scan",
    @id : 1,
    memo : "initial_scan",
   storageengine : "parquet-local",
    selection : [ {
      path : "sample-data/region.parquet"
    } ],
    ref : "_MAP"
  }, {
    op : "store",
    @id : 2,
    memo : "output sink",
    input : 1,
    target : {
      number : 0
    },
    partition : null,
    storageEngine : "queue"
  } ]
}
16:19:04.296 [WorkManager-4] DEBUG o.a.drill.common.config.DrillConfig - 
Loading configs at the following URLs 
[jar:file:/home/user/.m2/repository/org/apache/drill/exec/ref/1.0-SNAPSHOT/ref-1.0-SNAPSHOT.jar!/drill-module.conf,
 
jar:file:/home/user/.m2/repository/org/apache/drill/exec/java-exec/1.0-SNAPSHOT/java-exec-1.0-SNAPSHOT.jar!/drill-module.conf,
 
jar:file:/home/user/.m2/repository/org/apache/drill/common/1.0-SNAPSHOT/common-1.0-SNAPSHOT.jar!/drill-module.conf]
16:19:04.317 [WorkManager-4] DEBUG o.a.d.c.l.data.LogicalOperatorBase - Adding 
Logical Operator sub types: [class 
org.apache.drill.common.logical.data.Sequence, class 
org.apache.drill.common.logical.data.Transform, class 
org.apache.drill.common.logical.data.Flatten, class 
org.apache.drill.common.logical.data.Store, class 
org.apache.drill.common.logical.data.Filter, class 
org.apache.drill.common.logical.data.Order, class 
org.apache.drill.common.logical.data.WindowFrame, class 
org.apache.drill.common.logical.data.Segment, class 
org.apache.drill.common.logical.data.Join, class 
org.apache.drill.common.logical.data.Union, class 
org.apache.drill.common.logical.data.Limit, class 
org.apache.drill.common.logical.data.Constant, class 
org.apache.drill.common.logical.data.Project, class 
org.apache.drill.common.logical.data.Scan, class 
org.apache.drill.common.logical.data.RunningAggregate, class 
org.apache.drill.common.logical.data.CollapsingAggregate]
16:19:04.320 [WorkManager-4] DEBUG o.a.d.c.l.StorageEngineConfigBase - Adding 
Storage Engine Configs including [class 
org.apache.drill.exec.ref.rse.FileSystemRSE$FileSystemRSEConfig, class 
org.apache.drill.exec.ref.rse.ConsoleRSE$ConsoleRSEConfig, class 
org.apache.drill.exec.ref.rse.QueueRSE$QueueRSEConfig, class 
org.apache.drill.exec.store.json.JSONStorageEngineConfig, class 
org.apache.drill.exec.store.parquet.ParquetStorageEngineConfig, class 
org.apache.drill.exec.store.mock.MockStorageEngineConfig, class 
org.apache.drill.exec.ref.rse.ClasspathRSE$ClasspathRSEConfig]
16:19:04.352 [WorkManager-4] ERROR o.a.drill.exec.work.foreman.Foreman - Error 
4962bfd0-316a-43a7-af9f-9826a096c3c7: Failure while converting logical plan to 
physical plan.
java.io.FileNotFoundException: File sample-data/region.parquet does not exist.
                at 
org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:402)
 ~[hadoop-core-1.2.1.jar:na]
                at 
org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:255) 
~[hadoop-core-1.2.1.jar:na]
                at 
parquet.hadoop.ParquetFileReader.readFooters(ParquetFileReader.java:234) 
~[parquet-hadoop-1.0.1.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.exec.store.parquet.ParquetGroupScan.readFooter(ParquetGroupScan.java:131)
 ~[java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.exec.store.parquet.ParquetGroupScan.<init>(ParquetGroupScan.java:119)
 ~[java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.exec.store.parquet.ParquetStorageEngine.getPhysicalScan(ParquetStorageEngine.java:97)
 ~[java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.exec.store.parquet.ParquetStorageEngine.getPhysicalScan(ParquetStorageEngine.java:55)
 ~[java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.exec.opt.BasicOptimizer$LogicalConverter.visitScan(BasicOptimizer.java:189)
 ~[java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.exec.opt.BasicOptimizer$LogicalConverter.visitScan(BasicOptimizer.java:96)
 ~[java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.common.logical.data.Scan.accept(Scan.java:59) 
~[common-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.exec.opt.BasicOptimizer$LogicalConverter.visitStore(BasicOptimizer.java:200)
 ~[java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.exec.opt.BasicOptimizer$LogicalConverter.visitStore(BasicOptimizer.java:96)
 ~[java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.common.logical.data.Store.accept(Store.java:61) 
~[common-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.exec.opt.BasicOptimizer.optimize(BasicOptimizer.java:70) 
~[java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.exec.work.foreman.Foreman.convert(Foreman.java:242) 
[java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.exec.work.foreman.Foreman.parseAndRunLogicalPlan(Foreman.java:174)
 [java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:153) 
[java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) 
~[na:1.7.0_25]
                at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) 
~[na:1.7.0_25]
                at java.lang.Thread.run(Thread.java:724) ~[na:1.7.0_25]
16:19:04.354 [WorkManager-4] DEBUG o.a.drill.exec.rpc.user.UserServer - Sending 
result to client with QueryWritableBatch [header=query_state: FAILED
query_id {
  part1: -8933707500775258587
  part2: -7304718200146300837
}
is_last_chunk: true
error {
  error_id: "4962bfd0-316a-43a7-af9f-9826a096c3c7"
  endpoint {
    address: "user-VirtualBox"
    user_port: 31010
    bit_port: 31011
  }
  error_type: 0
  message: "Failure while converting logical plan to physical plan. < 
OptimizerException:[ Failure while attempting to retrieve storage engine. ] < 
FileNotFoundException:[ File sample-data/region.parquet does not exist. ]"
}
, buffers=[]]
16:19:04.387 [main] ERROR o.a.d.s.client.full.ResultEnumerator - Exception 
during query
org.apache.drill.exec.rpc.RpcException: Remote failure while running 
query.[error_id: "4962bfd0-316a-43a7-af9f-9826a096c3c7"
endpoint {
  address: "user-VirtualBox"
  user_port: 31010
  bit_port: 31011
}
error_type: 0
message: "Failure while converting logical plan to physical plan. < 
OptimizerException:[ Failure while attempting to retrieve storage engine. ] < 
FileNotFoundException:[ File sample-data/region.parquet does not exist. ]"
]
                at 
org.apache.drill.exec.rpc.user.QueryResultHandler.batchArrived(QueryResultHandler.java:72)
 ~[java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.exec.rpc.user.UserClient.handle(UserClient.java:79) 
~[java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.exec.rpc.BasicClientWithConnection.handle(BasicClientWithConnection.java:48)
 ~[java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.exec.rpc.BasicClientWithConnection.handle(BasicClientWithConnection.java:33)
 ~[java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.exec.rpc.RpcBus$InboundHandler.decode(RpcBus.java:142) 
~[java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
org.apache.drill.exec.rpc.RpcBus$InboundHandler.decode(RpcBus.java:127) 
~[java-exec-1.0-SNAPSHOT.jar:1.0-SNAPSHOT]
                at 
io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:89)
 ~[netty-codec-4.0.7.Final.jar:na]
                at 
io.netty.channel.DefaultChannelHandlerContext.invokeChannelRead(DefaultChannelHandlerContext.java:334)
 ~[netty-transport-4.0.7.Final.jar:na]
                at 
io.netty.channel.DefaultChannelHandlerContext.fireChannelRead(DefaultChannelHandlerContext.java:320)
 ~[netty-transport-4.0.7.Final.jar:na]
                at 
io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102)
 ~[netty-codec-4.0.7.Final.jar:na]
                at 
io.netty.channel.DefaultChannelHandlerContext.invokeChannelRead(DefaultChannelHandlerContext.java:334)
 ~[netty-transport-4.0.7.Final.jar:na]
                at 
io.netty.channel.DefaultChannelHandlerContext.fireChannelRead(DefaultChannelHandlerContext.java:320)
 ~[netty-transport-4.0.7.Final.jar:na]
                at 
io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:173)
 ~[netty-codec-4.0.7.Final.jar:na]
                at 
io.netty.channel.DefaultChannelHandlerContext.invokeChannelRead(DefaultChannelHandlerContext.java:334)
 ~[netty-transport-4.0.7.Final.jar:na]
                at 
io.netty.channel.DefaultChannelHandlerContext.fireChannelRead(DefaultChannelHandlerContext.java:320)
 ~[netty-transport-4.0.7.Final.jar:na]
                at 
io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:785)
 ~[netty-transport-4.0.7.Final.jar:na]
                at 
io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:100)
 ~[netty-transport-4.0.7.Final.jar:na]
                at 
io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:497) 
~[netty-transport-4.0.7.Final.jar:na]
                at 
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:465)
 ~[netty-transport-4.0.7.Final.jar:na]
                at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:359) 
~[netty-transport-4.0.7.Final.jar:na]
                at 
io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:101)
 ~[netty-common-4.0.7.Final.jar:na]
                at java.lang.Thread.run(Thread.java:724) ~[na:1.7.0_25]
java.lang.RuntimeException: org.apache.drill.exec.rpc.RpcException: Remote 
failure while running query.[error_id: "4962bfd0-316a-43a7-af9f-9826a096c3c7"
endpoint {
  address: "user-VirtualBox"
  user_port: 31010
  bit_port: 31011
}
error_type: 0
message: "Failure while converting logical plan to physical plan. < 
OptimizerException:[ Failure while attempting to retrieve storage engine. ] < 
FileNotFoundException:[ File sample-data/region.parquet does not exist. ]"
]
                at 
org.apache.drill.sql.client.full.ResultEnumerator.moveNext(ResultEnumerator.java:44)
                at 
net.hydromatic.optiq.runtime.ObjectEnumeratorCursor.next(ObjectEnumeratorCursor.java:44)
                at 
net.hydromatic.optiq.jdbc.OptiqResultSet.next(OptiqResultSet.java:162)
                at sqlline.SqlLine$BufferedRows.<init>(SqlLine.java:2499)
                at sqlline.SqlLine.print(SqlLine.java:1886)
                at sqlline.SqlLine$Commands.execute(SqlLine.java:3835)
                at sqlline.SqlLine$Commands.sql(SqlLine.java:3738)
                at sqlline.SqlLine.dispatch(SqlLine.java:882)
                at sqlline.SqlLine.begin(SqlLine.java:717)
                at sqlline.SqlLine.mainWithInputRedirection(SqlLine.java:460)
                at sqlline.SqlLine.main(SqlLine.java:443)
Caused by: org.apache.drill.exec.rpc.RpcException: Remote failure while running 
query.[error_id: "4962bfd0-316a-43a7-af9f-9826a096c3c7"
endpoint {
  address: "user-VirtualBox"
  user_port: 31010
  bit_port: 31011
}
error_type: 0
message: "Failure while converting logical plan to physical plan. < 
OptimizerException:[ Failure while attempting to retrieve storage engine. ] < 
FileNotFoundException:[ File sample-data/region.parquet does not exist. ]"
]
                at 
org.apache.drill.exec.rpc.user.QueryResultHandler.batchArrived(QueryResultHandler.java:72)
                at 
org.apache.drill.exec.rpc.user.UserClient.handle(UserClient.java:79)
                at 
org.apache.drill.exec.rpc.BasicClientWithConnection.handle(BasicClientWithConnection.java:48)
                at 
org.apache.drill.exec.rpc.BasicClientWithConnection.handle(BasicClientWithConnection.java:33)
                at 
org.apache.drill.exec.rpc.RpcBus$InboundHandler.decode(RpcBus.java:142)
                at 
org.apache.drill.exec.rpc.RpcBus$InboundHandler.decode(RpcBus.java:127)
                at 
io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:89)
                at 
io.netty.channel.DefaultChannelHandlerContext.invokeChannelRead(DefaultChannelHandlerContext.java:334)
                at 
io.netty.channel.DefaultChannelHandlerContext.fireChannelRead(DefaultChannelHandlerContext.java:320)
                at 
io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102)
                at 
io.netty.channel.DefaultChannelHandlerContext.invokeChannelRead(DefaultChannelHandlerContext.java:334)
                at 
io.netty.channel.DefaultChannelHandlerContext.fireChannelRead(DefaultChannelHandlerContext.java:320)
                at 
io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:173)
                at 
io.netty.channel.DefaultChannelHandlerContext.invokeChannelRead(DefaultChannelHandlerContext.java:334)
                at 
io.netty.channel.DefaultChannelHandlerContext.fireChannelRead(DefaultChannelHandlerContext.java:320)
                at 
io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:785)
                at 
io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:100)
                at 
io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:497)
                at 
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:465)
                at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:359)
                at 
io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:101)
                at java.lang.Thread.run(Thread.java:724)

Regards

Adam Szulinski | Trainee Software Engineer | Kainos | DD: +48 766 7567 | 
[email protected]<mailto:[email protected]>



Follow the Kainos buzz on: [Twitter] <https://twitter.com/kainossoftware>  
[Facebook] <https://www.facebook.com/KainosSoftware>  [Linkedin] 
<http://www.linkedin.com/company/kainos>  [Youtube] 
<http://www.youtube.com/user/KainosSoftware>

[Sunday 
Times]<http://www.kainos.com/news/kainos-catapulted-straight-in-at-55-in-the-sunday-times-best-companies-to-work-for-2013/>

This e-mail is intended solely for the addressee and is strictly confidential; 
if you are not the addressee please destroy the message and all copies. Any 
opinion or information contained in this email or its attachments that does not 
relate to the business of Kainos is personal to the sender and is not given by 
or endorsed by Kainos. Kainos is the trading name of the Kainos group of 
companies comprising the parent company, Kainos Software Limited, registered in 
Northern Ireland under company number: NI19370, having its registered offices 
at: Kainos House, 4-6 Upper Crescent, Belfast, BT7 1NT, Northern Ireland, 
registered in the UK for VAT under number 454598802, and its wholly owned 
subsidiaries (i) Kainos Software Ireland Limited, registered in Ireland with 
company number: 510943, having its registered office at 6-7 St Stephen's Green, 
Dublin 2, Ireland, registered in Ireland for VAT under number: 9824105M and 
(ii) Kainos Software Poland Spolka z.o.o, registered Poland with national 
business registry number: 221656287, having its registered office at Centrum 
Biurowe Gnilna 2, ul. Gnilna 2, IIIp, 80-847 Gdansk, Poland, registered in 
Poland with tax identification number 5833152587 (used for VAT purposes). This 
email has been scanned for all known viruses by MessageLabs but is not 
guaranteed to be virus free; further terms and conditions may be found on our 
website - www.kainos.com

Reply via email to