but i am able to connect to ECS via python using boto3 libraries without any issues, I am able to write files to the bucket and read them back ..
not sure why i am facing issues with drill though with the same credentials On Mon, Aug 19, 2019 at 11:53 PM Ted Dunning <[email protected]> wrote: > So that looks like the security tokens aren't getting through to AWS > correctly. > > On Mon, Aug 19, 2019 at 11:21 AM Prabu Mohan <[email protected]> > wrote: > > > Log info > > > > 2019-08-19 16:23:05,439 [22a54125-e1e7-29f6-a146-6cfd529b23d1:foreman] > > INFO o.a.drill.exec.work.foreman.Foreman - Query text for query with id > > 22a54125-e1e7-29f6-a146-6cfd529b23d1 issued by anonymous: use ecstest > > > > 2019-08-19 16:23:06,691 [22a54125-e1e7-29f6-a146-6cfd529b23d1:foreman] > > ERROR o.a.drill.exec.work.foreman.Foreman - SYSTEM ERROR: > > AmazonS3Exception: Status Code: 405, AWS Service: Amazon S3, AWS Request > > ID: a1552153:16baf0e3c53:351c7:13, AWS Error Code: null, AWS Error > Message: > > Method Not Allowed > > > > > > > > > > > > Please, refer to logs for more information. > > > > > > > > [Error Id: 76b95ca3-0ed7-4463-b3a1-7b0a8a21cd1a on ] > > > > org.apache.drill.common.exceptions.UserException: SYSTEM ERROR: > > AmazonS3Exception: Status Code: 405, AWS Service: Amazon S3, AWS Request > > ID: a1552153:16baf0e3c53:351c7:13, AWS Error Code: null, AWS Error > Message: > > Method Not Allowed > > > > > > > > > > > > Please, refer to logs for more information. > > > > > > > > [Error Id: 76b95ca3-0ed7-4463-b3a1-7b0a8a21cd1a on ] > > > > at > > > > > org.apache.drill.common.exceptions.UserException$Builder.build(UserException.java:630) > > ~[drill-common-1.16.0.jar:1.16.0] > > > > at > > org.apache.drill.exec.work > > .foreman.Foreman$ForemanResult.close(Foreman.java:789) > > [drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > org.apache.drill.exec.work > > > .foreman.QueryStateProcessor.checkCommonStates(QueryStateProcessor.java:325) > > [drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > org.apache.drill.exec.work > > .foreman.QueryStateProcessor.planning(QueryStateProcessor.java:221) > > [drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > org.apache.drill.exec.work > > .foreman.QueryStateProcessor.moveToState(QueryStateProcessor.java:83) > > [drill-java-exec-1.16.0.jar:1.16.0] > > > > at org.apache.drill.exec.work > > .foreman.Foreman.run(Foreman.java:304) > > [drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > > [na:1.8.0_92] > > > > at > > > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > > [na:1.8.0_92] > > > > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_92] > > > > Caused by: org.apache.drill.exec.work.foreman.ForemanException: > Unexpected > > exception during fragment initialization: Status Code: 405, AWS Service: > > Amazon S3, AWS Request ID: a1552153:16baf0e3c53:351c7:13, AWS Error Code: > > null, AWS Error Message: Method Not Allowed > > > > at org.apache.drill.exec.work > > .foreman.Foreman.run(Foreman.java:305) > > [drill-java-exec-1.16.0.jar:1.16.0] > > > > ... 3 common frames omitted > > > > Caused by: com.amazonaws.services.s3.model.AmazonS3Exception: Status > Code: > > 405, AWS Service: Amazon S3, AWS Request ID: > a1552153:16baf0e3c53:351c7:13, > > AWS Error Code: null, AWS Error Message: Method Not Allowed > > > > at > > > > > com.amazonaws.http.AmazonHttpClient.handleErrorResponse(AmazonHttpClient.java:798) > > ~[aws-java-sdk-1.7.4.jar:na] > > > > at > > > > > com.amazonaws.http.AmazonHttpClient.executeHelper(AmazonHttpClient.java:421) > > ~[aws-java-sdk-1.7.4.jar:na] > > > > at > > com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:232) > > ~[aws-java-sdk-1.7.4.jar:na] > > > > at > > com.amazonaws.services.s3.AmazonS3Client.invoke(AmazonS3Client.java:3528) > > ~[aws-java-sdk-1.7.4.jar:na] > > > > at > > > > > com.amazonaws.services.s3.AmazonS3Client.headBucket(AmazonS3Client.java:1031) > > ~[aws-java-sdk-1.7.4.jar:na] > > > > at > > > > > com.amazonaws.services.s3.AmazonS3Client.doesBucketExist(AmazonS3Client.java:994) > > ~[aws-java-sdk-1.7.4.jar:na] > > > > at > > org.apache.hadoop.fs.s3a.S3AFileSystem.initialize(S3AFileSystem.java:297) > > ~[hadoop-aws-2.7.4.jar:na] > > > > at > > org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2669) > > ~[hadoop-common-2.7.4.jar:na] > > > > at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94) > > ~[hadoop-common-2.7.4.jar:na] > > > > at > > org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703) > > ~[hadoop-common-2.7.4.jar:na] > > > > at > org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685) > > ~[hadoop-common-2.7.4.jar:na] > > > > at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373) > > ~[hadoop-common-2.7.4.jar:na] > > > > at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:172) > > ~[hadoop-common-2.7.4.jar:na] > > > > at > > > > > org.apache.drill.exec.store.dfs.DrillFileSystem.<init>(DrillFileSystem.java:93) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.util.ImpersonationUtil.lambda$createFileSystem$0(ImpersonationUtil.java:215) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at java.security.AccessController.doPrivileged(Native Method) > > ~[na:1.8.0_92] > > > > at javax.security.auth.Subject.doAs(Subject.java:422) > > ~[na:1.8.0_92] > > > > at > > > > > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1746) > > ~[hadoop-common-2.7.4.jar:na] > > > > at > > > > > org.apache.drill.exec.util.ImpersonationUtil.createFileSystem(ImpersonationUtil.java:213) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.util.ImpersonationUtil.createFileSystem(ImpersonationUtil.java:205) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.store.dfs.FileSystemSchemaFactory$FileSystemSchema.<init>(FileSystemSchemaFactory.java:84) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.store.dfs.FileSystemSchemaFactory.registerSchemas(FileSystemSchemaFactory.java:72) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.store.dfs.FileSystemPlugin.registerSchemas(FileSystemPlugin.java:189) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.calcite.jdbc.DynamicRootSchema.loadSchemaFactory(DynamicRootSchema.java:84) > > ~[drill-java-exec-1.16.0.jar:1.18.0-drill-r0] > > > > at > > > > > org.apache.calcite.jdbc.DynamicRootSchema.getImplicitSubSchema(DynamicRootSchema.java:69) > > ~[drill-java-exec-1.16.0.jar:1.18.0-drill-r0] > > > > at > > > org.apache.calcite.jdbc.CalciteSchema.getSubSchema(CalciteSchema.java:262) > > ~[calcite-core-1.18.0-drill-r0.jar:1.18.0-drill-r0] > > > > at > > > > > org.apache.calcite.jdbc.CalciteSchema$SchemaPlusImpl.getSubSchema(CalciteSchema.java:675) > > ~[calcite-core-1.18.0-drill-r0.jar:1.18.0-drill-r0] > > > > at > > > > > org.apache.drill.exec.planner.sql.SchemaUtilites.searchSchemaTree(SchemaUtilites.java:108) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.planner.sql.SchemaUtilites.findSchema(SchemaUtilites.java:52) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.rpc.user.UserSession.setDefaultSchemaPath(UserSession.java:223) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.planner.sql.handlers.UseSchemaHandler.getPlan(UseSchemaHandler.java:43) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan(DrillSqlWorker.java:216) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.planner.sql.DrillSqlWorker.convertPlan(DrillSqlWorker.java:130) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:87) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:593) > > [drill-java-exec-1.16.0.jar:1.16.0] > > > > at org.apache.drill.exec.work > > .foreman.Foreman.run(Foreman.java:276) > > [drill-java-exec-1.16.0.jar:1.16.0] > > > > ... 3 common frames omitted > > > > 2019-08-19 16:23:06,719 [Client-1] INFO > > o.a.d.j.i.DrillCursor$ResultsListener - [#1] Query failed: > > > > org.apache.drill.common.exceptions.UserRemoteException: SYSTEM ERROR: > > AmazonS3Exception: Status Code: 405, AWS Service: Amazon S3, AWS Request > > ID: a1552153:16baf0e3c53:351c7:13, AWS Error Code: null, AWS Error > Message: > > Method Not Allowed > > > > > > > > > > > > Please, refer to logs for more information. > > > > > > > > [Error Id: 76b95ca3-0ed7-4463-b3a1-7b0a8a21cd1a ] > > > > at > > > > > org.apache.drill.exec.rpc.user.QueryResultHandler.resultArrived(QueryResultHandler.java:123) > > [drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > org.apache.drill.exec.rpc.user.UserClient.handle(UserClient.java:422) > > [drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > org.apache.drill.exec.rpc.user.UserClient.handle(UserClient.java:96) > > [drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > org.apache.drill.exec.rpc.RpcBus$InboundHandler.decode(RpcBus.java:273) > > [drill-rpc-1.16.0.jar:1.16.0] > > > > at > > org.apache.drill.exec.rpc.RpcBus$InboundHandler.decode(RpcBus.java:243) > > [drill-rpc-1.16.0.jar:1.16.0] > > > > at > > > > > io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:88) > > [netty-codec-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:356) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:342) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:335) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:287) > > [netty-handler-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:356) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:342) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:335) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102) > > [netty-codec-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:356) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:342) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:335) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:312) > > [netty-codec-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:286) > > [netty-codec-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:356) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:342) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:335) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:86) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:356) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:342) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:335) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1294) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:356) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:342) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:911) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:645) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:580) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:497) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:459) > > [netty-transport-4.0.48.Final.jar:4.0.48.Final] > > > > at > > > > > io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131) > > [netty-common-4.0.48.Final.jar:4.0.48.Final] > > > > at java.lang.Thread.run(Thread.java:745) [na:1.8.0_92] > > > > org.apache.drill.exec.work.foreman.ForemanException: Unexpected > exception > > during fragment initialization: Status Code: 405, AWS Service: Amazon S3, > > AWS Request ID: a1552153:16baf0e3c53:351c7:13, AWS Error Code: null, AWS > > Error Message: Method Not Allowed > > > > at org.apache.drill.exec.work > > .foreman.Foreman.run(Foreman.java:305) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at .......(:0) ~[na:na] > > > > Caused by: com.amazonaws.services.s3.model.AmazonS3Exception: Status > Code: > > 0, AWS Service: null, AWS Request ID: null, AWS Error Code: null, AWS > Error > > Message: Status Code: 405, AWS Service: Amazon S3, AWS Request ID: > > a1552153:16baf0e3c53:351c7:13, AWS Error Code: null, AWS Error Message: > > Method Not Allowed > > > > at > > > > > com.amazonaws.http.AmazonHttpClient.handleErrorResponse(AmazonHttpClient.java:798) > > ~[aws-java-sdk-1.7.4.jar:na] > > > > at > > > > > com.amazonaws.http.AmazonHttpClient.executeHelper(AmazonHttpClient.java:421) > > ~[aws-java-sdk-1.7.4.jar:na] > > > > at > > com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:232) > > ~[aws-java-sdk-1.7.4.jar:na] > > > > at > > com.amazonaws.services.s3.AmazonS3Client.invoke(AmazonS3Client.java:3528) > > ~[aws-java-sdk-1.7.4.jar:na] > > > > at > > > > > com.amazonaws.services.s3.AmazonS3Client.headBucket(AmazonS3Client.java:1031) > > ~[aws-java-sdk-1.7.4.jar:na] > > > > at > > > > > com.amazonaws.services.s3.AmazonS3Client.doesBucketExist(AmazonS3Client.java:994) > > ~[aws-java-sdk-1.7.4.jar:na] > > > > at > > org.apache.hadoop.fs.s3a.S3AFileSystem.initialize(S3AFileSystem.java:297) > > ~[hadoop-aws-2.7.4.jar:na] > > > > at > > org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2669) > > ~[hadoop-common-2.7.4.jar:na] > > > > at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94) > > ~[hadoop-common-2.7.4.jar:na] > > > > at > > org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703) > > ~[hadoop-common-2.7.4.jar:na] > > > > at > org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685) > > ~[hadoop-common-2.7.4.jar:na] > > > > at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373) > > ~[hadoop-common-2.7.4.jar:na] > > > > at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:172) > > ~[hadoop-common-2.7.4.jar:na] > > > > at > > > > > org.apache.drill.exec.store.dfs.DrillFileSystem.<init>(DrillFileSystem.java:93) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.util.ImpersonationUtil.lambda$createFileSystem$0(ImpersonationUtil.java:215) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at .......(:0) ~[na:na] > > > > at > > > > > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1746) > > ~[hadoop-common-2.7.4.jar:na] > > > > at > > > > > org.apache.drill.exec.util.ImpersonationUtil.createFileSystem(ImpersonationUtil.java:213) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.util.ImpersonationUtil.createFileSystem(ImpersonationUtil.java:205) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.store.dfs.FileSystemSchemaFactory$FileSystemSchema.<init>(FileSystemSchemaFactory.java:84) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.store.dfs.FileSystemSchemaFactory.registerSchemas(FileSystemSchemaFactory.java:72) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.store.dfs.FileSystemPlugin.registerSchemas(FileSystemPlugin.java:189) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.calcite.jdbc.DynamicRootSchema.loadSchemaFactory(DynamicRootSchema.java:84) > > ~[drill-java-exec-1.16.0.jar:1.18.0-drill-r0] > > > > at > > > > > org.apache.calcite.jdbc.DynamicRootSchema.getImplicitSubSchema(DynamicRootSchema.java:69) > > ~[drill-java-exec-1.16.0.jar:1.18.0-drill-r0] > > > > at > > > org.apache.calcite.jdbc.CalciteSchema.getSubSchema(CalciteSchema.java:262) > > ~[calcite-core-1.18.0-drill-r0.jar:1.18.0-drill-r0] > > > > at > > > > > org.apache.calcite.jdbc.CalciteSchema$SchemaPlusImpl.getSubSchema(CalciteSchema.java:675) > > ~[calcite-core-1.18.0-drill-r0.jar:1.18.0-drill-r0] > > > > at > > > > > org.apache.drill.exec.planner.sql.SchemaUtilites.searchSchemaTree(SchemaUtilites.java:108) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.planner.sql.SchemaUtilites.findSchema(SchemaUtilites.java:52) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.rpc.user.UserSession.setDefaultSchemaPath(UserSession.java:223) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.planner.sql.handlers.UseSchemaHandler.getPlan(UseSchemaHandler.java:43) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan(DrillSqlWorker.java:216) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.planner.sql.DrillSqlWorker.convertPlan(DrillSqlWorker.java:130) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > > > > org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:87) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at > > org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:593) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > at org.apache.drill.exec.work > > .foreman.Foreman.run(Foreman.java:276) > > ~[drill-java-exec-1.16.0.jar:1.16.0] > > > > ... 1 common frames omitted > > > > > > > > On Mon, Aug 19, 2019 at 12:16 PM Ted Dunning <[email protected]> > > wrote: > > > > > Did you see anything in any logs? > > > > > > > > > > > > On Sun, Aug 18, 2019 at 10:16 PM Prabu Mohan <[email protected]> > > > wrote: > > > > > > > I am able to connect to the http endpoint using boto3 from python > (able > > > to > > > > retrieve files/store files), from IE with https and port 9021 , it > > comes > > > > back with 403 Forbidden indicating that it was able to connect to > > website > > > > but does not have permission to view the webpage ( may be due to > > > > credentials ). > > > > > > > > Regards, > > > > Prabu > > > > > > > > > > > > On Mon, Aug 19, 2019 at 12:59 AM Sorabh Apache< > > > [email protected] > > > > > > > > > wrote: > > > > > > > > > Are you able to use the same configured endpoint *http://:9020* > from > > > > your > > > > > browser ? > > > > > > > > > > Thanks, > > > > > > > > > > Sorabh. > > > > > > > > > > On Mon, Aug 19, 2019 at 12:19 AM Prabu Mohan < > [email protected] > > > > > > > wrote: > > > > > > > > > I'm trying to configure Apache Drill with EMC ECS, it was quite > easy > > > > > configuring with AWS S3 and GCP Cloud Storage, but I'm facing > issues > > > > > configuring with EMC ECS, > > > > > > > > > > When I use http with port 9020 as endpoint I face this error > > > > > > > > > > Error:SYSTEM ERROR: AmazonS3Exception: Status Code: 405, AWS > > > > > Service:Amazon S3, AWS Request ID:, AWS Error Code:null, AWS Error > > > > Message: > > > > > Method Not Allowed > > > > > > > > > > core-site.xml > > > > > > > > > > <configuration> > > > > > > > > > > <property> > > > > > <name>fs.s3a.access.key</name> > > > > > <value>accesskey</value> > > > > > </property> > > > > > > > > > > <property> > > > > > <name>fs.s3a.secret.key</name> > > > > > <value>secretkey</value> > > > > > </property> > > > > > > > > > > <property> > > > > > <name>fs.s3a.endpoint</name> > > > > > http://:9020 > > > > > </property> > > > > > > > > > > <property> > > > > > <name>fs.s3a.connection.ssl.enabled</name> > > > > > <value>false</value> > > > > > </property> > > > > > > > > > > </configuration> > > > > > > > > > > and for connecting to the bucket i use the following in the storage > > > > plugin > > > > > > > > > > { > > > > > "type" : "file", > > > > > "connection" : "s3a://bucketname/", > > > > > "config" : { > > > > > "fs.s3a.impl.disable.cache" : "true" > > > > > }, > > > > > "workspaces" : { > > > > > "tmp" : { > > > > > "location" : "/tmp", > > > > > "writable" : true, > > > > > "defaultInputFormat" : null, > > > > > "allowAccessOutsideWorkspace" : false > > > > > }, > > > > > "root" : { > > > > > "location" : "/", > > > > > "writable" : false, > > > > > "defaultInputFormat" : null, > > > > > "allowAccessOutsideWorkspace" : false > > > > > } > > > > > }, > > > > > "formats" : { > > > > > "psv" : { > > > > > "type" : "text", > > > > > "extensions" : [ "tbl" ], > > > > > "delimiter" : "|" > > > > > }, > > > > > "csv" : { > > > > > "type" : "text", > > > > > "extensions" : [ "csv" ], > > > > > "delimiter" : "," > > > > > }, > > > > > "tsv" : { > > > > > "type" : "text", > > > > > "extensions" : [ "tsv" ], > > > > > "delimiter" : "\t" > > > > > }, > > > > > "parquet" : { > > > > > "type" : "parquet" > > > > > }, > > > > > "json" : { > > > > > "type" : "json", > > > > > "extensions" : [ "json" ] > > > > > }, > > > > > "avro" : { > > > > > "type" : "avro" > > > > > }, > > > > > "sequencefile" : { > > > > > "type" : "sequencefile", > > > > > "extensions" : [ "seq" ] > > > > > }, > > > > > "csvh" : { > > > > > "type" : "text", > > > > > "extensions" : [ "csvh" ], > > > > > "extractHeader" : true, > > > > > "delimiter" : "," > > > > > } > > > > > }, > > > > > "enabled" : true > > > > > } > > > > > > > > > > Any suggestions on how to get this working ? > > > > > > > > > > > > > > >
