Log info 2019-08-19 16:23:05,439 [22a54125-e1e7-29f6-a146-6cfd529b23d1:foreman] INFO o.a.drill.exec.work.foreman.Foreman - Query text for query with id 22a54125-e1e7-29f6-a146-6cfd529b23d1 issued by anonymous: use ecstest
2019-08-19 16:23:06,691 [22a54125-e1e7-29f6-a146-6cfd529b23d1:foreman] ERROR o.a.drill.exec.work.foreman.Foreman - SYSTEM ERROR: AmazonS3Exception: Status Code: 405, AWS Service: Amazon S3, AWS Request ID: a1552153:16baf0e3c53:351c7:13, AWS Error Code: null, AWS Error Message: Method Not Allowed Please, refer to logs for more information. [Error Id: 76b95ca3-0ed7-4463-b3a1-7b0a8a21cd1a on ] org.apache.drill.common.exceptions.UserException: SYSTEM ERROR: AmazonS3Exception: Status Code: 405, AWS Service: Amazon S3, AWS Request ID: a1552153:16baf0e3c53:351c7:13, AWS Error Code: null, AWS Error Message: Method Not Allowed Please, refer to logs for more information. [Error Id: 76b95ca3-0ed7-4463-b3a1-7b0a8a21cd1a on ] at org.apache.drill.common.exceptions.UserException$Builder.build(UserException.java:630) ~[drill-common-1.16.0.jar:1.16.0] at org.apache.drill.exec.work.foreman.Foreman$ForemanResult.close(Foreman.java:789) [drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.work.foreman.QueryStateProcessor.checkCommonStates(QueryStateProcessor.java:325) [drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.work.foreman.QueryStateProcessor.planning(QueryStateProcessor.java:221) [drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.work.foreman.QueryStateProcessor.moveToState(QueryStateProcessor.java:83) [drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:304) [drill-java-exec-1.16.0.jar:1.16.0] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [na:1.8.0_92] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [na:1.8.0_92] at java.lang.Thread.run(Thread.java:745) [na:1.8.0_92] Caused by: org.apache.drill.exec.work.foreman.ForemanException: Unexpected exception during fragment initialization: Status Code: 405, AWS Service: Amazon S3, AWS Request ID: a1552153:16baf0e3c53:351c7:13, AWS Error Code: null, AWS Error Message: Method Not Allowed at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:305) [drill-java-exec-1.16.0.jar:1.16.0] ... 3 common frames omitted Caused by: com.amazonaws.services.s3.model.AmazonS3Exception: Status Code: 405, AWS Service: Amazon S3, AWS Request ID: a1552153:16baf0e3c53:351c7:13, AWS Error Code: null, AWS Error Message: Method Not Allowed at com.amazonaws.http.AmazonHttpClient.handleErrorResponse(AmazonHttpClient.java:798) ~[aws-java-sdk-1.7.4.jar:na] at com.amazonaws.http.AmazonHttpClient.executeHelper(AmazonHttpClient.java:421) ~[aws-java-sdk-1.7.4.jar:na] at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:232) ~[aws-java-sdk-1.7.4.jar:na] at com.amazonaws.services.s3.AmazonS3Client.invoke(AmazonS3Client.java:3528) ~[aws-java-sdk-1.7.4.jar:na] at com.amazonaws.services.s3.AmazonS3Client.headBucket(AmazonS3Client.java:1031) ~[aws-java-sdk-1.7.4.jar:na] at com.amazonaws.services.s3.AmazonS3Client.doesBucketExist(AmazonS3Client.java:994) ~[aws-java-sdk-1.7.4.jar:na] at org.apache.hadoop.fs.s3a.S3AFileSystem.initialize(S3AFileSystem.java:297) ~[hadoop-aws-2.7.4.jar:na] at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2669) ~[hadoop-common-2.7.4.jar:na] at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94) ~[hadoop-common-2.7.4.jar:na] at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703) ~[hadoop-common-2.7.4.jar:na] at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685) ~[hadoop-common-2.7.4.jar:na] at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373) ~[hadoop-common-2.7.4.jar:na] at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:172) ~[hadoop-common-2.7.4.jar:na] at org.apache.drill.exec.store.dfs.DrillFileSystem.<init>(DrillFileSystem.java:93) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.util.ImpersonationUtil.lambda$createFileSystem$0(ImpersonationUtil.java:215) ~[drill-java-exec-1.16.0.jar:1.16.0] at java.security.AccessController.doPrivileged(Native Method) ~[na:1.8.0_92] at javax.security.auth.Subject.doAs(Subject.java:422) ~[na:1.8.0_92] at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1746) ~[hadoop-common-2.7.4.jar:na] at org.apache.drill.exec.util.ImpersonationUtil.createFileSystem(ImpersonationUtil.java:213) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.util.ImpersonationUtil.createFileSystem(ImpersonationUtil.java:205) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.store.dfs.FileSystemSchemaFactory$FileSystemSchema.<init>(FileSystemSchemaFactory.java:84) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.store.dfs.FileSystemSchemaFactory.registerSchemas(FileSystemSchemaFactory.java:72) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.store.dfs.FileSystemPlugin.registerSchemas(FileSystemPlugin.java:189) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.calcite.jdbc.DynamicRootSchema.loadSchemaFactory(DynamicRootSchema.java:84) ~[drill-java-exec-1.16.0.jar:1.18.0-drill-r0] at org.apache.calcite.jdbc.DynamicRootSchema.getImplicitSubSchema(DynamicRootSchema.java:69) ~[drill-java-exec-1.16.0.jar:1.18.0-drill-r0] at org.apache.calcite.jdbc.CalciteSchema.getSubSchema(CalciteSchema.java:262) ~[calcite-core-1.18.0-drill-r0.jar:1.18.0-drill-r0] at org.apache.calcite.jdbc.CalciteSchema$SchemaPlusImpl.getSubSchema(CalciteSchema.java:675) ~[calcite-core-1.18.0-drill-r0.jar:1.18.0-drill-r0] at org.apache.drill.exec.planner.sql.SchemaUtilites.searchSchemaTree(SchemaUtilites.java:108) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.planner.sql.SchemaUtilites.findSchema(SchemaUtilites.java:52) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.rpc.user.UserSession.setDefaultSchemaPath(UserSession.java:223) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.planner.sql.handlers.UseSchemaHandler.getPlan(UseSchemaHandler.java:43) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan(DrillSqlWorker.java:216) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.planner.sql.DrillSqlWorker.convertPlan(DrillSqlWorker.java:130) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:87) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:593) [drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:276) [drill-java-exec-1.16.0.jar:1.16.0] ... 3 common frames omitted 2019-08-19 16:23:06,719 [Client-1] INFO o.a.d.j.i.DrillCursor$ResultsListener - [#1] Query failed: org.apache.drill.common.exceptions.UserRemoteException: SYSTEM ERROR: AmazonS3Exception: Status Code: 405, AWS Service: Amazon S3, AWS Request ID: a1552153:16baf0e3c53:351c7:13, AWS Error Code: null, AWS Error Message: Method Not Allowed Please, refer to logs for more information. [Error Id: 76b95ca3-0ed7-4463-b3a1-7b0a8a21cd1a ] at org.apache.drill.exec.rpc.user.QueryResultHandler.resultArrived(QueryResultHandler.java:123) [drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.rpc.user.UserClient.handle(UserClient.java:422) [drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.rpc.user.UserClient.handle(UserClient.java:96) [drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.rpc.RpcBus$InboundHandler.decode(RpcBus.java:273) [drill-rpc-1.16.0.jar:1.16.0] at org.apache.drill.exec.rpc.RpcBus$InboundHandler.decode(RpcBus.java:243) [drill-rpc-1.16.0.jar:1.16.0] at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:88) [netty-codec-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:356) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:342) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:335) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:287) [netty-handler-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:356) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:342) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:335) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102) [netty-codec-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:356) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:342) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:335) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:312) [netty-codec-4.0.48.Final.jar:4.0.48.Final] at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:286) [netty-codec-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:356) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:342) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:335) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:86) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:356) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:342) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:335) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1294) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:356) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:342) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:911) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:645) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:580) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:497) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:459) [netty-transport-4.0.48.Final.jar:4.0.48.Final] at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131) [netty-common-4.0.48.Final.jar:4.0.48.Final] at java.lang.Thread.run(Thread.java:745) [na:1.8.0_92] org.apache.drill.exec.work.foreman.ForemanException: Unexpected exception during fragment initialization: Status Code: 405, AWS Service: Amazon S3, AWS Request ID: a1552153:16baf0e3c53:351c7:13, AWS Error Code: null, AWS Error Message: Method Not Allowed at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:305) ~[drill-java-exec-1.16.0.jar:1.16.0] at .......(:0) ~[na:na] Caused by: com.amazonaws.services.s3.model.AmazonS3Exception: Status Code: 0, AWS Service: null, AWS Request ID: null, AWS Error Code: null, AWS Error Message: Status Code: 405, AWS Service: Amazon S3, AWS Request ID: a1552153:16baf0e3c53:351c7:13, AWS Error Code: null, AWS Error Message: Method Not Allowed at com.amazonaws.http.AmazonHttpClient.handleErrorResponse(AmazonHttpClient.java:798) ~[aws-java-sdk-1.7.4.jar:na] at com.amazonaws.http.AmazonHttpClient.executeHelper(AmazonHttpClient.java:421) ~[aws-java-sdk-1.7.4.jar:na] at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:232) ~[aws-java-sdk-1.7.4.jar:na] at com.amazonaws.services.s3.AmazonS3Client.invoke(AmazonS3Client.java:3528) ~[aws-java-sdk-1.7.4.jar:na] at com.amazonaws.services.s3.AmazonS3Client.headBucket(AmazonS3Client.java:1031) ~[aws-java-sdk-1.7.4.jar:na] at com.amazonaws.services.s3.AmazonS3Client.doesBucketExist(AmazonS3Client.java:994) ~[aws-java-sdk-1.7.4.jar:na] at org.apache.hadoop.fs.s3a.S3AFileSystem.initialize(S3AFileSystem.java:297) ~[hadoop-aws-2.7.4.jar:na] at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2669) ~[hadoop-common-2.7.4.jar:na] at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94) ~[hadoop-common-2.7.4.jar:na] at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703) ~[hadoop-common-2.7.4.jar:na] at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685) ~[hadoop-common-2.7.4.jar:na] at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373) ~[hadoop-common-2.7.4.jar:na] at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:172) ~[hadoop-common-2.7.4.jar:na] at org.apache.drill.exec.store.dfs.DrillFileSystem.<init>(DrillFileSystem.java:93) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.util.ImpersonationUtil.lambda$createFileSystem$0(ImpersonationUtil.java:215) ~[drill-java-exec-1.16.0.jar:1.16.0] at .......(:0) ~[na:na] at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1746) ~[hadoop-common-2.7.4.jar:na] at org.apache.drill.exec.util.ImpersonationUtil.createFileSystem(ImpersonationUtil.java:213) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.util.ImpersonationUtil.createFileSystem(ImpersonationUtil.java:205) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.store.dfs.FileSystemSchemaFactory$FileSystemSchema.<init>(FileSystemSchemaFactory.java:84) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.store.dfs.FileSystemSchemaFactory.registerSchemas(FileSystemSchemaFactory.java:72) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.store.dfs.FileSystemPlugin.registerSchemas(FileSystemPlugin.java:189) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.calcite.jdbc.DynamicRootSchema.loadSchemaFactory(DynamicRootSchema.java:84) ~[drill-java-exec-1.16.0.jar:1.18.0-drill-r0] at org.apache.calcite.jdbc.DynamicRootSchema.getImplicitSubSchema(DynamicRootSchema.java:69) ~[drill-java-exec-1.16.0.jar:1.18.0-drill-r0] at org.apache.calcite.jdbc.CalciteSchema.getSubSchema(CalciteSchema.java:262) ~[calcite-core-1.18.0-drill-r0.jar:1.18.0-drill-r0] at org.apache.calcite.jdbc.CalciteSchema$SchemaPlusImpl.getSubSchema(CalciteSchema.java:675) ~[calcite-core-1.18.0-drill-r0.jar:1.18.0-drill-r0] at org.apache.drill.exec.planner.sql.SchemaUtilites.searchSchemaTree(SchemaUtilites.java:108) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.planner.sql.SchemaUtilites.findSchema(SchemaUtilites.java:52) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.rpc.user.UserSession.setDefaultSchemaPath(UserSession.java:223) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.planner.sql.handlers.UseSchemaHandler.getPlan(UseSchemaHandler.java:43) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan(DrillSqlWorker.java:216) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.planner.sql.DrillSqlWorker.convertPlan(DrillSqlWorker.java:130) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:87) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:593) ~[drill-java-exec-1.16.0.jar:1.16.0] at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:276) ~[drill-java-exec-1.16.0.jar:1.16.0] ... 1 common frames omitted On Mon, Aug 19, 2019 at 12:16 PM Ted Dunning <ted.dunn...@gmail.com> wrote: > Did you see anything in any logs? > > > > On Sun, Aug 18, 2019 at 10:16 PM Prabu Mohan <prabu.ora...@gmail.com> > wrote: > > > I am able to connect to the http endpoint using boto3 from python (able > to > > retrieve files/store files), from IE with https and port 9021 , it comes > > back with 403 Forbidden indicating that it was able to connect to website > > but does not have permission to view the webpage ( may be due to > > credentials ). > > > > Regards, > > Prabu > > > > > > On Mon, Aug 19, 2019 at 12:59 AM Sorabh Apache< > sorabh.apa...@apache.org > > > > > wrote: > > > > > Are you able to use the same configured endpoint *http://:9020* from > > your > > > browser ? > > > > > > Thanks, > > > > > > Sorabh. > > > > > > On Mon, Aug 19, 2019 at 12:19 AM Prabu Mohan <prabu.ora...@gmail.com> > > wrote: > > > > > I'm trying to configure Apache Drill with EMC ECS, it was quite easy > > > configuring with AWS S3 and GCP Cloud Storage, but I'm facing issues > > > configuring with EMC ECS, > > > > > > When I use http with port 9020 as endpoint I face this error > > > > > > Error:SYSTEM ERROR: AmazonS3Exception: Status Code: 405, AWS > > > Service:Amazon S3, AWS Request ID:, AWS Error Code:null, AWS Error > > Message: > > > Method Not Allowed > > > > > > core-site.xml > > > > > > <configuration> > > > > > > <property> > > > <name>fs.s3a.access.key</name> > > > <value>accesskey</value> > > > </property> > > > > > > <property> > > > <name>fs.s3a.secret.key</name> > > > <value>secretkey</value> > > > </property> > > > > > > <property> > > > <name>fs.s3a.endpoint</name> > > > http://:9020 > > > </property> > > > > > > <property> > > > <name>fs.s3a.connection.ssl.enabled</name> > > > <value>false</value> > > > </property> > > > > > > </configuration> > > > > > > and for connecting to the bucket i use the following in the storage > > plugin > > > > > > { > > > "type" : "file", > > > "connection" : "s3a://bucketname/", > > > "config" : { > > > "fs.s3a.impl.disable.cache" : "true" > > > }, > > > "workspaces" : { > > > "tmp" : { > > > "location" : "/tmp", > > > "writable" : true, > > > "defaultInputFormat" : null, > > > "allowAccessOutsideWorkspace" : false > > > }, > > > "root" : { > > > "location" : "/", > > > "writable" : false, > > > "defaultInputFormat" : null, > > > "allowAccessOutsideWorkspace" : false > > > } > > > }, > > > "formats" : { > > > "psv" : { > > > "type" : "text", > > > "extensions" : [ "tbl" ], > > > "delimiter" : "|" > > > }, > > > "csv" : { > > > "type" : "text", > > > "extensions" : [ "csv" ], > > > "delimiter" : "," > > > }, > > > "tsv" : { > > > "type" : "text", > > > "extensions" : [ "tsv" ], > > > "delimiter" : "\t" > > > }, > > > "parquet" : { > > > "type" : "parquet" > > > }, > > > "json" : { > > > "type" : "json", > > > "extensions" : [ "json" ] > > > }, > > > "avro" : { > > > "type" : "avro" > > > }, > > > "sequencefile" : { > > > "type" : "sequencefile", > > > "extensions" : [ "seq" ] > > > }, > > > "csvh" : { > > > "type" : "text", > > > "extensions" : [ "csvh" ], > > > "extractHeader" : true, > > > "delimiter" : "," > > > } > > > }, > > > "enabled" : true > > > } > > > > > > Any suggestions on how to get this working ? > > > > > >