I'm using EC2 instances Thank you. Daniel
> On 9 Jun 2016, at 16:49, Gourav Sengupta <gourav.sengu...@gmail.com> wrote: > > Hi, > > are you using EC2 instances or local cluster behind firewall. > > > Regards, > Gourav Sengupta > >> On Wed, Jun 8, 2016 at 4:34 PM, Daniel Haviv >> <daniel.ha...@veracity-group.com> wrote: >> Hi, >> I'm trying to create a table on s3a but I keep hitting the following error: >> Exception in thread "main" org.apache.hadoop.hive.ql.metadata.HiveException: >> MetaException(message:com.cloudera.com.amazonaws.AmazonClientException: >> Unable to load AWS credentials from any provider in the chain) >> >> I tried setting the s3a keys using the configuration object but I might be >> hitting SPARK-11364 : >> conf.set("fs.s3a.access.key", accessKey) >> conf.set("fs.s3a.secret.key", secretKey) >> conf.set("spark.hadoop.fs.s3a.access.key",accessKey) >> conf.set("spark.hadoop.fs.s3a.secret.key",secretKey) >> val sc = new SparkContext(conf) >> >> I tried setting these propeties in hdfs-site.xml but i'm still getting this >> error. >> Finally I tried to set the AWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEY >> environment variables but with no luck. >> >> Any ideas on how to resolve this issue ? >> >> Thank you. >> Daniel >> >> Thank you. >> Daniel >