Steve Loughran created SPARK-19739:
--------------------------------------

             Summary: SparkHadoopUtil.appendS3AndSparkHadoopConfigurations to 
propagate full set of AWS env vars
                 Key: SPARK-19739
                 URL: https://issues.apache.org/jira/browse/SPARK-19739
             Project: Spark
          Issue Type: Improvement
          Components: Spark Core
    Affects Versions: 2.1.0
            Reporter: Steve Loughran
            Priority: Minor


{{SparkHadoopUtil.appendS3AndSparkHadoopConfigurations()}} propagates the AWS 
user and secret key to s3n and s3a config options, so getting secrets from the 
user to the cluster, if set.

AWS also supports session authentication (env var {{AWS_SESSION_TOKEN}}) and 
region endpoints {{AWS_DEFAULT_REGION}}, the latter being critical if you want 
to address V4-auth-only endpoints like frankfurt and Seol. 

These env vars should be picked up and passed down to S3a too. 4+ lines of 
code, though impossible to test unless the existing code is refactored to take 
the env var map[String, String], so allowing a test suite to set the values in 
itds own map.

side issue: what if only half the env vars are set and users are trying to 
understand why auth is failing? It may be good to build up a string identifying 
which env vars had their value propagate, and log that @ debug, while not 
logging the values, obviously.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to