Kengo Seki created AIRFLOW-2247:
-----------------------------------

             Summary: Fix RedshiftToS3Transfer not to fail with ValueError
                 Key: AIRFLOW-2247
                 URL: https://issues.apache.org/jira/browse/AIRFLOW-2247
             Project: Apache Airflow
          Issue Type: Bug
          Components: aws, redshift
            Reporter: Kengo Seki


I tried to use RedshiftToS3Transfer but it failed with:

{code}
/path/to/incubator-airflow/airflow/operators/redshift_to_s3_operator.py in 
execute(self, context)
     69         self.hook = PostgresHook(postgres_conn_id=self.redshift_conn_id)
     70         self.s3 = S3Hook(aws_conn_id=self.aws_conn_id)
---> 71         a_key, s_key = self.s3.get_credentials()
     72         unload_options = '\n\t\t\t'.join(self.unload_options)
     73 

ValueError: too many values to unpack
{code}

This is occurred by unmatch between the number of variables and return values. 
As AwsHook.get_credentials' docstring says, it returns three values:

{code}
    def get_credentials(self, region_name=None):
        """Get the underlying `botocore.Credentials` object.

        This contains the attributes: access_key, secret_key and token.
        """
{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to