[ 
https://issues.apache.org/jira/browse/AIRFLOW-2247?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kengo Seki reassigned AIRFLOW-2247:
-----------------------------------

    Assignee: Kengo Seki

> Fix RedshiftToS3Transfer not to fail with ValueError
> ----------------------------------------------------
>
>                 Key: AIRFLOW-2247
>                 URL: https://issues.apache.org/jira/browse/AIRFLOW-2247
>             Project: Apache Airflow
>          Issue Type: Bug
>          Components: aws, redshift
>            Reporter: Kengo Seki
>            Assignee: Kengo Seki
>            Priority: Major
>
> I tried to use RedshiftToS3Transfer but it failed with:
> {code}
> /path/to/incubator-airflow/airflow/operators/redshift_to_s3_operator.py in 
> execute(self, context)
>      69         self.hook = 
> PostgresHook(postgres_conn_id=self.redshift_conn_id)
>      70         self.s3 = S3Hook(aws_conn_id=self.aws_conn_id)
> ---> 71         a_key, s_key = self.s3.get_credentials()
>      72         unload_options = '\n\t\t\t'.join(self.unload_options)
>      73 
> ValueError: too many values to unpack
> {code}
> This is occurred by unmatch between the number of variables and return 
> values. As AwsHook.get_credentials' docstring says, it returns three values:
> {code}
>     def get_credentials(self, region_name=None):
>         """Get the underlying `botocore.Credentials` object.
>         This contains the attributes: access_key, secret_key and token.
>         """
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to