[ https://issues.apache.org/jira/browse/AIRFLOW-2247?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Kengo Seki reassigned AIRFLOW-2247: ----------------------------------- Assignee: Kengo Seki > Fix RedshiftToS3Transfer not to fail with ValueError > ---------------------------------------------------- > > Key: AIRFLOW-2247 > URL: https://issues.apache.org/jira/browse/AIRFLOW-2247 > Project: Apache Airflow > Issue Type: Bug > Components: aws, redshift > Reporter: Kengo Seki > Assignee: Kengo Seki > Priority: Major > > I tried to use RedshiftToS3Transfer but it failed with: > {code} > /path/to/incubator-airflow/airflow/operators/redshift_to_s3_operator.py in > execute(self, context) > 69 self.hook = > PostgresHook(postgres_conn_id=self.redshift_conn_id) > 70 self.s3 = S3Hook(aws_conn_id=self.aws_conn_id) > ---> 71 a_key, s_key = self.s3.get_credentials() > 72 unload_options = '\n\t\t\t'.join(self.unload_options) > 73 > ValueError: too many values to unpack > {code} > This is occurred by unmatch between the number of variables and return > values. As AwsHook.get_credentials' docstring says, it returns three values: > {code} > def get_credentials(self, region_name=None): > """Get the underlying `botocore.Credentials` object. > This contains the attributes: access_key, secret_key and token. > """ > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)