[ 
https://issues.apache.org/jira/browse/AIRFLOW-1756?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16219423#comment-16219423
 ] 

Ash Berlin-Taylor commented on AIRFLOW-1756:
--------------------------------------------

Ah right, yes boto3 is lower level than boto2. (We didn't get as far as testing 
this part today)

I wonder if get_key should return a s3.Object instead 
http://boto3.readthedocs.io/en/latest/reference/services/s3.html#object -- I 
feel that is closest to what the old API did.

This will need fixing before 1.9.0 goes out. Can you update the fix version on 
this ticket so we don't forget, please?

> S3 Task Handler Cannot Read Logs With New S3Hook
> ------------------------------------------------
>
>                 Key: AIRFLOW-1756
>                 URL: https://issues.apache.org/jira/browse/AIRFLOW-1756
>             Project: Apache Airflow
>          Issue Type: Bug
>    Affects Versions: 1.9.0
>            Reporter: Colin Son
>
> With the changes to the S3Hook, it seems like it cannot read the S3 task logs.
> In the `s3_read` in the S3TaskHandler.py:
> {code}
> s3_key = self.hook.get_key(remote_log_location)
> if s3_key:
>     return s3_key.get_contents_as_string().decode()
> {code}
> Since the s3_key object is now a dict, you cannot call 
> `get_contents_as_string()` on a dict object. You have to use the S3Hook's 
> `read_key()` method to read the contents of the task logs now. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

Reply via email to