[ 
https://issues.apache.org/jira/browse/HADOOP-13904?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15851644#comment-15851644
 ] 

Steve Loughran commented on HADOOP-13904:
-----------------------------------------

Yetus is unhappy with it...is it in sync with the branch?

* that fix to line 196 of the pom should go into branch-2...submit a separate 
patch for that and I'll get it in

h2. {{retryBackoff}}
* the retry policy should really detect and reject the auth failures as 
non-retryable. Looking @ the s3a block output stream, we get away with it only 
because you don't get as far as completing a multipart write without having the 
credentials —though I should add the check there too, to failfast on situations 
like session credential expiry during a multiday streaming app.
* Take a look at {{S3aBlockOutputStream.shouldRetry}} for some things to 
consider: (a) handle interruptions by interrupting thread again, and (b) 
handling any other exception by just returning false to the shouldRetry probe. 
Why? it means the caller can fail with whatever exception caused the initial 
problem, which is presumably the most useful.

other than that, LGTM.

> DynamoDBMetadataStore to handle DDB throttling failures through retry policy
> ----------------------------------------------------------------------------
>
>                 Key: HADOOP-13904
>                 URL: https://issues.apache.org/jira/browse/HADOOP-13904
>             Project: Hadoop Common
>          Issue Type: Sub-task
>          Components: fs/s3
>    Affects Versions: HADOOP-13345
>            Reporter: Steve Loughran
>            Assignee: Aaron Fabbri
>         Attachments: HADOOP-13904-HADOOP-13345.001.patch, 
> HADOOP-13904-HADOOP-13345.002.patch
>
>
> When you overload DDB, you get error messages warning of throttling, [as 
> documented by 
> AWS|http://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Programming.Errors.html#Programming.Errors.MessagesAndCodes]
> Reduce load on DDB by doing a table lookup before the create, then, in table 
> create/delete operations and in get/put actions, recognise the error codes 
> and retry using an appropriate retry policy (exponential backoff + ultimate 
> failure) 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

---------------------------------------------------------------------
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org

Reply via email to