[ 
https://issues.apache.org/jira/browse/HADOOP-17771?focusedWorklogId=613609&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-613609
 ]

ASF GitHub Bot logged work on HADOOP-17771:
-------------------------------------------

                Author: ASF GitHub Bot
            Created on: 22/Jun/21 17:25
            Start Date: 22/Jun/21 17:25
    Worklog Time Spent: 10m 
      Work Description: steveloughran commented on pull request #3133:
URL: https://github.com/apache/hadoop/pull/3133#issuecomment-866183472


   tested with  `-Dparallel-tests -DtestsThreadCount=7 -Dmarkers=keep -Dscale 
-Ds3guard -Ddynamo`
   
   transient read buffer underflow failure; one extra S3Guard write than 
expected. Both of those surface when there are too many records
   
   
   ```
   [ERROR] Failures:
   [ERROR]   
ITestS3AContractUnbuffer>AbstractContractUnbufferTest.testUnbufferOnClosedFile:83->AbstractContractUnbufferTest.validateFullFileContents:132->AbstractContractUnbufferTest.validateFileContents:139->Assert.assertEquals:647->Assert.failNotEquals:835->Assert.fail:89
 failed to read expected number of bytes from stream. This may be transient 
expected:<1024> but was:<515>
   [ERROR]   
ITestCommitOperations.testBulkCommitFiles:723->Assert.assertEquals:647->Assert.failNotEquals:835->Assert.fail:89
 Number of records written after commit #2; first commit had 4; first commit 
ancestors CommitContext{operationState=AncestorState{operation=Commitid=44; 
dest=s3a://stevel-london/fork-0007/test/DELAY_LISTING_ME/testBulkCommitFiles/out;
 size=6; 
paths={s3a://stevel-london/fork-0007/test/DELAY_LISTING_ME/testBulkCommitFiles/out
 
s3a://stevel-london/fork-0007/test/DELAY_LISTING_ME/testBulkCommitFiles/out/file1
 s3a://stevel-london/fork-0007 s3a://stevel-london/fork-0007/test 
s3a://stevel-london/fork-0007/test/DELAY_LISTING_ME/testBulkCommitFiles 
s3a://stevel-london/fork-0007/test/DELAY_LISTING_ME}}}; second commit 
ancestors: CommitContext{operationState=AncestorState{operation=Commitid=44; 
dest=s3a://stevel-london/fork-0007/test/DELAY_LISTING_ME/testBulkCommitFiles/out;
 size=8; 
paths={s3a://stevel-london/fork-0007/test/DELAY_LISTING_ME/testBulkCommitFiles/out
 
s3a://stevel-london/fork-0007/test/DELAY_LISTING_ME/testBulkCommitFiles/out/file1
 s3a://stevel-london/fork-0007 
s3a://stevel-london/fork-0007/test/DELAY_LISTING_ME/testBulkCommitFiles/out/subdir
 s3a://stevel-london/fork-0007/test 
s3a://stevel-london/fork-0007/test/DELAY_LISTING_ME/testBulkCommitFiles 
s3a://stevel-london/fork-0007/test/DELAY_LISTING_ME/testBulkCommitFiles/out/subdir/file2
 s3a://stevel-london/fork-0007/test/DELAY_LISTING_ME}}}: 
s3guard_metadatastore_record_writes expected:<2> but was:<3>
   [INFO]
   ```


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
-------------------

    Worklog Id:     (was: 613609)
    Time Spent: 1h 10m  (was: 1h)

> S3AFS creation fails without region set in ~/.aws/config
> --------------------------------------------------------
>
>                 Key: HADOOP-17771
>                 URL: https://issues.apache.org/jira/browse/HADOOP-17771
>             Project: Hadoop Common
>          Issue Type: Sub-task
>          Components: fs/s3
>    Affects Versions: 3.3.1
>         Environment: Host outside EC2 and without the file ~/.aws/config or 
> without a region set in it
>            Reporter: Steve Loughran
>            Assignee: Steve Loughran
>            Priority: Blocker
>              Labels: pull-request-available
>          Time Spent: 1h 10m
>  Remaining Estimate: 0h
>
> If you don't have {{fs.s3a.endpoint}} set and lack a region set in
> env var {{AWS_REGION_ENV_VAR}}, system property {{aws.region}} or the file  
> ~/.aws/config
> then S3A FS creation fails with  the message
> "Unable to find a region via the region provider chain."
> This is caused by the move to the AWS S3 client builder API in HADOOP-13551
> This is pretty dramatic and no doubt everyone will be asking "why didn't you 
> notice this?",
> But in fact there are some reasons.
> # when running in EC2, all is well. Meaning our big test runs were all happy.
> # if a developer has fs.s3a.endpoint set for the test bucket, all is well.
>    Those of us who work with buckets in the "regions tend to do this, not 
> least because it can save a HEAD request every time an FS is created.
> # if you have a region set in ~/.aws/config then all is well
> reason #3 is the real surprise and the one which has really caught out. Even 
> my tests against buckets in usw-2 through central didn't fail because of 
> course I, like my colleagues, have the AWS S3 client installed locally. This 
> was sufficient to make the problem go away. It is also why this has been an 
> intermittent problem on test clusters outside AWS infra: it really depended 
> on the VM/docker image whether things worked or not.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org

Reply via email to