Re: S3 Access in eu-central-1

2018-01-02 Thread Nico Kruber
Sorry for the late response,
but I finally got around adding this workaround to our "common issues"
section with PR https://github.com/apache/flink/pull/5231

Nico

On 29/11/17 09:31, Ufuk Celebi wrote:
> Hey Dominik,
> 
> yes, we should definitely add this to the docs.
> 
> @Nico: You recently updated the Flink S3 setup docs. Would you mind
> adding these hints for eu-central-1 from Steve? I think that would be
> super helpful!
> 
> Best,
> 
> Ufuk
> 
> On Tue, Nov 28, 2017 at 10:00 PM, Dominik Bruhn  wrote:
>> Hey Stephan, Hey Steve,
>> that was the right hint, adding that open to the Java-Options fixed the
>> problem. Maybe we should add this somehow to our Flink Wiki?
>>
>> Thanks!
>>
>> Dominik
>>
>> On 28/11/17 11:55, Stephan Ewen wrote:
>>>
>>> Got a pointer from Steve that this is answered on Stack Overflow here:
>>> https://stackoverflow.com/questions/36154484/aws-java-sdk-manually-set-signature-version
>>> 
>>>
>>> Flink 1.4 contains a specially bundled "fs-s3-hadoop" with smaller no
>>> footprint, compatible across Hadoop versions, and based on a later s3a and
>>> AWS sdk. In that connector, it should work out of the box because it uses a
>>> later AWS SDK. You can also use it with earlier Hadoop versions because
>>> dependencies are relocated, so it should not cash/conflict.
>>>
>>>
>>>
>>>
>>> On Mon, Nov 27, 2017 at 8:58 PM, Stephan Ewen >> > wrote:
>>>
>>> Hi!
>>>
>>> The endpoint config entry looks correct.
>>> I was looking at this issue to see if there are pointers to anything
>>> else, but it looks like the explicit endpoint entry is the most
>>> important thing: https://issues.apache.org/jira/browse/HADOOP-13324
>>> 
>>>
>>> I cc-ed Steve Loughran, who is Hadoop's S3 expert (sorry Steve for
>>> pulling you in again - listening and learning still about the subtle
>>> bits and pieces of S3).
>>> @Steve are S3 V4 endpoints supported in Hadoop 2.7.x already, or
>>> only in Hadoop 2.8?
>>>
>>> Best,
>>> Stephan
>>>
>>>
>>> On Mon, Nov 27, 2017 at 9:47 AM, Dominik Bruhn >> > wrote:
>>>
>>> Hey,
>>> can anyone give a hint? Does anyone have flink running with an
>>> S3 Bucket in Frankfurt/eu-central-1 and can share his config and
>>> setup?
>>>
>>> Thanks,
>>> Dominik
>>>
>>> On 22. Nov 2017, at 17:52, domi...@dbruhn.de
>>>  wrote:
>>>
 Hey everyone,
 I'm trying since hours to get Flink 1.3.2 (downloaded for
 hadoop 2.7) to snapshot/checkpoint to an S3 bucket which is
 hosted in the eu-central-1 region. Everything works fine for
 other regions. I'm running my job on a JobTracker in local
 mode. I googled the internet and found several hints, most of
 them telling that setting the `fs.s3a.endpoint` should solve
 it. It doesn't. I'm also sure that the core-site.xml (see
 below) is picked up, if I put garbage into the endpoint then I
 receive a hostname not found error.

 The exception I'm getting is:
 com.amazonaws.services.s3.model.AmazonS3Exception: Status
 Code: 400, AWS Service: Amazon S3, AWS Request ID:
 432415098B0994BC, AWS Error Code: null, AWS Error Message: Bad
 Request, S3 Extended Request ID:

 1PSDe4EOh7zvfNPdWrwoBKKOtsS/gf9atn5movRzcpvIH2WsR+ptXvXyFyEHXjDb3F9AniXgsBQ=

 I read the AWS FAQ but I don't think that

 https://ci.apache.org/projects/flink/flink-docs-release-1.3/setup/aws.html#ioexception-400-bad-request

 
 applies to me as I'm not running the NativeFileSystem.

 I suspect this is related to the v4 signing protocol which is
 required for S3 in Frankfurt. Could it be that the aws-sdk
 version is just too old? I tried to play around with it but
 the hadoop adapter is incompatible with newer versions.

 I have the following core-site.xml:

 
 

 fs.s3.implorg.apache.hadoop.fs.s3a.S3AFileSystem

 fs.s3a.buffer.dir/tmp

 fs.s3a.access.keysomething

 fs.s3a.secret.keywont-tell

 fs.s3a.endpoints3.eu-central-1.amazonaws.com
 
 >>>
 Here is my lib folder with the versions of the aws-sdk and the
 hadoop-aws integration:
 -rw---1 root root   11.4M Mar 20  2014
 aws-java-sdk-1.7.4.jar
 

Re: S3 Access in eu-central-1

2017-11-29 Thread Ufuk Celebi
Hey Dominik,

yes, we should definitely add this to the docs.

@Nico: You recently updated the Flink S3 setup docs. Would you mind
adding these hints for eu-central-1 from Steve? I think that would be
super helpful!

Best,

Ufuk

On Tue, Nov 28, 2017 at 10:00 PM, Dominik Bruhn  wrote:
> Hey Stephan, Hey Steve,
> that was the right hint, adding that open to the Java-Options fixed the
> problem. Maybe we should add this somehow to our Flink Wiki?
>
> Thanks!
>
> Dominik
>
> On 28/11/17 11:55, Stephan Ewen wrote:
>>
>> Got a pointer from Steve that this is answered on Stack Overflow here:
>> https://stackoverflow.com/questions/36154484/aws-java-sdk-manually-set-signature-version
>> 
>>
>> Flink 1.4 contains a specially bundled "fs-s3-hadoop" with smaller no
>> footprint, compatible across Hadoop versions, and based on a later s3a and
>> AWS sdk. In that connector, it should work out of the box because it uses a
>> later AWS SDK. You can also use it with earlier Hadoop versions because
>> dependencies are relocated, so it should not cash/conflict.
>>
>>
>>
>>
>> On Mon, Nov 27, 2017 at 8:58 PM, Stephan Ewen > > wrote:
>>
>> Hi!
>>
>> The endpoint config entry looks correct.
>> I was looking at this issue to see if there are pointers to anything
>> else, but it looks like the explicit endpoint entry is the most
>> important thing: https://issues.apache.org/jira/browse/HADOOP-13324
>> 
>>
>> I cc-ed Steve Loughran, who is Hadoop's S3 expert (sorry Steve for
>> pulling you in again - listening and learning still about the subtle
>> bits and pieces of S3).
>> @Steve are S3 V4 endpoints supported in Hadoop 2.7.x already, or
>> only in Hadoop 2.8?
>>
>> Best,
>> Stephan
>>
>>
>> On Mon, Nov 27, 2017 at 9:47 AM, Dominik Bruhn > > wrote:
>>
>> Hey,
>> can anyone give a hint? Does anyone have flink running with an
>> S3 Bucket in Frankfurt/eu-central-1 and can share his config and
>> setup?
>>
>> Thanks,
>> Dominik
>>
>> On 22. Nov 2017, at 17:52, domi...@dbruhn.de
>>  wrote:
>>
>>> Hey everyone,
>>> I'm trying since hours to get Flink 1.3.2 (downloaded for
>>> hadoop 2.7) to snapshot/checkpoint to an S3 bucket which is
>>> hosted in the eu-central-1 region. Everything works fine for
>>> other regions. I'm running my job on a JobTracker in local
>>> mode. I googled the internet and found several hints, most of
>>> them telling that setting the `fs.s3a.endpoint` should solve
>>> it. It doesn't. I'm also sure that the core-site.xml (see
>>> below) is picked up, if I put garbage into the endpoint then I
>>> receive a hostname not found error.
>>>
>>> The exception I'm getting is:
>>> com.amazonaws.services.s3.model.AmazonS3Exception: Status
>>> Code: 400, AWS Service: Amazon S3, AWS Request ID:
>>> 432415098B0994BC, AWS Error Code: null, AWS Error Message: Bad
>>> Request, S3 Extended Request ID:
>>>
>>> 1PSDe4EOh7zvfNPdWrwoBKKOtsS/gf9atn5movRzcpvIH2WsR+ptXvXyFyEHXjDb3F9AniXgsBQ=
>>>
>>> I read the AWS FAQ but I don't think that
>>>
>>> https://ci.apache.org/projects/flink/flink-docs-release-1.3/setup/aws.html#ioexception-400-bad-request
>>>
>>> 
>>> applies to me as I'm not running the NativeFileSystem.
>>>
>>> I suspect this is related to the v4 signing protocol which is
>>> required for S3 in Frankfurt. Could it be that the aws-sdk
>>> version is just too old? I tried to play around with it but
>>> the hadoop adapter is incompatible with newer versions.
>>>
>>> I have the following core-site.xml:
>>>
>>> 
>>> 
>>>
>>> fs.s3.implorg.apache.hadoop.fs.s3a.S3AFileSystem
>>>
>>> fs.s3a.buffer.dir/tmp
>>>
>>> fs.s3a.access.keysomething
>>>
>>> fs.s3a.secret.keywont-tell
>>>
>>> fs.s3a.endpoints3.eu-central-1.amazonaws.com
>>> 
>>> >>
>>> Here is my lib folder with the versions of the aws-sdk and the
>>> hadoop-aws integration:
>>> -rw---1 root root   11.4M Mar 20  2014
>>> aws-java-sdk-1.7.4.jar
>>> -rw-r--r--1 1005 1006   70.0M Aug  3 12:10
>>> flink-dist_2.11-1.3.2.jar
>>> -rw-rw-r--1 1005 1006   98.3K Aug  3 12:07
>>> flink-python_2.11-1.3.2.jar
>>> -rw-r--r--1 1005 1006   34.9M Aug  3 11:58
>>> flink-shaded-hadoop2-uber-1.3.2.jar
>>> -rw---   

Re: S3 Access in eu-central-1

2017-11-28 Thread Dominik Bruhn

Hey Stephan, Hey Steve,
that was the right hint, adding that open to the Java-Options fixed the 
problem. Maybe we should add this somehow to our Flink Wiki?


Thanks!

Dominik

On 28/11/17 11:55, Stephan Ewen wrote:
Got a pointer from Steve that this is answered on Stack Overflow here: 
https://stackoverflow.com/questions/36154484/aws-java-sdk-manually-set-signature-version 



Flink 1.4 contains a specially bundled "fs-s3-hadoop" with smaller no 
footprint, compatible across Hadoop versions, and based on a later s3a 
and AWS sdk. In that connector, it should work out of the box because it 
uses a later AWS SDK. You can also use it with earlier Hadoop versions 
because dependencies are relocated, so it should not cash/conflict.





On Mon, Nov 27, 2017 at 8:58 PM, Stephan Ewen > wrote:


Hi!

The endpoint config entry looks correct.
I was looking at this issue to see if there are pointers to anything
else, but it looks like the explicit endpoint entry is the most
important thing: https://issues.apache.org/jira/browse/HADOOP-13324


I cc-ed Steve Loughran, who is Hadoop's S3 expert (sorry Steve for
pulling you in again - listening and learning still about the subtle
bits and pieces of S3).
@Steve are S3 V4 endpoints supported in Hadoop 2.7.x already, or
only in Hadoop 2.8?

Best,
Stephan


On Mon, Nov 27, 2017 at 9:47 AM, Dominik Bruhn > wrote:

Hey,
can anyone give a hint? Does anyone have flink running with an
S3 Bucket in Frankfurt/eu-central-1 and can share his config and
setup?

Thanks,
Dominik

On 22. Nov 2017, at 17:52, domi...@dbruhn.de
 wrote:


Hey everyone,
I'm trying since hours to get Flink 1.3.2 (downloaded for
hadoop 2.7) to snapshot/checkpoint to an S3 bucket which is
hosted in the eu-central-1 region. Everything works fine for
other regions. I'm running my job on a JobTracker in local
mode. I googled the internet and found several hints, most of
them telling that setting the `fs.s3a.endpoint` should solve
it. It doesn't. I'm also sure that the core-site.xml (see
below) is picked up, if I put garbage into the endpoint then I
receive a hostname not found error.

The exception I'm getting is:
com.amazonaws.services.s3.model.AmazonS3Exception: Status
Code: 400, AWS Service: Amazon S3, AWS Request ID:
432415098B0994BC, AWS Error Code: null, AWS Error Message: Bad
Request, S3 Extended Request ID:

1PSDe4EOh7zvfNPdWrwoBKKOtsS/gf9atn5movRzcpvIH2WsR+ptXvXyFyEHXjDb3F9AniXgsBQ=

I read the AWS FAQ but I don't think that

https://ci.apache.org/projects/flink/flink-docs-release-1.3/setup/aws.html#ioexception-400-bad-request


applies to me as I'm not running the NativeFileSystem.

I suspect this is related to the v4 signing protocol which is
required for S3 in Frankfurt. Could it be that the aws-sdk
version is just too old? I tried to play around with it but
the hadoop adapter is incompatible with newer versions.

I have the following core-site.xml:



 
fs.s3.implorg.apache.hadoop.fs.s3a.S3AFileSystem
 fs.s3a.buffer.dir/tmp
 
fs.s3a.access.keysomething
 
fs.s3a.secret.keywont-tell
 
fs.s3a.endpoints3.eu-central-1.amazonaws.com







Re: S3 Access in eu-central-1

2017-11-28 Thread Stephan Ewen
Got a pointer from Steve that this is answered on Stack Overflow here:
https://stackoverflow.com/questions/36154484/aws-java-
sdk-manually-set-signature-version

Flink 1.4 contains a specially bundled "fs-s3-hadoop" with smaller no
footprint, compatible across Hadoop versions, and based on a later s3a and
AWS sdk. In that connector, it should work out of the box because it uses a
later AWS SDK. You can also use it with earlier Hadoop versions because
dependencies are relocated, so it should not cash/conflict.




On Mon, Nov 27, 2017 at 8:58 PM, Stephan Ewen  wrote:

> Hi!
>
> The endpoint config entry looks correct.
> I was looking at this issue to see if there are pointers to anything else,
> but it looks like the explicit endpoint entry is the most important thing:
> https://issues.apache.org/jira/browse/HADOOP-13324
>
> I cc-ed Steve Loughran, who is Hadoop's S3 expert (sorry Steve for pulling
> you in again - listening and learning still about the subtle bits and
> pieces of S3).
> @Steve are S3 V4 endpoints supported in Hadoop 2.7.x already, or only in
> Hadoop 2.8?
>
> Best,
> Stephan
>
>
> On Mon, Nov 27, 2017 at 9:47 AM, Dominik Bruhn  wrote:
>
>> Hey,
>> can anyone give a hint? Does anyone have flink running with an S3 Bucket
>> in Frankfurt/eu-central-1 and can share his config and setup?
>>
>> Thanks,
>> Dominik
>>
>> On 22. Nov 2017, at 17:52, domi...@dbruhn.de wrote:
>>
>> Hey everyone,
>> I'm trying since hours to get Flink 1.3.2 (downloaded for hadoop 2.7) to
>> snapshot/checkpoint to an S3 bucket which is hosted in the eu-central-1
>> region. Everything works fine for other regions. I'm running my job on a
>> JobTracker in local mode. I googled the internet and found several hints,
>> most of them telling that setting the `fs.s3a.endpoint` should solve it. It
>> doesn't. I'm also sure that the core-site.xml (see below) is picked up, if
>> I put garbage into the endpoint then I receive a hostname not found error.
>>
>> The exception I'm getting is:
>> com.amazonaws.services.s3.model.AmazonS3Exception: Status Code: 400, AWS
>> Service: Amazon S3, AWS Request ID: 432415098B0994BC, AWS Error Code: null,
>> AWS Error Message: Bad Request, S3 Extended Request ID:
>> 1PSDe4EOh7zvfNPdWrwoBKKOtsS/gf9atn5movRzcpvIH2WsR+ptXvXyFyEH
>> XjDb3F9AniXgsBQ=
>>
>> I read the AWS FAQ but I don't think that https://ci.apache.org/projects
>> /flink/flink-docs-release-1.3/setup/aws.html#ioexception-400-bad-request
>> applies to me as I'm not running the NativeFileSystem.
>>
>> I suspect this is related to the v4 signing protocol which is required
>> for S3 in Frankfurt. Could it be that the aws-sdk version is just too old?
>> I tried to play around with it but the hadoop adapter is incompatible with
>> newer versions.
>>
>> I have the following core-site.xml:
>>
>> 
>> 
>>  fs.s3.implorg.apache.hadoop.f
>> s.s3a.S3AFileSystem
>>  fs.s3a.buffer.dir/tmp
>>  fs.s3a.access.keysomething> value>
>>  fs.s3a.secret.keywont-tell> value>
>>  fs.s3a.endpoints3.eu-central-
>> 1.amazonaws.com
>> >
>> Here is my lib folder with the versions of the aws-sdk and the hadoop-aws
>> integration:
>> -rw---1 root root   11.4M Mar 20  2014
>> aws-java-sdk-1.7.4.jar
>> -rw-r--r--1 1005 1006   70.0M Aug  3 12:10
>> flink-dist_2.11-1.3.2.jar
>> -rw-rw-r--1 1005 1006   98.3K Aug  3 12:07
>> flink-python_2.11-1.3.2.jar
>> -rw-r--r--1 1005 1006   34.9M Aug  3 11:58
>> flink-shaded-hadoop2-uber-1.3.2.jar
>> -rw---1 root root  100.7K Jan 14  2016
>> hadoop-aws-2.7.2.jar
>> -rw---1 root root  414.7K May 17  2012 httpclient-4.2.jar
>> -rw---1 root root  218.0K May  1  2012 httpcore-4.2.jar
>> -rw-rw-r--1 1005 1006  478.4K Jul 28 14:50 log4j-1.2.17.jar
>> -rw-rw-r--1 1005 10068.7K Jul 28 14:50
>> slf4j-log4j12-1.7.7.jar
>>
>> Can anyone give me any hints?
>>
>> Thanks,
>> Dominik
>>
>>
>


Re: S3 Access in eu-central-1

2017-11-27 Thread Stephan Ewen
Hi!

The endpoint config entry looks correct.
I was looking at this issue to see if there are pointers to anything else,
but it looks like the explicit endpoint entry is the most important thing:
https://issues.apache.org/jira/browse/HADOOP-13324

I cc-ed Steve Loughran, who is Hadoop's S3 expert (sorry Steve for pulling
you in again - listening and learning still about the subtle bits and
pieces of S3).
@Steve are S3 V4 endpoints supported in Hadoop 2.7.x already, or only in
Hadoop 2.8?

Best,
Stephan


On Mon, Nov 27, 2017 at 9:47 AM, Dominik Bruhn  wrote:

> Hey,
> can anyone give a hint? Does anyone have flink running with an S3 Bucket
> in Frankfurt/eu-central-1 and can share his config and setup?
>
> Thanks,
> Dominik
>
> On 22. Nov 2017, at 17:52, domi...@dbruhn.de wrote:
>
> Hey everyone,
> I'm trying since hours to get Flink 1.3.2 (downloaded for hadoop 2.7) to
> snapshot/checkpoint to an S3 bucket which is hosted in the eu-central-1
> region. Everything works fine for other regions. I'm running my job on a
> JobTracker in local mode. I googled the internet and found several hints,
> most of them telling that setting the `fs.s3a.endpoint` should solve it. It
> doesn't. I'm also sure that the core-site.xml (see below) is picked up, if
> I put garbage into the endpoint then I receive a hostname not found error.
>
> The exception I'm getting is:
> com.amazonaws.services.s3.model.AmazonS3Exception: Status Code: 400, AWS
> Service: Amazon S3, AWS Request ID: 432415098B0994BC, AWS Error Code: null,
> AWS Error Message: Bad Request, S3 Extended Request ID:
> 1PSDe4EOh7zvfNPdWrwoBKKOtsS/gf9atn5movRzcpvIH2WsR+
> ptXvXyFyEHXjDb3F9AniXgsBQ=
>
> I read the AWS FAQ but I don't think that https://ci.apache.org/
> projects/flink/flink-docs-release-1.3/setup/aws.html#
> ioexception-400-bad-request applies to me as I'm not running the
> NativeFileSystem.
>
> I suspect this is related to the v4 signing protocol which is required for
> S3 in Frankfurt. Could it be that the aws-sdk version is just too old? I
> tried to play around with it but the hadoop adapter is incompatible with
> newer versions.
>
> I have the following core-site.xml:
>
> 
> 
>  fs.s3.implorg.apache.hadoop.
> fs.s3a.S3AFileSystem
>  fs.s3a.buffer.dir/tmp
>  fs.s3a.access.key
> something
>  fs.s3a.secret.keywont-
> tell
>  fs.s3a.endpoints3.eu-
> central-1.amazonaws.com
> 
> Here is my lib folder with the versions of the aws-sdk and the hadoop-aws
> integration:
> -rw---1 root root   11.4M Mar 20  2014
> aws-java-sdk-1.7.4.jar
> -rw-r--r--1 1005 1006   70.0M Aug  3 12:10
> flink-dist_2.11-1.3.2.jar
> -rw-rw-r--1 1005 1006   98.3K Aug  3 12:07
> flink-python_2.11-1.3.2.jar
> -rw-r--r--1 1005 1006   34.9M Aug  3 11:58
> flink-shaded-hadoop2-uber-1.3.2.jar
> -rw---1 root root  100.7K Jan 14  2016 hadoop-aws-2.7.2.jar
> -rw---1 root root  414.7K May 17  2012 httpclient-4.2.jar
> -rw---1 root root  218.0K May  1  2012 httpcore-4.2.jar
> -rw-rw-r--1 1005 1006  478.4K Jul 28 14:50 log4j-1.2.17.jar
> -rw-rw-r--1 1005 10068.7K Jul 28 14:50
> slf4j-log4j12-1.7.7.jar
>
> Can anyone give me any hints?
>
> Thanks,
> Dominik
>
>


Re: S3 Access in eu-central-1

2017-11-27 Thread Dominik Bruhn
Hey,
can anyone give a hint? Does anyone have flink running with an S3 Bucket in 
Frankfurt/eu-central-1 and can share his config and setup?

Thanks,
Dominik

> On 22. Nov 2017, at 17:52, domi...@dbruhn.de wrote:
> 
> Hey everyone,
> I'm trying since hours to get Flink 1.3.2 (downloaded for hadoop 2.7) to 
> snapshot/checkpoint to an S3 bucket which is hosted in the eu-central-1 
> region. Everything works fine for other regions. I'm running my job on a 
> JobTracker in local mode. I googled the internet and found several hints, 
> most of them telling that setting the `fs.s3a.endpoint` should solve it. It 
> doesn't. I'm also sure that the core-site.xml (see below) is picked up, if I 
> put garbage into the endpoint then I receive a hostname not found error.
> 
> The exception I'm getting is:
> com.amazonaws.services.s3.model.AmazonS3Exception: Status Code: 400, AWS 
> Service: Amazon S3, AWS Request ID: 432415098B0994BC, AWS Error Code: null, 
> AWS Error Message: Bad Request, S3 Extended Request ID: 
> 1PSDe4EOh7zvfNPdWrwoBKKOtsS/gf9atn5movRzcpvIH2WsR+ptXvXyFyEHXjDb3F9AniXgsBQ=
> 
> I read the AWS FAQ but I don't think that 
> https://ci.apache.org/projects/flink/flink-docs-release-1.3/setup/aws.html#ioexception-400-bad-request
>  applies to me as I'm not running the NativeFileSystem.
> 
> I suspect this is related to the v4 signing protocol which is required for S3 
> in Frankfurt. Could it be that the aws-sdk version is just too old? I tried 
> to play around with it but the hadoop adapter is incompatible with newer 
> versions.
> 
> I have the following core-site.xml:
> 
> 
> 
>  
> fs.s3.implorg.apache.hadoop.fs.s3a.S3AFileSystem
>  fs.s3a.buffer.dir/tmp
>  fs.s3a.access.keysomething
>  fs.s3a.secret.keywont-tell
>  
> fs.s3a.endpoints3.eu-central-1.amazonaws.com
>  
> Here is my lib folder with the versions of the aws-sdk and the hadoop-aws 
> integration:
> -rw---1 root root   11.4M Mar 20  2014 aws-java-sdk-1.7.4.jar
> -rw-r--r--1 1005 1006   70.0M Aug  3 12:10 
> flink-dist_2.11-1.3.2.jar
> -rw-rw-r--1 1005 1006   98.3K Aug  3 12:07 
> flink-python_2.11-1.3.2.jar
> -rw-r--r--1 1005 1006   34.9M Aug  3 11:58 
> flink-shaded-hadoop2-uber-1.3.2.jar
> -rw---1 root root  100.7K Jan 14  2016 hadoop-aws-2.7.2.jar
> -rw---1 root root  414.7K May 17  2012 httpclient-4.2.jar
> -rw---1 root root  218.0K May  1  2012 httpcore-4.2.jar
> -rw-rw-r--1 1005 1006  478.4K Jul 28 14:50 log4j-1.2.17.jar
> -rw-rw-r--1 1005 10068.7K Jul 28 14:50 slf4j-log4j12-1.7.7.jar
> 
> Can anyone give me any hints?
> 
> Thanks,
> Dominik


Re: S3 Access in eu-central-1

2017-11-22 Thread Timo Walther

@Patrick: Do you have an advice?


Am 11/22/17 um 5:52 PM schrieb domi...@dbruhn.de:

Hey everyone,
I'm trying since hours to get Flink 1.3.2 (downloaded for hadoop 2.7) 
to snapshot/checkpoint to an S3 bucket which is hosted in the 
eu-central-1 region. Everything works fine for other regions. I'm 
running my job on a JobTracker in local mode. I googled the internet 
and found several hints, most of them telling that setting the 
`fs.s3a.endpoint` should solve it. It doesn't. I'm also sure that the 
core-site.xml (see below) is picked up, if I put garbage into the 
endpoint then I receive a hostname not found error.


The exception I'm getting is:
com.amazonaws.services.s3.model.AmazonS3Exception: Status Code: 400, 
AWS Service: Amazon S3, AWS Request ID: 432415098B0994BC, AWS Error 
Code: null, AWS Error Message: Bad Request, S3 Extended Request ID: 
1PSDe4EOh7zvfNPdWrwoBKKOtsS/gf9atn5movRzcpvIH2WsR+ptXvXyFyEHXjDb3F9AniXgsBQ=


I read the AWS FAQ but I don't think that 
https://ci.apache.org/projects/flink/flink-docs-release-1.3/setup/aws.html#ioexception-400-bad-request 
applies to me as I'm not running the NativeFileSystem.


I suspect this is related to the v4 signing protocol which is required 
for S3 in Frankfurt. Could it be that the aws-sdk version is just too 
old? I tried to play around with it but the hadoop adapter is 
incompatible with newer versions.


I have the following core-site.xml:



fs.s3.implorg.apache.hadoop.fs.s3a.S3AFileSystem
fs.s3a.buffer.dir/tmp
fs.s3a.access.keysomething
fs.s3a.secret.keywont-tell
fs.s3a.endpoints3.eu-central-1.amazonaws.com
Here is my lib folder with the versions of the aws-sdk and the 
hadoop-aws integration:
-rw---    1 root root   11.4M Mar 20  2014 
aws-java-sdk-1.7.4.jar
-rw-r--r--    1 1005 1006   70.0M Aug  3 12:10 
flink-dist_2.11-1.3.2.jar
-rw-rw-r--    1 1005 1006   98.3K Aug  3 12:07 
flink-python_2.11-1.3.2.jar
-rw-r--r--    1 1005 1006   34.9M Aug  3 11:58 
flink-shaded-hadoop2-uber-1.3.2.jar
-rw---    1 root root  100.7K Jan 14  2016 
hadoop-aws-2.7.2.jar

-rw---    1 root root  414.7K May 17  2012 httpclient-4.2.jar
-rw---    1 root root  218.0K May  1  2012 httpcore-4.2.jar
-rw-rw-r--    1 1005 1006  478.4K Jul 28 14:50 log4j-1.2.17.jar
-rw-rw-r--    1 1005 1006    8.7K Jul 28 14:50 
slf4j-log4j12-1.7.7.jar


Can anyone give me any hints?

Thanks,
Dominik





S3 Access in eu-central-1

2017-11-22 Thread dominik

Hey everyone,
I'm trying since hours to get Flink 1.3.2 (downloaded for hadoop 2.7) to 
snapshot/checkpoint to an S3 bucket which is hosted in the eu-central-1 
region. Everything works fine for other regions. I'm running my job on a 
JobTracker in local mode. I googled the internet and found several 
hints, most of them telling that setting the `fs.s3a.endpoint` should 
solve it. It doesn't. I'm also sure that the core-site.xml (see below) 
is picked up, if I put garbage into the endpoint then I receive a 
hostname not found error.


The exception I'm getting is:
com.amazonaws.services.s3.model.AmazonS3Exception: Status Code: 400, AWS 
Service: Amazon S3, AWS Request ID: 432415098B0994BC, AWS Error Code: 
null, AWS Error Message: Bad Request, S3 Extended Request ID: 
1PSDe4EOh7zvfNPdWrwoBKKOtsS/gf9atn5movRzcpvIH2WsR+ptXvXyFyEHXjDb3F9AniXgsBQ=


I read the AWS FAQ but I don't think that 
https://ci.apache.org/projects/flink/flink-docs-release-1.3/setup/aws.html#ioexception-400-bad-request 
applies to me as I'm not running the NativeFileSystem.


I suspect this is related to the v4 signing protocol which is required 
for S3 in Frankfurt. Could it be that the aws-sdk version is just too 
old? I tried to play around with it but the hadoop adapter is 
incompatible with newer versions.


I have the following core-site.xml:



  
fs.s3.implorg.apache.hadoop.fs.s3a.S3AFileSystem

  fs.s3a.buffer.dir/tmp
  
fs.s3a.access.keysomething
  
fs.s3a.secret.keywont-tell
  
fs.s3a.endpoints3.eu-central-1.amazonaws.com

Here is my lib folder with the versions of the aws-sdk and the 
hadoop-aws integration:
-rw---1 root root   11.4M Mar 20  2014 
aws-java-sdk-1.7.4.jar
-rw-r--r--1 1005 1006   70.0M Aug  3 12:10 
flink-dist_2.11-1.3.2.jar
-rw-rw-r--1 1005 1006   98.3K Aug  3 12:07 
flink-python_2.11-1.3.2.jar
-rw-r--r--1 1005 1006   34.9M Aug  3 11:58 
flink-shaded-hadoop2-uber-1.3.2.jar
-rw---1 root root  100.7K Jan 14  2016 
hadoop-aws-2.7.2.jar
-rw---1 root root  414.7K May 17  2012 
httpclient-4.2.jar

-rw---1 root root  218.0K May  1  2012 httpcore-4.2.jar
-rw-rw-r--1 1005 1006  478.4K Jul 28 14:50 log4j-1.2.17.jar
-rw-rw-r--1 1005 10068.7K Jul 28 14:50 
slf4j-log4j12-1.7.7.jar


Can anyone give me any hints?

Thanks,
Dominik