[jira] [Resolved] (HADOOP-15182) Support to change back to signature version 2 of AWS SDK

2018-01-21 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-15182?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger resolved HADOOP-15182.
-
Resolution: Not A Problem

> Support to change back to signature version 2 of AWS SDK
> 
>
> Key: HADOOP-15182
> URL: https://issues.apache.org/jira/browse/HADOOP-15182
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.9.0
> Environment:  
>  
>  
>Reporter: Yonger
>Priority: Minor
>
> Current s3a depend on  aws-java-sdk-bundle-1.11.199 which use signature v4. 
> So for some s3-compatible system(Ceph) which still using v2, Hadoop can't 
> work on them.
> s3cmd can use v2 via specify option like :
> {code:java}
> s3cmd --signature-v2 ls s3://xxx/{code}
>  
> maybe we can add a parameter to allow back to use signature v2 in s3a.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15182) Support to change back to signature version 2 of AWS SDK

2018-01-21 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-15182?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16333778#comment-16333778
 ] 

Yonger commented on HADOOP-15182:
-

Just know how to roll back to signature v2, it's a bit hard to find out this 
approach for me.
{code:java}

 fs.s3a.signing-algorithm
 S3SignerType
 {code}
Sorry to wast your time, I will close this ticket.

> Support to change back to signature version 2 of AWS SDK
> 
>
> Key: HADOOP-15182
> URL: https://issues.apache.org/jira/browse/HADOOP-15182
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.9.0
> Environment:  
>  
>  
>Reporter: Yonger
>Priority: Minor
>
> Current s3a depend on  aws-java-sdk-bundle-1.11.199 which use signature v4. 
> So for some s3-compatible system(Ceph) which still using v2, Hadoop can't 
> work on them.
> s3cmd can use v2 via specify option like :
> {code:java}
> s3cmd --signature-v2 ls s3://xxx/{code}
>  
> maybe we can add a parameter to allow back to use signature v2 in s3a.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15182) Support to change back to signature version 2 of AWS SDK

2018-01-18 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-15182?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16331554#comment-16331554
 ] 

Yonger commented on HADOOP-15182:
-

You means  "fs.s3a.signing-algorithm"?, but I can't find an algorithm to match 
v2:
{code:java}
static {
 // Register the standard signer types.
 SIGNERS.put(QUERY_STRING_SIGNER, QueryStringSigner.class);
 SIGNERS.put(VERSION_THREE_SIGNER, AWS3Signer.class);
 SIGNERS.put(VERSION_FOUR_SIGNER, AWS4Signer.class);
 SIGNERS.put(VERSION_FOUR_UNSIGNED_PAYLOAD_SIGNER, 
AWS4UnsignedPayloadSigner.class);
 SIGNERS.put(NO_OP_SIGNER, NoOpSigner.class);
}{code}

> Support to change back to signature version 2 of AWS SDK
> 
>
> Key: HADOOP-15182
> URL: https://issues.apache.org/jira/browse/HADOOP-15182
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.9.0
> Environment:  
>  
>  
>Reporter: Yonger
>Priority: Minor
>
> Current s3a depend on  aws-java-sdk-bundle-1.11.199 which use signature v4. 
> So for some s3-compatible system(Ceph) which still using v2, Hadoop can't 
> work on them.
> s3cmd can use v2 via specify option like :
> {code:java}
> s3cmd --signature-v2 ls s3://xxx/{code}
>  
> maybe we can add a parameter to allow back to use signature v2 in s3a.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14943) Add common getFileBlockLocations() emulation for object stores, including S3A

2017-12-26 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14943?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16303790#comment-16303790
 ] 

Yonger commented on HADOOP-14943:
-

[~ste...@apache.org] As Hadoop running on s3-compatible storage, compute and 
storage separately, so I called the host that running MR/Spark/Hive application 
"compute host". 

> Add common getFileBlockLocations() emulation for object stores, including S3A
> -
>
> Key: HADOOP-14943
> URL: https://issues.apache.org/jira/browse/HADOOP-14943
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.1
>Reporter: Steve Loughran
>Assignee: Steve Loughran
> Attachments: HADOOP-14943-001.patch, HADOOP-14943-002.patch, 
> HADOOP-14943-002.patch, HADOOP-14943-003.patch
>
>
> It looks suspiciously like S3A isn't providing the partitioning data needed 
> in {{listLocatedStatus}} and {{getFileBlockLocations()}} needed to break up a 
> file by the blocksize. This will stop tools using the MRv1 APIS doing the 
> partitioning properly if the input format isn't doing it own split logic.
> FileInputFormat in MRv2 is a bit more configurable about input split 
> calculation & will split up large files. but otherwise, the partitioning is 
> being done more by the default values of the executing engine, rather than 
> any config data from the filesystem about what its "block size" is,
> NativeAzureFS does a better job; maybe that could be factored out to 
> hadoop-common and reused?



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-15087) Write directly without creating temp directory to avoid rename

2017-12-03 Thread Yonger (JIRA)
Yonger created HADOOP-15087:
---

 Summary: Write directly without creating temp directory to avoid 
rename 
 Key: HADOOP-15087
 URL: https://issues.apache.org/jira/browse/HADOOP-15087
 Project: Hadoop Common
  Issue Type: Improvement
  Components: fs/s3
Reporter: Yonger


Rename in workloads like Teragen/Terasort who use Hadoop default 
outputcommitters really hurt performance a lot. 
Stocator announce it doesn't create the temporary directories any all, and 
still preserves Hadoop's fault tolerance. I add a switch when creating file via 
integrating it's code into s3a, I got 5x performance gain in Teragen and 15% 
performance improvement in Terasort.

 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-11-27 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16268148#comment-16268148
 ] 

Yonger commented on HADOOP-14475:
-

[~mackrorysd] Thank you.  I start to understand your logic, and I think you are 
right.
{code:java}
 There's no guarantee that metrics source names would even be consistent among 
all JVMs for a given bucket, since they're assigned numbers in the order that 
they're created
{code}
I can get the info what i want to by aggregating based on bucket instead of 
metric source name which is not still unique in multiple JVM processes.


> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: HADOOP-14475-003.patch, HADOOP-14475.002.patch, 
> HADOOP-14475.005.patch, HADOOP-14475.006.patch, HADOOP-14475.008.patch, 
> HADOOP-14475.009.patch, HADOOP-14475.010.patch, HADOOP-14475.011.patch, 
> HADOOP-14475.012.patch, HADOOP-14475.013.patch, HADOOP-14475.014.patch, 
> HADOOP-14475.015.patch, HADOOP-14775.007.patch, failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, failsafe-report-scale.html, 
> failsafe-report-scale.zip, s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-11-27 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16268051#comment-16268051
 ] 

Yonger commented on HADOOP-14475:
-

[~mackrorysd] So our final output format of record doesn't include any metric 
source info? if yes, I think it is not friendly to make a statistic chart like 
via InfluxDB+Grafana, because only fsId is unique in a record, especially there 
are multiple metric sources registered with the same bucket(I don't know why, 
but they were exist in my test), whose output records can't be easy to 
distinguish except fsid, then your chart in Grafana is hard readable.

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: HADOOP-14475-003.patch, HADOOP-14475.002.patch, 
> HADOOP-14475.005.patch, HADOOP-14475.006.patch, HADOOP-14475.008.patch, 
> HADOOP-14475.009.patch, HADOOP-14475.010.patch, HADOOP-14475.011.patch, 
> HADOOP-14475.012.patch, HADOOP-14475.013.patch, HADOOP-14475.014.patch, 
> HADOOP-14475.015.patch, HADOOP-14775.007.patch, failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, failsafe-report-scale.html, 
> failsafe-report-scale.zip, s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-11-27 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16266866#comment-16266866
 ] 

Yonger commented on HADOOP-14475:
-

azure just give the registry name to the record, which is confused. We can't 
distinguish the different records if there are more than one relative metric 
source registered within a process, what I given the example in previous 
comments.  

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: HADOOP-14475-003.patch, HADOOP-14475.002.patch, 
> HADOOP-14475.005.patch, HADOOP-14475.006.patch, HADOOP-14475.008.patch, 
> HADOOP-14475.009.patch, HADOOP-14475.010.patch, HADOOP-14475.011.patch, 
> HADOOP-14475.012.patch, HADOOP-14475.013.patch, HADOOP-14775.007.patch, 
> failsafe-report-s3a-it.html, failsafe-report-s3a-scale.html, 
> failsafe-report-scale.html, failsafe-report-scale.zip, s3a-metrics.patch1, 
> stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14943) Add common getFileBlockLocations() emulation for object stores, including S3A

2017-11-27 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14943?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16266856#comment-16266856
 ] 

Yonger commented on HADOOP-14943:
-

[~ste...@apache.org]I remember there are some discussion about how to configure 
the fake host list, such as returning endpoint, compute hosts and a star, is 
this right? I am not sure whether i understand these points totally.

I just test these four cases with 1TB dataset on query42 of TPC-DS, results are 
below(seconds):

||default localhost||endpoint||star||compute host list||
|16|16l 16|28|

>From this result, performance are equal in these cases except returning 
>compute host list.



> Add common getFileBlockLocations() emulation for object stores, including S3A
> -
>
> Key: HADOOP-14943
> URL: https://issues.apache.org/jira/browse/HADOOP-14943
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.1
>Reporter: Steve Loughran
>Assignee: Steve Loughran
> Attachments: HADOOP-14943-001.patch, HADOOP-14943-002.patch, 
> HADOOP-14943-002.patch, HADOOP-14943-003.patch
>
>
> It looks suspiciously like S3A isn't providing the partitioning data needed 
> in {{listLocatedStatus}} and {{getFileBlockLocations()}} needed to break up a 
> file by the blocksize. This will stop tools using the MRv1 APIS doing the 
> partitioning properly if the input format isn't doing it own split logic.
> FileInputFormat in MRv2 is a bit more configurable about input split 
> calculation & will split up large files. but otherwise, the partitioning is 
> being done more by the default values of the executing engine, rather than 
> any config data from the filesystem about what its "block size" is,
> NativeAzureFS does a better job; maybe that could be factored out to 
> hadoop-common and reused?



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-14943) Add common getFileBlockLocations() emulation for object stores, including S3A

2017-11-27 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14943?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16266856#comment-16266856
 ] 

Yonger edited comment on HADOOP-14943 at 11/27/17 2:15 PM:
---

[~ste...@apache.org]I remember there are some discussion about how to configure 
the fake host list, such as returning endpoint, compute hosts and a star, is 
this right? I am not sure whether i understand these points totally.

I just test these four cases with 1TB dataset on query42 of TPC-DS, results are 
below(seconds):

||default localhost||endpoint||star||compute host list||
|16|16|16|28|

>From this result, performance are equal in these cases except returning 
>compute host list.




was (Author: iyonger):
[~ste...@apache.org]I remember there are some discussion about how to configure 
the fake host list, such as returning endpoint, compute hosts and a star, is 
this right? I am not sure whether i understand these points totally.

I just test these four cases with 1TB dataset on query42 of TPC-DS, results are 
below(seconds):

||default localhost||endpoint||star||compute host list||
|16|16l 16|28|

>From this result, performance are equal in these cases except returning 
>compute host list.



> Add common getFileBlockLocations() emulation for object stores, including S3A
> -
>
> Key: HADOOP-14943
> URL: https://issues.apache.org/jira/browse/HADOOP-14943
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.1
>Reporter: Steve Loughran
>Assignee: Steve Loughran
> Attachments: HADOOP-14943-001.patch, HADOOP-14943-002.patch, 
> HADOOP-14943-002.patch, HADOOP-14943-003.patch
>
>
> It looks suspiciously like S3A isn't providing the partitioning data needed 
> in {{listLocatedStatus}} and {{getFileBlockLocations()}} needed to break up a 
> file by the blocksize. This will stop tools using the MRv1 APIS doing the 
> partitioning properly if the input format isn't doing it own split logic.
> FileInputFormat in MRv2 is a bit more configurable about input split 
> calculation & will split up large files. but otherwise, the partitioning is 
> being done more by the default values of the executing engine, rather than 
> any config data from the filesystem about what its "block size" is,
> NativeAzureFS does a better job; maybe that could be factored out to 
> hadoop-common and reused?



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-11-21 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16260436#comment-16260436
 ] 

Yonger commented on HADOOP-14475:
-

[~mackrorysd]Thanks for refining the code, that looks pretty good for me.
The only thing want to discuss with you: 
>From the latest code, 

{code:java}
+String msName = METRICS_SOURCE_BASENAME + number;
+metricsSourceName = msName + "-" + name.getHost();
+this.recordName = metricsSourceName; 
{code}
I think we don't need to add name.getHost to record name, because each record 
has a field "fsURI" that include the host name/bucket name.




> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: HADOOP-14475-003.patch, HADOOP-14475.002.patch, 
> HADOOP-14475.005.patch, HADOOP-14475.006.patch, HADOOP-14475.008.patch, 
> HADOOP-14475.009.patch, HADOOP-14475.010.patch, HADOOP-14475.011.patch, 
> HADOOP-14475.012.patch, HADOOP-14475.013.patch, HADOOP-14775.007.patch, 
> failsafe-report-s3a-it.html, failsafe-report-s3a-scale.html, 
> failsafe-report-scale.html, failsafe-report-scale.zip, s3a-metrics.patch1, 
> stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-11-14 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16252976#comment-16252976
 ] 

Yonger commented on HADOOP-14475:
-

[~mackrorysd]Thank you.

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: HADOOP-14475-003.patch, HADOOP-14475.002.patch, 
> HADOOP-14475.005.patch, HADOOP-14475.006.patch, HADOOP-14475.008.patch, 
> HADOOP-14775.007.patch, failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, failsafe-report-scale.html, 
> failsafe-report-scale.zip, s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-11-14 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16252804#comment-16252804
 ] 

Yonger edited comment on HADOOP-14475 at 11/15/17 1:44 AM:
---

For the # 2 point of my update
--
without this update,the metrics be flush into file like this(each record with 
same URI mark the same "context+registry name" as begin):

{code:java}
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=dd721794-8269-44eb-911f-a39caa2e6b34-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=8fe27f60-8b3d-41c1-aa84-8bb812315d55-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=6bcf9a67-2dcd-4533-aa5b-2ded7a8ba038-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
{code}
Note, these three records written at the same time instead of different 
intervals, but they should be consider three different kinds of 
metrics(different fsid,different values in real logs).
(Additional, I don't know why the metrics be registered many times within a 
process,like why we need to method "newMetricsSourceName")

with this update, each record will be easy to distinguish:

{code:java}
1510631110244 s3afs.S3aFileSystemMetrics: Context=s3afs, 
FileSystemId=dd721794-8269-44eb-911f-a39caa2e6b34-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3aFileSystemMetrics2: Context=s3afs, 
FileSystemId=8fe27f60-8b3d-41c1-aa84-8bb812315d55-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3aFileSystemMetrics3: Context=s3afs, 
FileSystemId=6bcf9a67-2dcd-4533-aa5b-2ded7a8ba038-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
{code}
 


was (Author: iyonger):
For the # 2 point of my update
--
without this update,the metrics be flush into file like this(each record with 
same URI mark the same "context+registry name" as begin):

{code:java}
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=dd721794-8269-44eb-911f-a39caa2e6b34-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=8fe27f60-8b3d-41c1-aa84-8bb812315d55-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=6bcf9a67-2dcd-4533-aa5b-2ded7a8ba038-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
{code}
Note, these three records written at the same time instead of different 
intervals, so they should be consider three different kinds of metrics.
(Additional, I don't know why the metrics be registered many times within a 
process,like why we need to method "newMetricsSourceName")

with this update, each record will be easy to distinguish:

{code:java}
1510631110244 s3afs.S3aFileSystemMetrics: Context=s3afs, 
FileSystemId=dd721794-8269-44eb-911f-a39caa2e6b34-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3aFileSystemMetrics2: Context=s3afs, 
FileSystemId=8fe27f60-8b3d-41c1-aa84-8bb812315d55-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3aFileSystemMetrics3: Context=s3afs, 
FileSystemId=6bcf9a67-2dcd-4533-aa5b-2ded7a8ba038-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
{code}
 

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: HADOOP-14475-003.patch, HADOOP-14475.002.patch, 
> HADOOP-14475.005.patch, HADOOP-14475.006.patch, HADOOP-14475.008.patch, 
> HADOOP-14775.007.patch, failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, failsafe-report-scale.html, 
> failsafe-report-scale.zip, s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.si

[jira] [Comment Edited] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-11-14 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16252804#comment-16252804
 ] 

Yonger edited comment on HADOOP-14475 at 11/15/17 1:42 AM:
---

For the # 2 point of my update
--
without this update,the metrics be flush into file like this(each record with 
same URI mark the same "context+registry name" as begin):

{code:java}
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=dd721794-8269-44eb-911f-a39caa2e6b34-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=8fe27f60-8b3d-41c1-aa84-8bb812315d55-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=6bcf9a67-2dcd-4533-aa5b-2ded7a8ba038-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
{code}
Note, these three records written at the same time instead of different 
intervals, so they should be consider three different kinds of metrics.
(Additional, I don't know why the metrics be registered many times within a 
process,like why we need to method "newMetricsSourceName")

with this update, each record will be easy to distinguish:

{code:java}
1510631110244 s3afs.S3aFileSystemMetrics: Context=s3afs, 
FileSystemId=dd721794-8269-44eb-911f-a39caa2e6b34-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3aFileSystemMetrics2: Context=s3afs, 
FileSystemId=8fe27f60-8b3d-41c1-aa84-8bb812315d55-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3aFileSystemMetrics3: Context=s3afs, 
FileSystemId=6bcf9a67-2dcd-4533-aa5b-2ded7a8ba038-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
{code}
 


was (Author: iyonger):
For the # 2 point of my update
--
without this update,the metrics be flush into file like this(each record with 
same URI mark the same "context+registry name" as begin):

{code:java}
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=dd721794-8269-44eb-911f-a39caa2e6b34-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=8fe27f60-8b3d-41c1-aa84-8bb812315d55-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=6bcf9a67-2dcd-4533-aa5b-2ded7a8ba038-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
{code}


with this update, each record will be easy to distinguish:

{code:java}
1510631110244 s3afs.S3aFileSystemMetrics: Context=s3afs, 
FileSystemId=dd721794-8269-44eb-911f-a39caa2e6b34-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3aFileSystemMetrics2: Context=s3afs, 
FileSystemId=8fe27f60-8b3d-41c1-aa84-8bb812315d55-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3aFileSystemMetrics3: Context=s3afs, 
FileSystemId=6bcf9a67-2dcd-4533-aa5b-2ded7a8ba038-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
{code}
 

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: HADOOP-14475-003.patch, HADOOP-14475.002.patch, 
> HADOOP-14475.005.patch, HADOOP-14475.006.patch, HADOOP-14475.008.patch, 
> HADOOP-14775.007.patch, failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, failsafe-report-scale.html, 
> failsafe-report-scale.zip, s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent

[jira] [Comment Edited] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-11-14 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16252804#comment-16252804
 ] 

Yonger edited comment on HADOOP-14475 at 11/15/17 1:31 AM:
---

For the # 2 point of my update
--
without this update,the metrics be flush into file like this(each record with 
same URI mark the same "context+registry name" as begin):

{code:java}
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=dd721794-8269-44eb-911f-a39caa2e6b34-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=8fe27f60-8b3d-41c1-aa84-8bb812315d55-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=6bcf9a67-2dcd-4533-aa5b-2ded7a8ba038-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
{code}


with this update, each record will be easy to distinguish:

{code:java}
1510631110244 s3afs.S3aFileSystemMetrics: Context=s3afs, 
FileSystemId=dd721794-8269-44eb-911f-a39caa2e6b34-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3aFileSystemMetrics2: Context=s3afs, 
FileSystemId=8fe27f60-8b3d-41c1-aa84-8bb812315d55-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3aFileSystemMetrics3: Context=s3afs, 
FileSystemId=6bcf9a67-2dcd-4533-aa5b-2ded7a8ba038-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
{code}
 


was (Author: iyonger):
For the # 2 point of my update
--
without this update,the metrics be flush into file like this(each record mark 
the same "context+registry name" as begin):

{code:java}
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=dd721794-8269-44eb-911f-a39caa2e6b34-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=8fe27f60-8b3d-41c1-aa84-8bb812315d55-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=6bcf9a67-2dcd-4533-aa5b-2ded7a8ba038-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
{code}


with this update, each record will be easy to distinguish:

{code:java}
1510631110244 s3afs.S3aFileSystemMetrics: Context=s3afs, 
FileSystemId=dd721794-8269-44eb-911f-a39caa2e6b34-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3aFileSystemMetrics2: Context=s3afs, 
FileSystemId=8fe27f60-8b3d-41c1-aa84-8bb812315d55-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3aFileSystemMetrics3: Context=s3afs, 
FileSystemId=6bcf9a67-2dcd-4533-aa5b-2ded7a8ba038-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
{code}
 

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: HADOOP-14475-003.patch, HADOOP-14475.002.patch, 
> HADOOP-14475.005.patch, HADOOP-14475.006.patch, HADOOP-14475.008.patch, 
> HADOOP-14775.007.patch, failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, failsafe-report-scale.html, 
> failsafe-report-scale.zip, s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-11-14 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16252804#comment-16252804
 ] 

Yonger commented on HADOOP-14475:
-

For the # 2 point of my update
--
without this update,the metrics be flush into file like this(each record mark 
the same "context+registry name" as begin):

{code:java}
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=dd721794-8269-44eb-911f-a39caa2e6b34-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=8fe27f60-8b3d-41c1-aa84-8bb812315d55-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3AFileSystem: Context=s3afs, 
FileSystemId=6bcf9a67-2dcd-4533-aa5b-2ded7a8ba038-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
{code}


with this update, each record will be easy to distinguish:

{code:java}
1510631110244 s3afs.S3aFileSystemMetrics: Context=s3afs, 
FileSystemId=dd721794-8269-44eb-911f-a39caa2e6b34-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3aFileSystemMetrics2: Context=s3afs, 
FileSystemId=8fe27f60-8b3d-41c1-aa84-8bb812315d55-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
1510631110244 s3afs.S3aFileSystemMetrics3: Context=s3afs, 
FileSystemId=6bcf9a67-2dcd-4533-aa5b-2ded7a8ba038-bdaas-demo-dfs, 
fsURI=s3a://bdaas-demo-dfs/, Hostname=client01, xxx
{code}
 

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: HADOOP-14475-003.patch, HADOOP-14475.002.patch, 
> HADOOP-14475.005.patch, HADOOP-14475.006.patch, HADOOP-14475.008.patch, 
> HADOOP-14775.007.patch, failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, failsafe-report-scale.html, 
> failsafe-report-scale.zip, s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-11-14 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16252798#comment-16252798
 ] 

Yonger commented on HADOOP-14475:
-

[~macdonsp]the .008 patch is newer than .007 patch, I think it's the correct 
code that I want to give you guys.

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: HADOOP-14475-003.patch, HADOOP-14475.002.patch, 
> HADOOP-14475.005.patch, HADOOP-14475.006.patch, HADOOP-14475.008.patch, 
> HADOOP-14775.007.patch, failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, failsafe-report-scale.html, 
> failsafe-report-scale.zip, s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-11-13 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16251020#comment-16251020
 ] 

Yonger commented on HADOOP-14475:
-

[~mackrorysd] Thank you very much.
For the test case, I removed the assertion for file status counter while it's 
not necessary.  
For  hard cord, that I copy from some other place(maybe old azure fs 
implementation, forgot), now I don't use fix length string to append, so this 
hard code also removed.

At last, I just only put this setting into metrics property file:
s3afilesystem.sink.file.filename=s3afilesystem-metrics.out
which would print all metrics info into the specify file.


> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: HADOOP-14475-003.patch, HADOOP-14475.002.patch, 
> HADOOP-14475.005.patch, HADOOP-14475.006.patch, HADOOP-14475.008.patch, 
> HADOOP-14775.007.patch, failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, failsafe-report-scale.html, 
> failsafe-report-scale.zip, s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-11-13 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Attachment: HADOOP-14475.008.patch

update:
1.chang prefix and context name to dfs style
2.give each record a different name by the generated metric source name
3.add s3a metric system destroy logic in file system close method

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: HADOOP-14475-003.patch, HADOOP-14475.002.patch, 
> HADOOP-14475.005.patch, HADOOP-14475.006.patch, HADOOP-14475.008.patch, 
> HADOOP-14775.007.patch, failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, failsafe-report-scale.html, 
> failsafe-report-scale.zip, s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15000) s3a new getdefaultblocksize be called in getFileStatus which has not been implemented in s3afilesystem yet

2017-11-01 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-15000?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16235037#comment-16235037
 ] 

Yonger commented on HADOOP-15000:
-

Thanks for your detail interpretation.
But s3a has it's own configuration for block size, if we don't implement the 
newer, callers will use default dfs block size instead.

So we still need to add the newer implementation into s3a code, right ?

> s3a new getdefaultblocksize be called in getFileStatus which has not been 
> implemented in s3afilesystem yet
> --
>
> Key: HADOOP-15000
> URL: https://issues.apache.org/jira/browse/HADOOP-15000
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Affects Versions: 2.9.0
>Reporter: Yonger
>Priority: Minor
>
> new implementation of getting block size has been called in getFileStatus 
> method: 
> {code:java}
>   return new S3AFileStatus(meta.getContentLength(),
>   dateToLong(meta.getLastModified()),
>   path,
>   getDefaultBlockSize(path),
>   username);
> }
> {code}
> while we don't implement it in our s3afilesystem currently, also we need to 
> implement this new method as the old one deprecated.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-15000) s3a new getdefaultblocksize be called in getFileStatus which has not been implemented in s3afilesystem yet

2017-11-01 Thread Yonger (JIRA)
Yonger created HADOOP-15000:
---

 Summary: s3a new getdefaultblocksize be called in getFileStatus 
which has not been implemented in s3afilesystem yet
 Key: HADOOP-15000
 URL: https://issues.apache.org/jira/browse/HADOOP-15000
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs/s3
Reporter: Yonger
Priority: Normal


new implementation of getting block size has been called in getFileStatus 
method: 
{code:java}
  return new S3AFileStatus(meta.getContentLength(),
  dateToLong(meta.getLastModified()),
  path,
  getDefaultBlockSize(path),
  username);
}
{code}

while we don't implement it in our s3afilesystem currently, also we need to 
implement this new method as the old one deprecated.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-10-25 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Status: Patch Available  (was: Open)

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: HADOOP-14475-003.patch, HADOOP-14475.002.patch, 
> HADOOP-14475.005.patch, HADOOP-14475.006.patch, HADOOP-14775.007.patch, 
> failsafe-report-s3a-it.html, failsafe-report-s3a-scale.html, 
> failsafe-report-scale.html, failsafe-report-scale.zip, s3a-metrics.patch1, 
> stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-10-25 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Attachment: HADOOP-14775.007.patch

based trunk.

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: HADOOP-14475-003.patch, HADOOP-14475.002.patch, 
> HADOOP-14475.005.patch, HADOOP-14475.006.patch, HADOOP-14775.007.patch, 
> failsafe-report-s3a-it.html, failsafe-report-s3a-scale.html, 
> failsafe-report-scale.html, failsafe-report-scale.zip, s3a-metrics.patch1, 
> stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-10-25 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Status: Open  (was: Patch Available)

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: HADOOP-14475-003.patch, HADOOP-14475.002.patch, 
> HADOOP-14475.005.patch, HADOOP-14475.006.patch, failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, failsafe-report-scale.html, 
> failsafe-report-scale.zip, s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-14770) S3A http connection in s3a driver not reuse in Spark application

2017-09-21 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14770?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger resolved HADOOP-14770.
-
Resolution: Duplicate

Thanks Steve. when we apply the read input policy with random in our workload 
after upgrade to Hadoop 2.8.1, it works as my expect, connections not destroyed 
every time but reused.

> S3A http connection in s3a driver not reuse in Spark application
> 
>
> Key: HADOOP-14770
> URL: https://issues.apache.org/jira/browse/HADOOP-14770
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.7.3
>Reporter: Yonger
>Assignee: Yonger
>Priority: Minor
>
> I print out connection stats every 2 s when running Spark application against 
> s3-compatible storage:
> {code}
> ESTAB  0  0 :::10.0.2.36:6
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44454
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44374
> :::10.0.2.254:80 
> ESTAB  159724 0 :::10.0.2.36:44436
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:8
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44338
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44438
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44414
> :::10.0.2.254:80 
> ESTAB  0  480   :::10.0.2.36:44450
> :::10.0.2.254:80  timer:(on,170ms,0)
> ESTAB  0  0 :::10.0.2.36:2
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44390
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44326
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44452
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44394
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:4
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44456
> :::10.0.2.254:80 
> ==
> ESTAB  0  0 :::10.0.2.36:44508
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44476
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44524
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44374
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44500
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44504
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44512
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44506
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44464
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44518
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44510
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:2
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44526
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44472
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44466
> :::10.0.2.254:80 
> {code}
> the connection in the above of "=" and below were changed all the time. But 
> this haven't seen in MR application. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14774) S3A case "testRandomReadOverBuffer" failed due to improper range parameter

2017-09-07 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14774?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16156959#comment-16156959
 ] 

Yonger commented on HADOOP-14774:
-

[~ste...@apache.org] Is the latest patch OK?

> S3A case "testRandomReadOverBuffer" failed due to improper range parameter
> --
>
> Key: HADOOP-14774
> URL: https://issues.apache.org/jira/browse/HADOOP-14774
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: Hadoop 2.8.0  
> s3-compatible storage 
>Reporter: Yonger
>Assignee: Yonger
>Priority: Minor
> Attachments: failsafe-report.html, HADOOP-14774.001.patch, 
> HADOOP-14774.002.patch, HADOOP-14774.003.patch
>
>
> {code:java}
> Running org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 2.821 sec <<< 
> FAILURE! - in org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> testRandomReadOverBuffer(org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance)
>   Time elapsed: 2.605 sec  <<< FAILURE!
> java.lang.AssertionError: expected:<8192> but was:<8193>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotEquals(Assert.java:743)
>   at org.junit.Assert.assertEquals(Assert.java:118)
>   at org.junit.Assert.assertEquals(Assert.java:555)
>   at org.junit.Assert.assertEquals(Assert.java:542)
>   at 
> org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance.testRandomReadOverBuffer(ITestS3AInputStreamPerformance.java:533)
> {code}
> From log, the length of content is exceed than our expect:
> {code:java}
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(280)) - >> GET 
> /test-aws-s3a/test/testReadOverBuffer.bin HTTP/1.1
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Host: 10.0.2.254
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> 
> x-amz-content-sha256: 
> e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Authorization: 
> AWS4-HMAC-SHA256 
> Credential=JFDAM9KF9IY8S5P0JIV6/20170815/us-east-1/s3/aws4_request, 
> SignedHeaders=content-type;host;range;user-agent;x-amz-content-sha256;x-amz-date,
>  Signature=42bce4a43d2b1bf6e6d599613c60812e6716514da4ef5b3839ef0566c31279ee
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> X-Amz-Date: 
> 20170815T085316Z
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> User-Agent: Hadoop 
> 2.8.0, aws-sdk-java/1.10.6 Linux/3.10.0-514.21.2.el7.x86_64 
> Java_HotSpot(TM)_64-Bit_Server_VM/25.131-b11/1.8.0_131
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Range: bytes=0-8192
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Content-Type: 
> application/x-www-form-urlencoded; charset=utf-8
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Connection: 
> Keep-Alive
> 2017-08-15 16:53:16,473 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "HTTP/1.1 206 Partial Content[\r][\n]"
> 2017-08-15 16:53:16,475 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Range: bytes 0-8192/32768[\r][\n]"
> 2017-08-15 16:53:16,476 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Length: 8193[\r][\n]"
> 2017-08-15 16:53:16,477 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Accept-Ranges: bytes[\r][\n]"
> 2017-08-15 16:53:16,478 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Last-Modified: Tue, 15 Aug 2017 08:51:39 
> GMT[\r][\n]"
> 2017-08-15 16:53:16,479 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "ETag: "e7191764798ba504d6671d4c434d2f4d"[\r][\n]"
> 2017-08-15 16:53:16,480 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "x-amz-request-id: 
> tx0001e-005992b67e-27a45-default[\r][\n]"
> 2017-08-15 16:53:16,481 [JUnit-testRandomRe

[jira] [Updated] (HADOOP-14774) S3A case "testRandomReadOverBuffer" failed due to improper range parameter

2017-09-07 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14774?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14774:

Attachment: HADOOP-14774.003.patch

> S3A case "testRandomReadOverBuffer" failed due to improper range parameter
> --
>
> Key: HADOOP-14774
> URL: https://issues.apache.org/jira/browse/HADOOP-14774
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: Hadoop 2.8.0  
> s3-compatible storage 
>Reporter: Yonger
>Assignee: Yonger
>Priority: Minor
> Attachments: failsafe-report.html, HADOOP-14774.001.patch, 
> HADOOP-14774.002.patch, HADOOP-14774.003.patch
>
>
> {code:java}
> Running org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 2.821 sec <<< 
> FAILURE! - in org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> testRandomReadOverBuffer(org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance)
>   Time elapsed: 2.605 sec  <<< FAILURE!
> java.lang.AssertionError: expected:<8192> but was:<8193>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotEquals(Assert.java:743)
>   at org.junit.Assert.assertEquals(Assert.java:118)
>   at org.junit.Assert.assertEquals(Assert.java:555)
>   at org.junit.Assert.assertEquals(Assert.java:542)
>   at 
> org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance.testRandomReadOverBuffer(ITestS3AInputStreamPerformance.java:533)
> {code}
> From log, the length of content is exceed than our expect:
> {code:java}
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(280)) - >> GET 
> /test-aws-s3a/test/testReadOverBuffer.bin HTTP/1.1
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Host: 10.0.2.254
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> 
> x-amz-content-sha256: 
> e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Authorization: 
> AWS4-HMAC-SHA256 
> Credential=JFDAM9KF9IY8S5P0JIV6/20170815/us-east-1/s3/aws4_request, 
> SignedHeaders=content-type;host;range;user-agent;x-amz-content-sha256;x-amz-date,
>  Signature=42bce4a43d2b1bf6e6d599613c60812e6716514da4ef5b3839ef0566c31279ee
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> X-Amz-Date: 
> 20170815T085316Z
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> User-Agent: Hadoop 
> 2.8.0, aws-sdk-java/1.10.6 Linux/3.10.0-514.21.2.el7.x86_64 
> Java_HotSpot(TM)_64-Bit_Server_VM/25.131-b11/1.8.0_131
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Range: bytes=0-8192
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Content-Type: 
> application/x-www-form-urlencoded; charset=utf-8
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Connection: 
> Keep-Alive
> 2017-08-15 16:53:16,473 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "HTTP/1.1 206 Partial Content[\r][\n]"
> 2017-08-15 16:53:16,475 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Range: bytes 0-8192/32768[\r][\n]"
> 2017-08-15 16:53:16,476 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Length: 8193[\r][\n]"
> 2017-08-15 16:53:16,477 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Accept-Ranges: bytes[\r][\n]"
> 2017-08-15 16:53:16,478 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Last-Modified: Tue, 15 Aug 2017 08:51:39 
> GMT[\r][\n]"
> 2017-08-15 16:53:16,479 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "ETag: "e7191764798ba504d6671d4c434d2f4d"[\r][\n]"
> 2017-08-15 16:53:16,480 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "x-amz-request-id: 
> tx0001e-005992b67e-27a45-default[\r][\n]"
> 2017-08-15 16:53:16,481 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Con

[jira] [Updated] (HADOOP-14774) S3A case "testRandomReadOverBuffer" failed due to improper range parameter

2017-09-05 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14774?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14774:

Attachment: HADOOP-14774.002.patch

> S3A case "testRandomReadOverBuffer" failed due to improper range parameter
> --
>
> Key: HADOOP-14774
> URL: https://issues.apache.org/jira/browse/HADOOP-14774
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: Hadoop 2.8.0  
> s3-compatible storage 
>Reporter: Yonger
>Assignee: Yonger
>Priority: Minor
> Attachments: failsafe-report.html, HADOOP-14774.001.patch, 
> HADOOP-14774.002.patch
>
>
> {code:java}
> Running org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 2.821 sec <<< 
> FAILURE! - in org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> testRandomReadOverBuffer(org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance)
>   Time elapsed: 2.605 sec  <<< FAILURE!
> java.lang.AssertionError: expected:<8192> but was:<8193>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotEquals(Assert.java:743)
>   at org.junit.Assert.assertEquals(Assert.java:118)
>   at org.junit.Assert.assertEquals(Assert.java:555)
>   at org.junit.Assert.assertEquals(Assert.java:542)
>   at 
> org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance.testRandomReadOverBuffer(ITestS3AInputStreamPerformance.java:533)
> {code}
> From log, the length of content is exceed than our expect:
> {code:java}
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(280)) - >> GET 
> /test-aws-s3a/test/testReadOverBuffer.bin HTTP/1.1
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Host: 10.0.2.254
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> 
> x-amz-content-sha256: 
> e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Authorization: 
> AWS4-HMAC-SHA256 
> Credential=JFDAM9KF9IY8S5P0JIV6/20170815/us-east-1/s3/aws4_request, 
> SignedHeaders=content-type;host;range;user-agent;x-amz-content-sha256;x-amz-date,
>  Signature=42bce4a43d2b1bf6e6d599613c60812e6716514da4ef5b3839ef0566c31279ee
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> X-Amz-Date: 
> 20170815T085316Z
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> User-Agent: Hadoop 
> 2.8.0, aws-sdk-java/1.10.6 Linux/3.10.0-514.21.2.el7.x86_64 
> Java_HotSpot(TM)_64-Bit_Server_VM/25.131-b11/1.8.0_131
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Range: bytes=0-8192
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Content-Type: 
> application/x-www-form-urlencoded; charset=utf-8
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Connection: 
> Keep-Alive
> 2017-08-15 16:53:16,473 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "HTTP/1.1 206 Partial Content[\r][\n]"
> 2017-08-15 16:53:16,475 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Range: bytes 0-8192/32768[\r][\n]"
> 2017-08-15 16:53:16,476 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Length: 8193[\r][\n]"
> 2017-08-15 16:53:16,477 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Accept-Ranges: bytes[\r][\n]"
> 2017-08-15 16:53:16,478 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Last-Modified: Tue, 15 Aug 2017 08:51:39 
> GMT[\r][\n]"
> 2017-08-15 16:53:16,479 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "ETag: "e7191764798ba504d6671d4c434d2f4d"[\r][\n]"
> 2017-08-15 16:53:16,480 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "x-amz-request-id: 
> tx0001e-005992b67e-27a45-default[\r][\n]"
> 2017-08-15 16:53:16,481 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Type: application/o

[jira] [Updated] (HADOOP-14774) S3A case "testRandomReadOverBuffer" failed due to improper range parameter

2017-09-05 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14774?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14774:

Attachment: failsafe-report.html

test against on Ceph object store over s3a.

> S3A case "testRandomReadOverBuffer" failed due to improper range parameter
> --
>
> Key: HADOOP-14774
> URL: https://issues.apache.org/jira/browse/HADOOP-14774
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: Hadoop 2.8.0  
> s3-compatible storage 
>Reporter: Yonger
>Assignee: Yonger
>Priority: Minor
> Attachments: failsafe-report.html, HADOOP-14774.001.patch
>
>
> {code:java}
> Running org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 2.821 sec <<< 
> FAILURE! - in org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> testRandomReadOverBuffer(org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance)
>   Time elapsed: 2.605 sec  <<< FAILURE!
> java.lang.AssertionError: expected:<8192> but was:<8193>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotEquals(Assert.java:743)
>   at org.junit.Assert.assertEquals(Assert.java:118)
>   at org.junit.Assert.assertEquals(Assert.java:555)
>   at org.junit.Assert.assertEquals(Assert.java:542)
>   at 
> org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance.testRandomReadOverBuffer(ITestS3AInputStreamPerformance.java:533)
> {code}
> From log, the length of content is exceed than our expect:
> {code:java}
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(280)) - >> GET 
> /test-aws-s3a/test/testReadOverBuffer.bin HTTP/1.1
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Host: 10.0.2.254
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> 
> x-amz-content-sha256: 
> e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Authorization: 
> AWS4-HMAC-SHA256 
> Credential=JFDAM9KF9IY8S5P0JIV6/20170815/us-east-1/s3/aws4_request, 
> SignedHeaders=content-type;host;range;user-agent;x-amz-content-sha256;x-amz-date,
>  Signature=42bce4a43d2b1bf6e6d599613c60812e6716514da4ef5b3839ef0566c31279ee
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> X-Amz-Date: 
> 20170815T085316Z
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> User-Agent: Hadoop 
> 2.8.0, aws-sdk-java/1.10.6 Linux/3.10.0-514.21.2.el7.x86_64 
> Java_HotSpot(TM)_64-Bit_Server_VM/25.131-b11/1.8.0_131
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Range: bytes=0-8192
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Content-Type: 
> application/x-www-form-urlencoded; charset=utf-8
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Connection: 
> Keep-Alive
> 2017-08-15 16:53:16,473 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "HTTP/1.1 206 Partial Content[\r][\n]"
> 2017-08-15 16:53:16,475 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Range: bytes 0-8192/32768[\r][\n]"
> 2017-08-15 16:53:16,476 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Length: 8193[\r][\n]"
> 2017-08-15 16:53:16,477 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Accept-Ranges: bytes[\r][\n]"
> 2017-08-15 16:53:16,478 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Last-Modified: Tue, 15 Aug 2017 08:51:39 
> GMT[\r][\n]"
> 2017-08-15 16:53:16,479 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "ETag: "e7191764798ba504d6671d4c434d2f4d"[\r][\n]"
> 2017-08-15 16:53:16,480 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "x-amz-request-id: 
> tx0001e-005992b67e-27a45-default[\r][\n]"
> 2017-08-15 16:53:16,481 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Typ

[jira] [Updated] (HADOOP-14774) S3A case "testRandomReadOverBuffer" failed due to improper range parameter

2017-08-28 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14774?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14774:

Attachment: HADOOP-14774.001.patch

upload patch first since my s3-compatible storage backend not works well, then 
i will upload my test results if the storage come back.

> S3A case "testRandomReadOverBuffer" failed due to improper range parameter
> --
>
> Key: HADOOP-14774
> URL: https://issues.apache.org/jira/browse/HADOOP-14774
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: Hadoop 2.8.0  
> s3-compatible storage 
>Reporter: Yonger
>Assignee: Yonger
>Priority: Minor
> Attachments: HADOOP-14774.001.patch
>
>
> {code:java}
> Running org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 2.821 sec <<< 
> FAILURE! - in org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> testRandomReadOverBuffer(org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance)
>   Time elapsed: 2.605 sec  <<< FAILURE!
> java.lang.AssertionError: expected:<8192> but was:<8193>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotEquals(Assert.java:743)
>   at org.junit.Assert.assertEquals(Assert.java:118)
>   at org.junit.Assert.assertEquals(Assert.java:555)
>   at org.junit.Assert.assertEquals(Assert.java:542)
>   at 
> org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance.testRandomReadOverBuffer(ITestS3AInputStreamPerformance.java:533)
> {code}
> From log, the length of content is exceed than our expect:
> {code:java}
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(280)) - >> GET 
> /test-aws-s3a/test/testReadOverBuffer.bin HTTP/1.1
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Host: 10.0.2.254
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> 
> x-amz-content-sha256: 
> e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Authorization: 
> AWS4-HMAC-SHA256 
> Credential=JFDAM9KF9IY8S5P0JIV6/20170815/us-east-1/s3/aws4_request, 
> SignedHeaders=content-type;host;range;user-agent;x-amz-content-sha256;x-amz-date,
>  Signature=42bce4a43d2b1bf6e6d599613c60812e6716514da4ef5b3839ef0566c31279ee
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> X-Amz-Date: 
> 20170815T085316Z
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> User-Agent: Hadoop 
> 2.8.0, aws-sdk-java/1.10.6 Linux/3.10.0-514.21.2.el7.x86_64 
> Java_HotSpot(TM)_64-Bit_Server_VM/25.131-b11/1.8.0_131
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Range: bytes=0-8192
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Content-Type: 
> application/x-www-form-urlencoded; charset=utf-8
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Connection: 
> Keep-Alive
> 2017-08-15 16:53:16,473 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "HTTP/1.1 206 Partial Content[\r][\n]"
> 2017-08-15 16:53:16,475 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Range: bytes 0-8192/32768[\r][\n]"
> 2017-08-15 16:53:16,476 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Length: 8193[\r][\n]"
> 2017-08-15 16:53:16,477 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Accept-Ranges: bytes[\r][\n]"
> 2017-08-15 16:53:16,478 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Last-Modified: Tue, 15 Aug 2017 08:51:39 
> GMT[\r][\n]"
> 2017-08-15 16:53:16,479 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "ETag: "e7191764798ba504d6671d4c434d2f4d"[\r][\n]"
> 2017-08-15 16:53:16,480 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "x-amz-request-id: 
> tx0001e-005992b67e-27a45-default[\r][\n]"
> 2017-08-15 16:53:16,481 [JUnit-testRandomRe

[jira] [Issue Comment Deleted] (HADOOP-14774) S3A case "testRandomReadOverBuffer" failed due to improper range parameter

2017-08-16 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14774?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14774:

Comment: was deleted

(was: I saw the doc of withRange said :
 * The first byte in an object has
 * position 0; as an example, the first ten bytes of an object can be
 * downloaded by specifying a range of 0 to 9.
So, s3 and s3-compatible store should follow this semantics, doesn't it? and 
this is the issue?)

> S3A case "testRandomReadOverBuffer" failed due to improper range parameter
> --
>
> Key: HADOOP-14774
> URL: https://issues.apache.org/jira/browse/HADOOP-14774
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: Hadoop 2.8.0  
> s3-compatible storage 
>Reporter: Yonger
>Assignee: Yonger
>Priority: Minor
>
> {code:java}
> Running org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 2.821 sec <<< 
> FAILURE! - in org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> testRandomReadOverBuffer(org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance)
>   Time elapsed: 2.605 sec  <<< FAILURE!
> java.lang.AssertionError: expected:<8192> but was:<8193>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotEquals(Assert.java:743)
>   at org.junit.Assert.assertEquals(Assert.java:118)
>   at org.junit.Assert.assertEquals(Assert.java:555)
>   at org.junit.Assert.assertEquals(Assert.java:542)
>   at 
> org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance.testRandomReadOverBuffer(ITestS3AInputStreamPerformance.java:533)
> {code}
> From log, the length of content is exceed than our expect:
> {code:java}
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(280)) - >> GET 
> /test-aws-s3a/test/testReadOverBuffer.bin HTTP/1.1
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Host: 10.0.2.254
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> 
> x-amz-content-sha256: 
> e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Authorization: 
> AWS4-HMAC-SHA256 
> Credential=JFDAM9KF9IY8S5P0JIV6/20170815/us-east-1/s3/aws4_request, 
> SignedHeaders=content-type;host;range;user-agent;x-amz-content-sha256;x-amz-date,
>  Signature=42bce4a43d2b1bf6e6d599613c60812e6716514da4ef5b3839ef0566c31279ee
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> X-Amz-Date: 
> 20170815T085316Z
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> User-Agent: Hadoop 
> 2.8.0, aws-sdk-java/1.10.6 Linux/3.10.0-514.21.2.el7.x86_64 
> Java_HotSpot(TM)_64-Bit_Server_VM/25.131-b11/1.8.0_131
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Range: bytes=0-8192
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Content-Type: 
> application/x-www-form-urlencoded; charset=utf-8
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Connection: 
> Keep-Alive
> 2017-08-15 16:53:16,473 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "HTTP/1.1 206 Partial Content[\r][\n]"
> 2017-08-15 16:53:16,475 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Range: bytes 0-8192/32768[\r][\n]"
> 2017-08-15 16:53:16,476 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Length: 8193[\r][\n]"
> 2017-08-15 16:53:16,477 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Accept-Ranges: bytes[\r][\n]"
> 2017-08-15 16:53:16,478 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Last-Modified: Tue, 15 Aug 2017 08:51:39 
> GMT[\r][\n]"
> 2017-08-15 16:53:16,479 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "ETag: "e7191764798ba504d6671d4c434d2f4d"[\r][\n]"
> 2017-08-15 16:53:16,480 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "x-amz-request-id: 
>

[jira] [Commented] (HADOOP-14774) S3A case "testRandomReadOverBuffer" failed due to improper range parameter

2017-08-16 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14774?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16128838#comment-16128838
 ] 

Yonger commented on HADOOP-14774:
-

I saw the doc of withRange said :
 * The first byte in an object has
 * position 0; as an example, the first ten bytes of an object can be
 * downloaded by specifying a range of 0 to 9.
So, s3 and s3-compatible store should follow this semantics, doesn't it? and 
this is the issue?

> S3A case "testRandomReadOverBuffer" failed due to improper range parameter
> --
>
> Key: HADOOP-14774
> URL: https://issues.apache.org/jira/browse/HADOOP-14774
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: Hadoop 2.8.0  
> s3-compatible storage 
>Reporter: Yonger
>Assignee: Yonger
>Priority: Minor
>
> {code:java}
> Running org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 2.821 sec <<< 
> FAILURE! - in org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> testRandomReadOverBuffer(org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance)
>   Time elapsed: 2.605 sec  <<< FAILURE!
> java.lang.AssertionError: expected:<8192> but was:<8193>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotEquals(Assert.java:743)
>   at org.junit.Assert.assertEquals(Assert.java:118)
>   at org.junit.Assert.assertEquals(Assert.java:555)
>   at org.junit.Assert.assertEquals(Assert.java:542)
>   at 
> org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance.testRandomReadOverBuffer(ITestS3AInputStreamPerformance.java:533)
> {code}
> From log, the length of content is exceed than our expect:
> {code:java}
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(280)) - >> GET 
> /test-aws-s3a/test/testReadOverBuffer.bin HTTP/1.1
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Host: 10.0.2.254
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> 
> x-amz-content-sha256: 
> e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Authorization: 
> AWS4-HMAC-SHA256 
> Credential=JFDAM9KF9IY8S5P0JIV6/20170815/us-east-1/s3/aws4_request, 
> SignedHeaders=content-type;host;range;user-agent;x-amz-content-sha256;x-amz-date,
>  Signature=42bce4a43d2b1bf6e6d599613c60812e6716514da4ef5b3839ef0566c31279ee
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> X-Amz-Date: 
> 20170815T085316Z
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> User-Agent: Hadoop 
> 2.8.0, aws-sdk-java/1.10.6 Linux/3.10.0-514.21.2.el7.x86_64 
> Java_HotSpot(TM)_64-Bit_Server_VM/25.131-b11/1.8.0_131
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Range: bytes=0-8192
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Content-Type: 
> application/x-www-form-urlencoded; charset=utf-8
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Connection: 
> Keep-Alive
> 2017-08-15 16:53:16,473 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "HTTP/1.1 206 Partial Content[\r][\n]"
> 2017-08-15 16:53:16,475 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Range: bytes 0-8192/32768[\r][\n]"
> 2017-08-15 16:53:16,476 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Length: 8193[\r][\n]"
> 2017-08-15 16:53:16,477 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Accept-Ranges: bytes[\r][\n]"
> 2017-08-15 16:53:16,478 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Last-Modified: Tue, 15 Aug 2017 08:51:39 
> GMT[\r][\n]"
> 2017-08-15 16:53:16,479 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "ETag: "e7191764798ba504d6671d4c434d2f4d"[\r][\n]"
> 2017-08-15 16:53:16,480 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) - 

[jira] [Commented] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-08-16 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16128548#comment-16128548
 ] 

Yonger commented on HADOOP-14475:
-

[~ste...@apache.org] Can you help to verify and review this patch? 

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, failsafe-report-scale.html, 
> failsafe-report-scale.zip, HADOOP-14475.002.patch, HADOOP-14475-003.patch, 
> HADOOP-14475.005.patch, HADOOP-14475.006.patch, s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14770) S3A http connection in s3a driver not reuse in Spark application

2017-08-15 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14770?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16127218#comment-16127218
 ] 

Yonger commented on HADOOP-14770:
-

Sorry, not yet. I am working with multiple partners on our big data cluster, so 
it's not easy to move to 2.8.  But I will complete it ASAP.

> S3A http connection in s3a driver not reuse in Spark application
> 
>
> Key: HADOOP-14770
> URL: https://issues.apache.org/jira/browse/HADOOP-14770
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.7.3
>Reporter: Yonger
>Assignee: Yonger
>Priority: Minor
>
> I print out connection stats every 2 s when running Spark application against 
> s3-compatible storage:
> {code}
> ESTAB  0  0 :::10.0.2.36:6
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44454
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44374
> :::10.0.2.254:80 
> ESTAB  159724 0 :::10.0.2.36:44436
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:8
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44338
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44438
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44414
> :::10.0.2.254:80 
> ESTAB  0  480   :::10.0.2.36:44450
> :::10.0.2.254:80  timer:(on,170ms,0)
> ESTAB  0  0 :::10.0.2.36:2
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44390
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44326
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44452
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44394
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:4
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44456
> :::10.0.2.254:80 
> ==
> ESTAB  0  0 :::10.0.2.36:44508
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44476
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44524
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44374
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44500
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44504
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44512
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44506
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44464
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44518
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44510
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:2
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44526
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44472
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44466
> :::10.0.2.254:80 
> {code}
> the connection in the above of "=" and below were changed all the time. But 
> this haven't seen in MR application. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14774) S3A case "testRandomReadOverBuffer" failed due to improper range parameter

2017-08-15 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14774?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16126990#comment-16126990
 ] 

Yonger commented on HADOOP-14774:
-


{code:java}
  GetObjectRequest request = new GetObjectRequest(bucket, key)
  .withRange(targetPos, contentRangeFinish);
{code}
we should pass contentRangeFinish-1 instead of contentRangeFinish into 
withRange method

> S3A case "testRandomReadOverBuffer" failed due to improper range parameter
> --
>
> Key: HADOOP-14774
> URL: https://issues.apache.org/jira/browse/HADOOP-14774
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: Hadoop 2.8.0  
> s3-compatible storage 
>Reporter: Yonger
>Assignee: Yonger
>
> {code:java}
> Running org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 2.821 sec <<< 
> FAILURE! - in org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> testRandomReadOverBuffer(org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance)
>   Time elapsed: 2.605 sec  <<< FAILURE!
> java.lang.AssertionError: expected:<8192> but was:<8193>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotEquals(Assert.java:743)
>   at org.junit.Assert.assertEquals(Assert.java:118)
>   at org.junit.Assert.assertEquals(Assert.java:555)
>   at org.junit.Assert.assertEquals(Assert.java:542)
>   at 
> org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance.testRandomReadOverBuffer(ITestS3AInputStreamPerformance.java:533)
> {code}
> From log, the length of content is exceed than our expect:
> {code:java}
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(280)) - >> GET 
> /test-aws-s3a/test/testReadOverBuffer.bin HTTP/1.1
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Host: 10.0.2.254
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> 
> x-amz-content-sha256: 
> e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Authorization: 
> AWS4-HMAC-SHA256 
> Credential=JFDAM9KF9IY8S5P0JIV6/20170815/us-east-1/s3/aws4_request, 
> SignedHeaders=content-type;host;range;user-agent;x-amz-content-sha256;x-amz-date,
>  Signature=42bce4a43d2b1bf6e6d599613c60812e6716514da4ef5b3839ef0566c31279ee
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> X-Amz-Date: 
> 20170815T085316Z
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> User-Agent: Hadoop 
> 2.8.0, aws-sdk-java/1.10.6 Linux/3.10.0-514.21.2.el7.x86_64 
> Java_HotSpot(TM)_64-Bit_Server_VM/25.131-b11/1.8.0_131
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Range: bytes=0-8192
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Content-Type: 
> application/x-www-form-urlencoded; charset=utf-8
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Connection: 
> Keep-Alive
> 2017-08-15 16:53:16,473 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "HTTP/1.1 206 Partial Content[\r][\n]"
> 2017-08-15 16:53:16,475 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Range: bytes 0-8192/32768[\r][\n]"
> 2017-08-15 16:53:16,476 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Length: 8193[\r][\n]"
> 2017-08-15 16:53:16,477 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Accept-Ranges: bytes[\r][\n]"
> 2017-08-15 16:53:16,478 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Last-Modified: Tue, 15 Aug 2017 08:51:39 
> GMT[\r][\n]"
> 2017-08-15 16:53:16,479 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "ETag: "e7191764798ba504d6671d4c434d2f4d"[\r][\n]"
> 2017-08-15 16:53:16,480 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "x-amz-request-id: 
> tx0001e-005992b67e-27a45-default[\r][\n]"
> 2017-08-15 16:

[jira] [Updated] (HADOOP-14774) S3A case "testRandomReadOverBuffer" failed due to improper range parameter

2017-08-15 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14774?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14774:

Description: 

{code:java}
Running org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 2.821 sec <<< 
FAILURE! - in org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
testRandomReadOverBuffer(org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance)
  Time elapsed: 2.605 sec  <<< FAILURE!
java.lang.AssertionError: expected:<8192> but was:<8193>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:743)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:555)
at org.junit.Assert.assertEquals(Assert.java:542)
at 
org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance.testRandomReadOverBuffer(ITestS3AInputStreamPerformance.java:533)
{code}


>From log, the length of content is exceed than our expect:

{code:java}
2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(280)) - >> GET 
/test-aws-s3a/test/testReadOverBuffer.bin HTTP/1.1
2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(283)) - >> Host: 10.0.2.254
2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(283)) - >> 
x-amz-content-sha256: 
e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855
2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(283)) - >> Authorization: 
AWS4-HMAC-SHA256 
Credential=JFDAM9KF9IY8S5P0JIV6/20170815/us-east-1/s3/aws4_request, 
SignedHeaders=content-type;host;range;user-agent;x-amz-content-sha256;x-amz-date,
 Signature=42bce4a43d2b1bf6e6d599613c60812e6716514da4ef5b3839ef0566c31279ee
2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(283)) - >> X-Amz-Date: 
20170815T085316Z
2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(283)) - >> User-Agent: Hadoop 
2.8.0, aws-sdk-java/1.10.6 Linux/3.10.0-514.21.2.el7.x86_64 
Java_HotSpot(TM)_64-Bit_Server_VM/25.131-b11/1.8.0_131
2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(283)) - >> Range: bytes=0-8192
2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(283)) - >> Content-Type: 
application/x-www-form-urlencoded; charset=utf-8
2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(283)) - >> Connection: 
Keep-Alive
2017-08-15 16:53:16,473 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "HTTP/1.1 206 Partial Content[\r][\n]"
2017-08-15 16:53:16,475 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "Content-Range: bytes 0-8192/32768[\r][\n]"
2017-08-15 16:53:16,476 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "Content-Length: 8193[\r][\n]"
2017-08-15 16:53:16,477 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "Accept-Ranges: bytes[\r][\n]"
2017-08-15 16:53:16,478 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "Last-Modified: Tue, 15 Aug 2017 08:51:39 
GMT[\r][\n]"
2017-08-15 16:53:16,479 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "ETag: "e7191764798ba504d6671d4c434d2f4d"[\r][\n]"
2017-08-15 16:53:16,480 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "x-amz-request-id: 
tx0001e-005992b67e-27a45-default[\r][\n]"
2017-08-15 16:53:16,481 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "Content-Type: application/octet-stream[\r][\n]"
2017-08-15 16:53:16,482 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "Date: Tue, 15 Aug 2017 08:53:18 GMT[\r][\n]"
2017-08-15 16:53:16,483 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "[\r][\n]"
{code}

 

  was:
Running org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 2.821 sec <<< 
FAILURE! - in org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
testRandomReadOverBuffer(org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance)
  Time elapsed: 2.605 sec  <<< FAILURE!
java.lang.AssertionError: expected:<8192> but was:<8193>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.A

[jira] [Created] (HADOOP-14774) S3A case "testRandomReadOverBuffer" failed due to improper range parameter

2017-08-15 Thread Yonger (JIRA)
Yonger created HADOOP-14774:
---

 Summary: S3A case "testRandomReadOverBuffer" failed due to 
improper range parameter
 Key: HADOOP-14774
 URL: https://issues.apache.org/jira/browse/HADOOP-14774
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs/s3
Affects Versions: 2.8.0
 Environment: Hadoop 2.8.0  
s3-compatible storage 
Reporter: Yonger


Running org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 2.821 sec <<< 
FAILURE! - in org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
testRandomReadOverBuffer(org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance)
  Time elapsed: 2.605 sec  <<< FAILURE!
java.lang.AssertionError: expected:<8192> but was:<8193>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:743)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:555)
at org.junit.Assert.assertEquals(Assert.java:542)
at 
org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance.testRandomReadOverBuffer(ITestS3AInputStreamPerformance.java:533)

>From log, the length of content is exceed than our expect:
2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(280)) - >> GET 
/test-aws-s3a/test/testReadOverBuffer.bin HTTP/1.1
2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(283)) - >> Host: 10.0.2.254
2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(283)) - >> 
x-amz-content-sha256: 
e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855
2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(283)) - >> Authorization: 
AWS4-HMAC-SHA256 
Credential=JFDAM9KF9IY8S5P0JIV6/20170815/us-east-1/s3/aws4_request, 
SignedHeaders=content-type;host;range;user-agent;x-amz-content-sha256;x-amz-date,
 Signature=42bce4a43d2b1bf6e6d599613c60812e6716514da4ef5b3839ef0566c31279ee
2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(283)) - >> X-Amz-Date: 
20170815T085316Z
2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(283)) - >> User-Agent: Hadoop 
2.8.0, aws-sdk-java/1.10.6 Linux/3.10.0-514.21.2.el7.x86_64 
Java_HotSpot(TM)_64-Bit_Server_VM/25.131-b11/1.8.0_131
2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(283)) - >> Range: bytes=0-8192
2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(283)) - >> Content-Type: 
application/x-www-form-urlencoded; charset=utf-8
2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
(DefaultClientConnection.java:sendRequestHeader(283)) - >> Connection: 
Keep-Alive
2017-08-15 16:53:16,473 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "HTTP/1.1 206 Partial Content[\r][\n]"
2017-08-15 16:53:16,475 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "Content-Range: bytes 0-8192/32768[\r][\n]"
2017-08-15 16:53:16,476 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "Content-Length: 8193[\r][\n]"
2017-08-15 16:53:16,477 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "Accept-Ranges: bytes[\r][\n]"
2017-08-15 16:53:16,478 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "Last-Modified: Tue, 15 Aug 2017 08:51:39 
GMT[\r][\n]"
2017-08-15 16:53:16,479 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "ETag: "e7191764798ba504d6671d4c434d2f4d"[\r][\n]"
2017-08-15 16:53:16,480 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "x-amz-request-id: 
tx0001e-005992b67e-27a45-default[\r][\n]"
2017-08-15 16:53:16,481 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "Content-Type: application/octet-stream[\r][\n]"
2017-08-15 16:53:16,482 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "Date: Tue, 15 Aug 2017 08:53:18 GMT[\r][\n]"
2017-08-15 16:53:16,483 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
(Wire.java:wire(72)) -  << "[\r][\n]"
 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.

[jira] [Assigned] (HADOOP-14774) S3A case "testRandomReadOverBuffer" failed due to improper range parameter

2017-08-15 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14774?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger reassigned HADOOP-14774:
---

Assignee: Yonger

> S3A case "testRandomReadOverBuffer" failed due to improper range parameter
> --
>
> Key: HADOOP-14774
> URL: https://issues.apache.org/jira/browse/HADOOP-14774
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: Hadoop 2.8.0  
> s3-compatible storage 
>Reporter: Yonger
>Assignee: Yonger
>
> Running org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 2.821 sec <<< 
> FAILURE! - in org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance
> testRandomReadOverBuffer(org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance)
>   Time elapsed: 2.605 sec  <<< FAILURE!
> java.lang.AssertionError: expected:<8192> but was:<8193>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotEquals(Assert.java:743)
>   at org.junit.Assert.assertEquals(Assert.java:118)
>   at org.junit.Assert.assertEquals(Assert.java:555)
>   at org.junit.Assert.assertEquals(Assert.java:542)
>   at 
> org.apache.hadoop.fs.s3a.scale.ITestS3AInputStreamPerformance.testRandomReadOverBuffer(ITestS3AInputStreamPerformance.java:533)
> From log, the length of content is exceed than our expect:
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(280)) - >> GET 
> /test-aws-s3a/test/testReadOverBuffer.bin HTTP/1.1
> 2017-08-15 16:53:16,464 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Host: 10.0.2.254
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> 
> x-amz-content-sha256: 
> e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855
> 2017-08-15 16:53:16,465 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Authorization: 
> AWS4-HMAC-SHA256 
> Credential=JFDAM9KF9IY8S5P0JIV6/20170815/us-east-1/s3/aws4_request, 
> SignedHeaders=content-type;host;range;user-agent;x-amz-content-sha256;x-amz-date,
>  Signature=42bce4a43d2b1bf6e6d599613c60812e6716514da4ef5b3839ef0566c31279ee
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> X-Amz-Date: 
> 20170815T085316Z
> 2017-08-15 16:53:16,466 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> User-Agent: Hadoop 
> 2.8.0, aws-sdk-java/1.10.6 Linux/3.10.0-514.21.2.el7.x86_64 
> Java_HotSpot(TM)_64-Bit_Server_VM/25.131-b11/1.8.0_131
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Range: bytes=0-8192
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Content-Type: 
> application/x-www-form-urlencoded; charset=utf-8
> 2017-08-15 16:53:16,467 [JUnit-testRandomReadOverBuffer] DEBUG http.headers 
> (DefaultClientConnection.java:sendRequestHeader(283)) - >> Connection: 
> Keep-Alive
> 2017-08-15 16:53:16,473 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "HTTP/1.1 206 Partial Content[\r][\n]"
> 2017-08-15 16:53:16,475 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Range: bytes 0-8192/32768[\r][\n]"
> 2017-08-15 16:53:16,476 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Length: 8193[\r][\n]"
> 2017-08-15 16:53:16,477 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Accept-Ranges: bytes[\r][\n]"
> 2017-08-15 16:53:16,478 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Last-Modified: Tue, 15 Aug 2017 08:51:39 
> GMT[\r][\n]"
> 2017-08-15 16:53:16,479 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "ETag: "e7191764798ba504d6671d4c434d2f4d"[\r][\n]"
> 2017-08-15 16:53:16,480 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "x-amz-request-id: 
> tx0001e-005992b67e-27a45-default[\r][\n]"
> 2017-08-15 16:53:16,481 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Content-Type: application/octet-stream[\r][\n]"
> 2017-08-15 16:53:16,482 [JUnit-testRandomReadOverBuffer] DEBUG http.wire 
> (Wire.java:wire(72)) -  << "Date: Tue, 15 Aug 2017 08:53:18 GMT[\r][\n]"
> 2017-

[jira] [Commented] (HADOOP-14770) S3A http connection in s3a driver not reuse in Spark application

2017-08-14 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14770?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16125536#comment-16125536
 ] 

Yonger commented on HADOOP-14770:
-

Thanks Steve, the application running on Hadoop 2.7.3 and against ORC file 
format. I will upgrade to Hadoop 2.8.0 to verify.

> S3A http connection in s3a driver not reuse in Spark application
> 
>
> Key: HADOOP-14770
> URL: https://issues.apache.org/jira/browse/HADOOP-14770
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.7.3
>Reporter: Yonger
>Assignee: Yonger
>
> I print out connection stats every 2 s when running Spark application against 
> s3-compatible storage:
> {code}
> ESTAB  0  0 :::10.0.2.36:6
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44454
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44374
> :::10.0.2.254:80 
> ESTAB  159724 0 :::10.0.2.36:44436
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:8
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44338
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44438
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44414
> :::10.0.2.254:80 
> ESTAB  0  480   :::10.0.2.36:44450
> :::10.0.2.254:80  timer:(on,170ms,0)
> ESTAB  0  0 :::10.0.2.36:2
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44390
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44326
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44452
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44394
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:4
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44456
> :::10.0.2.254:80 
> ==
> ESTAB  0  0 :::10.0.2.36:44508
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44476
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44524
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44374
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44500
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44504
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44512
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44506
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44464
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44518
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44510
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:2
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44526
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44472
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44466
> :::10.0.2.254:80 
> {code}
> the connection in the above of "=" and below were changed all the time. But 
> this haven't seen in MR application. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14770) S3A http connection in s3a driver not reuse in Spark application

2017-08-14 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14770?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14770:

Affects Version/s: 2.7.3

> S3A http connection in s3a driver not reuse in Spark application
> 
>
> Key: HADOOP-14770
> URL: https://issues.apache.org/jira/browse/HADOOP-14770
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.7.3
>Reporter: Yonger
>Assignee: Yonger
>
> I print out connection stats every 2 s when running Spark application against 
> s3-compatible storage:
> {code}
> ESTAB  0  0 :::10.0.2.36:6
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44454
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44374
> :::10.0.2.254:80 
> ESTAB  159724 0 :::10.0.2.36:44436
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:8
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44338
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44438
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44414
> :::10.0.2.254:80 
> ESTAB  0  480   :::10.0.2.36:44450
> :::10.0.2.254:80  timer:(on,170ms,0)
> ESTAB  0  0 :::10.0.2.36:2
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44390
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44326
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44452
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44394
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:4
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44456
> :::10.0.2.254:80 
> ==
> ESTAB  0  0 :::10.0.2.36:44508
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44476
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44524
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44374
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44500
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44504
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44512
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44506
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44464
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44518
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44510
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:2
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44526
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44472
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44466
> :::10.0.2.254:80 
> {code}
> the connection in the above of "=" and below were changed all the time. But 
> this haven't seen in MR application. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14770) S3A http connection in s3a driver not reuse in Spark application

2017-08-14 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14770?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14770:

Component/s: fs/s3

> S3A http connection in s3a driver not reuse in Spark application
> 
>
> Key: HADOOP-14770
> URL: https://issues.apache.org/jira/browse/HADOOP-14770
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.7.3
>Reporter: Yonger
>Assignee: Yonger
>
> I print out connection stats every 2 s when running Spark application against 
> s3-compatible storage:
> {code}
> ESTAB  0  0 :::10.0.2.36:6
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44454
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44374
> :::10.0.2.254:80 
> ESTAB  159724 0 :::10.0.2.36:44436
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:8
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44338
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44438
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44414
> :::10.0.2.254:80 
> ESTAB  0  480   :::10.0.2.36:44450
> :::10.0.2.254:80  timer:(on,170ms,0)
> ESTAB  0  0 :::10.0.2.36:2
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44390
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44326
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44452
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44394
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:4
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44456
> :::10.0.2.254:80 
> ==
> ESTAB  0  0 :::10.0.2.36:44508
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44476
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44524
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44374
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44500
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44504
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44512
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44506
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44464
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44518
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44510
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:2
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44526
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44472
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44466
> :::10.0.2.254:80 
> {code}
> the connection in the above of "=" and below were changed all the time. But 
> this haven't seen in MR application. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Assigned] (HADOOP-14770) S3A http connection in s3a driver not reuse in Spark application

2017-08-13 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14770?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger reassigned HADOOP-14770:
---

Assignee: Yonger

> S3A http connection in s3a driver not reuse in Spark application
> 
>
> Key: HADOOP-14770
> URL: https://issues.apache.org/jira/browse/HADOOP-14770
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Yonger
>Assignee: Yonger
>
> I print out connection stats every 2 s when running Spark application against 
> s3-compatible storage:
> ESTAB  0  0 :::10.0.2.36:6
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44454
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44374
> :::10.0.2.254:80 
> ESTAB  159724 0 :::10.0.2.36:44436
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:8
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44338
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44438
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44414
> :::10.0.2.254:80 
> ESTAB  0  480   :::10.0.2.36:44450
> :::10.0.2.254:80  timer:(on,170ms,0)
> ESTAB  0  0 :::10.0.2.36:2
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44390
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44326
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44452
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44394
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:4
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44456
> :::10.0.2.254:80 
> ==
> ESTAB  0  0 :::10.0.2.36:44508
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44476
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44524
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44374
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44500
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44504
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44512
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44506
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44464
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44518
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44510
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:2
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44526
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44472
> :::10.0.2.254:80 
> ESTAB  0  0 :::10.0.2.36:44466
> :::10.0.2.254:80 
> the connection in the above of "=" and below were changed all the time. But 
> this haven't seen in MR application. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-14770) S3A http connection in s3a driver not reuse in Spark application

2017-08-13 Thread Yonger (JIRA)
Yonger created HADOOP-14770:
---

 Summary: S3A http connection in s3a driver not reuse in Spark 
application
 Key: HADOOP-14770
 URL: https://issues.apache.org/jira/browse/HADOOP-14770
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Yonger


I print out connection stats every 2 s when running Spark application against 
s3-compatible storage:

ESTAB  0  0 :::10.0.2.36:6
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44454
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44374
:::10.0.2.254:80 
ESTAB  159724 0 :::10.0.2.36:44436
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:8
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44338
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44438
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44414
:::10.0.2.254:80 
ESTAB  0  480   :::10.0.2.36:44450
:::10.0.2.254:80  timer:(on,170ms,0)
ESTAB  0  0 :::10.0.2.36:2
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44390
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44326
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44452
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44394
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:4
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44456
:::10.0.2.254:80 
==
ESTAB  0  0 :::10.0.2.36:44508
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44476
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44524
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44374
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44500
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44504
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44512
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44506
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44464
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44518
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44510
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:2
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44526
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44472
:::10.0.2.254:80 
ESTAB  0  0 :::10.0.2.36:44466
:::10.0.2.254:80 

the connection in the above of "=" and below were changed all the time. But 
this haven't seen in MR application. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14745) s3a getFileStatus can't return expect result when existing a file and directory with the same name

2017-08-08 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16118177#comment-16118177
 ] 

Yonger commented on HADOOP-14745:
-

I create that path structure through s3cmd, and yes , we can't do this by s3a 
itself. 
But, s3 or s3-compatible storage allow them existed within the same folder, and 
the common workload running on Hadoop on Ceph is data analysis, which means the 
data in these storage should be store by other way instead of s3a, and read 
data from storage through s3a, which can't stop this issue happen again.  

> s3a getFileStatus can't return expect result when existing a file and 
> directory with the same name
> --
>
> Key: HADOOP-14745
> URL: https://issues.apache.org/jira/browse/HADOOP-14745
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
>Reporter: Yonger
>Assignee: Yonger
>
> {code}
> [ hadoop-aws]# /root/hadoop/s3cmd/s3cmd ls s3://test-aws-s3a/user/root/
>DIR   s3://test-aws-s3a/user/root/ccc/
> 2017-08-08 07:04 0   s3://test-aws-s3a/user/root/ccc
> {code}
> if we expect to ccc is a directory by code :
> {code}
> Path test=new Path("ccc");
> fs.getFileStatus(test);
> {code}
> actually, it will tell us it is a file:
> {code}
> 2017-08-08 15:08:40,566 [JUnit-case1] DEBUG s3a.S3AFileSystem 
> (S3AFileSystem.java:getFileStatus(1576)) - Getting path status for 
> s3a://test-aws-s3a/user/root/ccc  (user/root/ccc)
> 2017-08-08 15:08:40,566 [JUnit-case1] DEBUG s3a.S3AFileSystem 
> (S3AStorageStatistics.java:incrementCounter(60)) - object_metadata_requests 
> += 1  ->  3
> 2017-08-08 15:08:40,580 [JUnit-case1] DEBUG s3a.S3AFileSystem 
> (S3AFileSystem.java:getFileStatus(1585)) - Found exact file: normal file
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14745) s3a getFileStatus can't return expect result when existing a file and directory with the same name

2017-08-08 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16118026#comment-16118026
 ] 

Yonger commented on HADOOP-14745:
-

On the other hand, I saw hdfs implementation not allow existing any file with 
the same name of directory under it's parent path.
 

> s3a getFileStatus can't return expect result when existing a file and 
> directory with the same name
> --
>
> Key: HADOOP-14745
> URL: https://issues.apache.org/jira/browse/HADOOP-14745
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
>Reporter: Yonger
>Assignee: Yonger
>
> [ hadoop-aws]# /root/hadoop/s3cmd/s3cmd ls s3://test-aws-s3a/user/root/
>DIR   s3://test-aws-s3a/user/root/ccc/
> 2017-08-08 07:04 0   s3://test-aws-s3a/user/root/ccc
> if we expect to ccc is a directory by code :
> Path test=new Path("ccc");
> fs.getFileStatus(test);
> actually, it will tell us it is a file:
> 2017-08-08 15:08:40,566 [JUnit-case1] DEBUG s3a.S3AFileSystem 
> (S3AFileSystem.java:getFileStatus(1576)) - Getting path status for 
> s3a://test-aws-s3a/user/root/ccc  (user/root/ccc)
> 2017-08-08 15:08:40,566 [JUnit-case1] DEBUG s3a.S3AFileSystem 
> (S3AStorageStatistics.java:incrementCounter(60)) - object_metadata_requests 
> += 1  ->  3
> 2017-08-08 15:08:40,580 [JUnit-case1] DEBUG s3a.S3AFileSystem 
> (S3AFileSystem.java:getFileStatus(1585)) - Found exact file: normal file



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14745) s3a getFileStatus can't return expect result when existing a file and directory with the same name

2017-08-08 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16118009#comment-16118009
 ] 

Yonger commented on HADOOP-14745:
-

So i think we should call getFileStatus with explicitly expect,  e.g.
getFileStatus(path,true)
true means we think the path we input is a directory

In internal of getFileStatus, i think the call like above will skip the first 
two getObjectMetadata call, and only list the object under this given path, 
which also benefit the performance of this network consuming function. 



> s3a getFileStatus can't return expect result when existing a file and 
> directory with the same name
> --
>
> Key: HADOOP-14745
> URL: https://issues.apache.org/jira/browse/HADOOP-14745
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
>Reporter: Yonger
>Assignee: Yonger
>
> [ hadoop-aws]# /root/hadoop/s3cmd/s3cmd ls s3://test-aws-s3a/user/root/
>DIR   s3://test-aws-s3a/user/root/ccc/
> 2017-08-08 07:04 0   s3://test-aws-s3a/user/root/ccc
> if we expect to ccc is a directory by code :
> Path test=new Path("ccc");
> fs.getFileStatus(test);
> actually, it will tell us it is a file:
> 2017-08-08 15:08:40,566 [JUnit-case1] DEBUG s3a.S3AFileSystem 
> (S3AFileSystem.java:getFileStatus(1576)) - Getting path status for 
> s3a://test-aws-s3a/user/root/ccc  (user/root/ccc)
> 2017-08-08 15:08:40,566 [JUnit-case1] DEBUG s3a.S3AFileSystem 
> (S3AStorageStatistics.java:incrementCounter(60)) - object_metadata_requests 
> += 1  ->  3
> 2017-08-08 15:08:40,580 [JUnit-case1] DEBUG s3a.S3AFileSystem 
> (S3AFileSystem.java:getFileStatus(1585)) - Found exact file: normal file



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-14745) s3a getFileStatus can't return expect result when existing a file and directory with the same name

2017-08-08 Thread Yonger (JIRA)
Yonger created HADOOP-14745:
---

 Summary: s3a getFileStatus can't return expect result when 
existing a file and directory with the same name
 Key: HADOOP-14745
 URL: https://issues.apache.org/jira/browse/HADOOP-14745
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs/s3
Affects Versions: 2.8.0
Reporter: Yonger
Assignee: Yonger


[ hadoop-aws]# /root/hadoop/s3cmd/s3cmd ls s3://test-aws-s3a/user/root/
   DIR   s3://test-aws-s3a/user/root/ccc/
2017-08-08 07:04 0   s3://test-aws-s3a/user/root/ccc

if we expect to ccc is a directory by code :
Path test=new Path("ccc");
fs.getFileStatus(test);

actually, it will tell us it is a file:

2017-08-08 15:08:40,566 [JUnit-case1] DEBUG s3a.S3AFileSystem 
(S3AFileSystem.java:getFileStatus(1576)) - Getting path status for 
s3a://test-aws-s3a/user/root/ccc  (user/root/ccc)
2017-08-08 15:08:40,566 [JUnit-case1] DEBUG s3a.S3AFileSystem 
(S3AStorageStatistics.java:incrementCounter(60)) - object_metadata_requests += 
1  ->  3
2017-08-08 15:08:40,580 [JUnit-case1] DEBUG s3a.S3AFileSystem 
(S3AFileSystem.java:getFileStatus(1585)) - Found exact file: normal file





--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13430) Optimize and fix getFileStatus in S3A

2017-08-07 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13430?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16116700#comment-16116700
 ] 

Yonger commented on HADOOP-13430:
-

[~ste...@apache.org] I really can't see any optimization for getFileStatus in 
ticket [https://issues.apache.org/jira/browse/HADOOP-13208] , with my 
understand, there is still exist 2 head +1 list operations in this function. 

And I scan the code where call getFileStatus, and found that in some place, we 
can know the path that input is a file or a directory, e.g. in create and open 
function, we don't need to check the path is a directory with calling 
getfileStatus, just consider the path is a file(according the implementation of 
Presto), thus when calling getFileStatus, we know the input path is a file so 
that it don't need to call getmetadata method again with suffix "/".

Totally, can we reduce the s3 call over network as possible by tell 
getFileStatus the path is a file or directory explicitly? 

> Optimize and fix getFileStatus in S3A
> -
>
> Key: HADOOP-13430
> URL: https://issues.apache.org/jira/browse/HADOOP-13430
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0
>Reporter: Steven K. Wong
>Assignee: Steven K. Wong
>Priority: Minor
> Attachments: HADOOP-13430.001.WIP.patch
>
>
> Currently, S3AFileSystem.getFileStatus(Path f) sends up to 3 requests to S3 
> when pathToKey(f) = key = "foo/bar" is a directory:
> 1. HEAD key=foo/bar \[continue if not found]
> 2. HEAD key=foo/bar/ \[continue if not found]
> 3. LIST prefix=foo/bar/ delimiter=/ max-keys=1
> My experience (and generally true, I reckon) is that almost all directories 
> are nonempty directories without a "fake directory" file (e.g. "foo/bar/"). 
> Under this condition, request #2 is mostly unhelpful; it only slows down 
> getFileStatus. Therefore, I propose swapping the order of requests #2 and #3. 
> The swapped HEAD request will be skipped in practically all cases.
> Furthermore, when key = "foo/bar" is a nonempty directory that contains a 
> "fake directory" file (in addition to actual files), getFileStatus currently 
> returns an S3AFileStatus with isEmptyDirectory=true, which is wrong. Swapping 
> will fix this. The swapped LIST request will use max-keys=2 to determine 
> isEmptyDirectory correctly. (Removing the delimiter from the LIST request 
> should make the logic a little simpler than otherwise.)
> Note that key = "foo/bar/" has the same problem with isEmptyDirectory. To fix 
> it, I propose skipping request #1 when key ends with "/". The price is this 
> will, for an empty directory, replace a HEAD request with a LIST request 
> that's generally more taxing on S3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-08-02 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Attachment: HADOOP-14475.006.patch

update:
1.add ASF license for new test class 

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, failsafe-report-scale.html, 
> failsafe-report-scale.zip, HADOOP-14475.002.patch, HADOOP-14475-003.patch, 
> HADOOP-14475.005.patch, HADOOP-14475.006.patch, s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-08-02 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Attachment: HADOOP-14475.005.patch

update:
1.remove streamMetrics hashmap, register all metrics into registry
2.add missing metric STREAM_WRITE_QUEUE_DURATION
3.add test case in ITestS3AMetrics to verify metrics resource registration 

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, failsafe-report-scale.html, 
> failsafe-report-scale.zip, HADOOP-14475.002.patch, HADOOP-14475-003.patch, 
> HADOOP-14475.005.patch, s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-07-25 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16100142#comment-16100142
 ] 

Yonger commented on HADOOP-14475:
-

[~ste...@apache.org] I found the output of metrics missing part of metrics in 
streamMetrics HashMap, because they didn't register into MetricsRegistry.
With my understand, i will remove streamMetrics HashMap, and register stream 
related metrics directly like method "counter". 
But i note there is function "dump" will read from this HashMap, and few other 
class call this dump method to print stream metrics.  I don't think it's a good 
idea to print any metrics to log file, we should put all performance metrics 
into other place(e.g. a separate file or database) .
Any suggestion?

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, failsafe-report-scale.html, 
> failsafe-report-scale.zip, HADOOP-14475.002.patch, HADOOP-14475-003.patch, 
> s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13845) s3a to instrument duration of HTTP calls

2017-07-24 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13845?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16098416#comment-16098416
 ] 

Yonger commented on HADOOP-13845:
-

Does this make sense?

{code:java}
Duration duration = new Duration();
ObjectListing listing= s3.listObjects(request);
duration.finished();
durationStats.add(method.getName()+" " + reason, duration, success);

return listing;
{code}


> s3a to instrument duration of HTTP calls
> 
>
> Key: HADOOP-13845
> URL: https://issues.apache.org/jira/browse/HADOOP-13845
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.9.0
>Reporter: Steve Loughran
>Priority: Minor
>
> HADOOP-13844 proposes pulling out the swift duration classes for reuse; this 
> patch proposes instrumenting s3a with it.
> One interesting question: what to do with the values. For now, they could 
> just be printed, but it might be interesting to include in FS stats collected 
> at the end of a run. However, those are all assumed to be simple counters 
> where merging is a matter of addition. These are more metrics



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13845) s3a to instrument duration of HTTP calls

2017-07-20 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13845?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16095706#comment-16095706
 ] 

Yonger commented on HADOOP-13845:
-

Another usage is push http duration into database(such as influxdb), and get a 
realtime show via grafana. that you can know the duration of one kind(path) of 
request when is longest when suffering some bottleneck, and then you can do 
trouble shooting work in the range of higher duration time. which can not be 
done if you just only output the final result in FS stats.

And what's going on for HADOOP-13844? i expect to put http duration into s3a 
metrics. 

> s3a to instrument duration of HTTP calls
> 
>
> Key: HADOOP-13845
> URL: https://issues.apache.org/jira/browse/HADOOP-13845
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.9.0
>Reporter: Steve Loughran
>Priority: Minor
>
> HADOOP-13844 proposes pulling out the swift duration classes for reuse; this 
> patch proposes instrumenting s3a with it.
> One interesting question: what to do with the values. For now, they could 
> just be printed, but it might be interesting to include in FS stats collected 
> at the end of a run. However, those are all assumed to be simple counters 
> where merging is a matter of addition. These are more metrics



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13762) S3A: Set thread names with more specific information about the call.

2017-07-20 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13762?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16095684#comment-16095684
 ] 

Yonger commented on HADOOP-13762:
-

I agree a thread name with meaningful information is a great idea, but I also 
think logging the information you mentioned(userid, timestamp, and path)  is 
better than to insert into thread name from a comprehensive perspective. You 
just need to put information into log component and without additional 
complexity and performance consideration. And with these meaningful information 
in logs, we also start a trouble shooting for performance problem, also this 
approach applied in big company with high throughput computer system.   

> S3A: Set thread names with more specific information about the call.
> 
>
> Key: HADOOP-13762
> URL: https://issues.apache.org/jira/browse/HADOOP-13762
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Reporter: Chris Nauroth
>
> Running {{jstack}} on a hung process and reading the stack traces is a 
> helpful way to determine exactly what code in the process is stuck.  This 
> would be even more helpful if we included more descriptive information about 
> the specific file system method call.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-13637) improve setting of max connections in AWS client

2017-07-20 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13637?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16094592#comment-16094592
 ] 

Yonger edited comment on HADOOP-13637 at 7/20/17 12:13 PM:
---

[~ste...@apache.org] I don't think we should create a connection pool bigger 
than a work thread pool.
Assume the application run on Hadoop kick of request with X rate, each request 
cost T1 time on thread and T1 time on connection, obviously, T1 will bigger 
than T2 due to a request always get response to release a connection and then 
release a thread later . That means XT1>XT2, and according little's law, 
N1(threads are needed concurrently) >N2 (connections are needed concurrently). 
So the result is opposite of yours, we should set minimum thread pool size base 
on max connection + n.


was (Author: iyonger):
[~steve_l] I don't think we should create a connection pool bigger than a work 
thread pool.
Assume the application run on Hadoop kick of request with X rate, each request 
cost T1 time on thread and T1 time on connection, obviously, T1 will bigger 
than T2 due to a request always get response to release a connection and then 
release a thread later . That means XT1>XT2, and according little's law, 
N1(threads are needed concurrently) >N2 (connections are needed concurrently). 
So the result is opposite of yours, we should set minimum thread pool size base 
on max connection + n.

> improve setting of max connections in AWS client
> 
>
> Key: HADOOP-13637
> URL: https://issues.apache.org/jira/browse/HADOOP-13637
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0
>Reporter: Steve Loughran
>Priority: Minor
>
> things can go badly wrong if the S3A FS creates a thread pool for IO > than 
> the number of pooled AWS http connections (set by property 
> MAXIMUM_CONNECTIONS); you also need some for any other IO requests coming in.
> The max connections property is currently independent of thread pool size, 
> and has a default value of 1. 
> this is why there is a troubleshooting section in the docs showing the stack 
> trace and instructions to fix".
> Better: have a dynamic minimum like thread pool size + n, for a value of n to 
> be chosen.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13757) Remove verifyBuckets overhead in S3AFileSystem::initialize()

2017-07-20 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13757?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16094598#comment-16094598
 ] 

Yonger commented on HADOOP-13757:
-

[~ste...@apache.org]Can you give more specific suggestion on this issue, then I 
can try to work on it in my spare time.

> Remove verifyBuckets overhead in S3AFileSystem::initialize()
> 
>
> Key: HADOOP-13757
> URL: https://issues.apache.org/jira/browse/HADOOP-13757
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Reporter: Rajesh Balamohan
>Priority: Minor
>
> {{S3AFileSystem.initialize()}} invokes verifyBuckets, but in case the bucket 
> does not exist and gets a 403 error message, it ends up returning {{true}} 
> for {{s3.doesBucketExists(bucketName}}.  In that aspect,  verifyBuckets() is 
> an unnecessary call during initialization. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13637) improve setting of max connections in AWS client

2017-07-20 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13637?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16094592#comment-16094592
 ] 

Yonger commented on HADOOP-13637:
-

[~steve_l] I don't think we should create a connection pool bigger than a work 
thread pool.
Assume the application run on Hadoop kick of request with X rate, each request 
cost T1 time on thread and T1 time on connection, obviously, T1 will bigger 
than T2 due to a request always get response to release a connection and then 
release a thread later . That means XT1>XT2, and according little's law, 
N1(threads are needed concurrently) >N2 (connections are needed concurrently). 
So the result is opposite of yours, we should set minimum thread pool size base 
on max connection + n.

> improve setting of max connections in AWS client
> 
>
> Key: HADOOP-13637
> URL: https://issues.apache.org/jira/browse/HADOOP-13637
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0
>Reporter: Steve Loughran
>Priority: Minor
>
> things can go badly wrong if the S3A FS creates a thread pool for IO > than 
> the number of pooled AWS http connections (set by property 
> MAXIMUM_CONNECTIONS); you also need some for any other IO requests coming in.
> The max connections property is currently independent of thread pool size, 
> and has a default value of 1. 
> this is why there is a troubleshooting section in the docs showing the stack 
> trace and instructions to fix".
> Better: have a dynamic minimum like thread pool size + n, for a value of n to 
> be chosen.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-07-19 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Attachment: HADOOP-14475-003.patch

compliant with raw code format and update the logic of metrics register name 

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, failsafe-report-scale.html, 
> failsafe-report-scale.zip, HADOOP-14475.002.patch, HADOOP-14475-003.patch, 
> s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-07-19 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Attachment: failsafe-report-scale.html
failsafe-report-scale.zip

Only one case failed.
mvn -Dscale clean verify and skip s3 and  s3n cases 

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, failsafe-report-scale.html, 
> failsafe-report-scale.zip, HADOOP-14475.002.patch, s3a-metrics.patch1, 
> stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-07-06 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16076511#comment-16076511
 ] 

Yonger edited comment on HADOOP-14475 at 7/6/17 1:42 PM:
-

@steve the method you mentioned give an empty url to skip the landsat-pds tests 
is not work, also I upload the gz file into my bucket according to the guide, 
but it failed too.
when giving the empty string, error message:
Tests run: 9, Failures: 0, Errors: 6, Skipped: 0, Time elapsed: 0.325 sec <<< 
FAILURE! - in org.apache.hadoop.fs.s3a.TestS3AAWSCredentialsProvider
testInstantiationChain(org.apache.hadoop.fs.s3a.TestS3AAWSCredentialsProvider)  
Time elapsed: 0.018 sec  <<< ERROR!
java.lang.IllegalArgumentException: Can not create a Path from an empty string
at org.apache.hadoop.fs.Path.checkPathArg(Path.java:163)
at org.apache.hadoop.fs.Path.(Path.java:175)
at 
org.apache.hadoop.fs.s3a.TestS3AAWSCredentialsProvider.testInstantiationChain(TestS3AAWSCredentialsProvider.java:92)

and if i use default value and upload the gz file, which give me a error 
message with code 403.




was (Author: iyonger):
[~stevea] the method you mentioned give an empty url to skip the landsat-pds 
tests is not work, also I upload the gz file into my bucket according to the 
guide, but it failed too.
when giving the empty string, error message:
Tests run: 9, Failures: 0, Errors: 6, Skipped: 0, Time elapsed: 0.325 sec <<< 
FAILURE! - in org.apache.hadoop.fs.s3a.TestS3AAWSCredentialsProvider
testInstantiationChain(org.apache.hadoop.fs.s3a.TestS3AAWSCredentialsProvider)  
Time elapsed: 0.018 sec  <<< ERROR!
java.lang.IllegalArgumentException: Can not create a Path from an empty string
at org.apache.hadoop.fs.Path.checkPathArg(Path.java:163)
at org.apache.hadoop.fs.Path.(Path.java:175)
at 
org.apache.hadoop.fs.s3a.TestS3AAWSCredentialsProvider.testInstantiationChain(TestS3AAWSCredentialsProvider.java:92)

and if i use default value and upload the gz file, which give me a error 
message with code 403.



> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, HADOOP-14475.002.patch, s3a-metrics.patch1, 
> stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-07-06 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16076511#comment-16076511
 ] 

Yonger commented on HADOOP-14475:
-

[~stevea] the method you mentioned give an empty url to skip the landsat-pds 
tests is not work, also I upload the gz file into my bucket according to the 
guide, but it failed too.
when giving the empty string, error message:
Tests run: 9, Failures: 0, Errors: 6, Skipped: 0, Time elapsed: 0.325 sec <<< 
FAILURE! - in org.apache.hadoop.fs.s3a.TestS3AAWSCredentialsProvider
testInstantiationChain(org.apache.hadoop.fs.s3a.TestS3AAWSCredentialsProvider)  
Time elapsed: 0.018 sec  <<< ERROR!
java.lang.IllegalArgumentException: Can not create a Path from an empty string
at org.apache.hadoop.fs.Path.checkPathArg(Path.java:163)
at org.apache.hadoop.fs.Path.(Path.java:175)
at 
org.apache.hadoop.fs.s3a.TestS3AAWSCredentialsProvider.testInstantiationChain(TestS3AAWSCredentialsProvider.java:92)

and if i use default value and upload the gz file, which give me a error 
message with code 403.



> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, HADOOP-14475.002.patch, s3a-metrics.patch1, 
> stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-06-21 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Attachment: failsafe-report-s3a-it.html

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, HADOOP-14475.002.patch, s3a-metrics.patch1, 
> stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-06-21 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Attachment: (was: failsafe-report-it.html)

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: failsafe-report-s3a-it.html, 
> failsafe-report-s3a-scale.html, HADOOP-14475.002.patch, s3a-metrics.patch1, 
> stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-06-21 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Attachment: failsafe-report-it.html
failsafe-report-s3a-scale.html

I run cases on Ceph storage, so part of them failed due to can't find something 
in s3 storage.

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: failsafe-report-it.html, failsafe-report-s3a-scale.html, 
> HADOOP-14475.002.patch, s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-06-21 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16058815#comment-16058815
 ] 

Yonger edited comment on HADOOP-14475 at 6/22/17 6:18 AM:
--

Fixed this issue by initializing metrics system and registering s3a metrics 
source implementation


was (Author: iyonger):
Fixed this issue by initializing metrics system and registering s3a metrcis 
source implementation

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: HADOOP-14475.002.patch, s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-06-21 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Attachment: HADOOP-14475.002.patch

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: HADOOP-14475.002.patch, s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-06-21 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Release Note:   (was: fix S3AInstrumentation registration issue )
  Status: Patch Available  (was: Open)

Fixed this issue by initializing metrics system and registering s3a metrcis 
source implementation

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-06-21 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Status: Open  (was: Patch Available)

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>Assignee: Yonger
> Attachments: s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-06-20 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16056807#comment-16056807
 ] 

Yonger commented on HADOOP-14475:
-

@Deng, each task should have an individual s3a file system that i agree with 
you. But i seen the s3a file system had been initialized multiple times with 
the same path in driver application, you can also check in my last uploaded 
logs.
BTW, the new patch that fix this issue totally is in the process of verifying,  
i don't know when it available to upload since i have no aws s3 test 
environment, i am running on Ceph now. 

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
> Attachments: s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-06-06 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16040018#comment-16040018
 ] 

Yonger commented on HADOOP-14475:
-

In my case, i just run DFSIO with 20 map/reduce tasks in a 4-nodes cluster. And 
i enable debug log to console, if you grep by "Metrics system initialized" the 
attached log file, you will find the s3a file system has been initialized 
multiple times under one bucket(s3a://test-bucket). 

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
> Attachments: s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-06-06 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Attachment: stdout.zip

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
> Attachments: s3a-metrics.patch1, stdout.zip
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-06-06 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16038864#comment-16038864
 ] 

Yonger commented on HADOOP-14475:
-

Thanks for your suggestion. I will take case for code style and do test by 
myself before submitting.

And answer your questions:
1. the name change of context just for distinguish with other attributes, such 
as MetricsRegistry and Metrics name. From the following log, it shows using 
different names is better than ones with the same name:
17/06/05 20:32:54 DEBUG impl.MetricsSinkAdapter: Pushing record 
S3AFileSystemMetrics.s3a.s3afilesystem to file
is it?

2.after i make a collection the relationship of those classes, i also think the 
functions of class S3AFileSystemMetricsSystem can be merge into some existed 
class, maybe S3AFileSystem.

3.that is the issue confused me. I still don't know why the 
filesystem(S3AFileSystem) be initialized multiple times in a MR job. From 
AzureFileSystem and DataNodeMetric, their filesystem and MetricSystem should be 
only initialized once.




> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
> Attachments: s3a-metrics.patch1
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-06-05 Thread Yonger (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16037032#comment-16037032
 ] 

Yonger edited comment on HADOOP-14475 at 6/5/17 2:28 PM:
-

This patch can fix the issue that s3a metrics don't print out to sink(file).But 
there also introduce an issue that the metrics info only print out once to file 
instead of printing out periodically.



was (Author: iyonger):
This patch can fix the issue that s3a metrics don't print out to sink(file).But 
there also introduce an issue that the metrics info only print out once in file 
instead of printing our periodically.


> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
> Attachments: s3a-metrics.patch1
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-06-05 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Attachment: s3a-metrics.patch1

This patch can fix the issue that s3a metrics don't print out to sink(file).But 
there also introduce an issue that the metrics info only print out once in file 
instead of printing our periodically.


> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
> Attachments: s3a-metrics.patch1
>
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-06-05 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Release Note: fix S3AInstrumentation registration issue 
  Status: Patch Available  (was: Open)

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-06-01 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Component/s: (was: s3)
 fs/s3

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A don't print out when enable it in Hadoop metrics property file

2017-06-01 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Summary: Metrics of S3A don't print out  when enable it in Hadoop metrics 
property file  (was: Metrics of S3A doesn't print out  when enable it in Hadoop 
metrics property file)

> Metrics of S3A don't print out  when enable it in Hadoop metrics property file
> --
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A doesn't print out when enable it in Hadoop metrics property file

2017-06-01 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Description: 
*.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
#*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
#*.sink.influxdb.url=http:/xx
#*.sink.influxdb.influxdb_port=8086
#*.sink.influxdb.database=hadoop
#*.sink.influxdb.influxdb_username=hadoop
#*.sink.influxdb.influxdb_password=hadoop
#*.sink.ingluxdb.cluster=c1
## default sampling period, in seconds
*.period=10
#namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
#S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out

I can't find the out put file even i run a MR job which should be used s3.


  was:
*.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
#*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
#*.sink.influxdb.url=http:/xx
#*.sink.influxdb.influxdb_port=8086
#*.sink.influxdb.database=hadoop
#*.sink.influxdb.influxdb_username=hadoop
#*.sink.influxdb.influxdb_password=hadoop
#*.sink.ingluxdb.cluster=c1
# default sampling period, in seconds
*.period=10
#namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
#S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out

I can't find the out put file even i run a MR job which should be used s3.



> Metrics of S3A doesn't print out  when enable it in Hadoop metrics property 
> file
> 
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> ## default sampling period, in seconds
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14475) Metrics of S3A doesn't print out when enable it in Hadoop metrics property file

2017-06-01 Thread Yonger (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonger updated HADOOP-14475:

Description: 
*.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
#*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
#*.sink.influxdb.url=http:/xx
#*.sink.influxdb.influxdb_port=8086
#*.sink.influxdb.database=hadoop
#*.sink.influxdb.influxdb_username=hadoop
#*.sink.influxdb.influxdb_password=hadoop
#*.sink.ingluxdb.cluster=c1

*.period=10
#namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
#S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out

I can't find the out put file even i run a MR job which should be used s3.


  was:
*.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
#*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
#*.sink.influxdb.url=http:/xx
#*.sink.influxdb.influxdb_port=8086
#*.sink.influxdb.database=hadoop
#*.sink.influxdb.influxdb_username=hadoop
#*.sink.influxdb.influxdb_password=hadoop
#*.sink.ingluxdb.cluster=c1
## default sampling period, in seconds
*.period=10
#namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
#S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out

I can't find the out put file even i run a MR job which should be used s3.



> Metrics of S3A doesn't print out  when enable it in Hadoop metrics property 
> file
> 
>
> Key: HADOOP-14475
> URL: https://issues.apache.org/jira/browse/HADOOP-14475
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: s3
>Affects Versions: 2.8.0
> Environment: uname -a
> Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
> x86_64 x86_64 x86_64 GNU/Linux
>  cat /etc/issue
> Ubuntu 16.04.2 LTS \n \l
>Reporter: Yonger
>
> *.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
> #*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #*.sink.influxdb.url=http:/xx
> #*.sink.influxdb.influxdb_port=8086
> #*.sink.influxdb.database=hadoop
> #*.sink.influxdb.influxdb_username=hadoop
> #*.sink.influxdb.influxdb_password=hadoop
> #*.sink.ingluxdb.cluster=c1
> *.period=10
> #namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> #S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
> S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out
> I can't find the out put file even i run a MR job which should be used s3.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-14475) Metrics of S3A doesn't print out when enable it in Hadoop metrics property file

2017-06-01 Thread Yonger (JIRA)
Yonger created HADOOP-14475:
---

 Summary: Metrics of S3A doesn't print out  when enable it in 
Hadoop metrics property file
 Key: HADOOP-14475
 URL: https://issues.apache.org/jira/browse/HADOOP-14475
 Project: Hadoop Common
  Issue Type: Bug
  Components: s3
Affects Versions: 2.8.0
 Environment: uname -a
Linux client01 4.4.0-74-generic #95-Ubuntu SMP Wed Apr 12 09:50:34 UTC 2017 
x86_64 x86_64 x86_64 GNU/Linux

 cat /etc/issue
Ubuntu 16.04.2 LTS \n \l
Reporter: Yonger


*.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
#*.sink.file.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
#*.sink.influxdb.url=http:/xx
#*.sink.influxdb.influxdb_port=8086
#*.sink.influxdb.database=hadoop
#*.sink.influxdb.influxdb_username=hadoop
#*.sink.influxdb.influxdb_password=hadoop
#*.sink.ingluxdb.cluster=c1
# default sampling period, in seconds
*.period=10
#namenode.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
#S3AFileSystem.sink.influxdb.class=org.apache.hadoop.metrics2.sink.influxdb.InfluxdbSink
S3AFileSystem.sink.file.filename=s3afilesystem-metrics.out

I can't find the out put file even i run a MR job which should be used s3.




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org