[jira] [Resolved] (HDFS-4925) TestQuorumJournalManager.testPurgeLogs intermittently Fails assertNoThreadsMatching

2016-02-09 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4925?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA resolved HDFS-4925.
-
   Resolution: Duplicate
Fix Version/s: (was: 2.6.5)
   (was: 2.7.3)
   (was: 2.8.0)

> TestQuorumJournalManager.testPurgeLogs intermittently Fails 
> assertNoThreadsMatching
> ---
>
> Key: HDFS-4925
> URL: https://issues.apache.org/jira/browse/HDFS-4925
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: qjm, test
>Affects Versions: 3.0.0
>Reporter: Tao Luo
>Priority: Minor
> Attachments: TestQuorumJournalManager.rtf
>
>
> On Jenkins, I saw this assert fail in the @After shutdown() for 
> TestQuorumJournalManager.testPurgeLogs():
> {code}// Should not leak clients between tests -- this can cause flaky 
> tests.
> // (See HDFS-4643)
> GenericTestUtils.assertNoThreadsMatching(".*IPC Client.*");
> {code}
> https://builds.apache.org/job/PreCommit-HDFS-Build/4546//testReport/
> I could not reproduce locally.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-4925) TestQuorumJournalManager.testPurgeLogs intermittently Fails assertNoThreadsMatching

2016-02-09 Thread Akira AJISAKA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4925?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140461#comment-15140461
 ] 

Akira AJISAKA commented on HDFS-4925:
-

Thanks [~jojochuang] for fixing the test. I'll reopen and (re)close it to 
change the resolution from "Fixed" to "Duplicate".

> TestQuorumJournalManager.testPurgeLogs intermittently Fails 
> assertNoThreadsMatching
> ---
>
> Key: HDFS-4925
> URL: https://issues.apache.org/jira/browse/HDFS-4925
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: qjm, test
>Affects Versions: 3.0.0
>Reporter: Tao Luo
>Priority: Minor
> Attachments: TestQuorumJournalManager.rtf
>
>
> On Jenkins, I saw this assert fail in the @After shutdown() for 
> TestQuorumJournalManager.testPurgeLogs():
> {code}// Should not leak clients between tests -- this can cause flaky 
> tests.
> // (See HDFS-4643)
> GenericTestUtils.assertNoThreadsMatching(".*IPC Client.*");
> {code}
> https://builds.apache.org/job/PreCommit-HDFS-Build/4546//testReport/
> I could not reproduce locally.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Reopened] (HDFS-4925) TestQuorumJournalManager.testPurgeLogs intermittently Fails assertNoThreadsMatching

2016-02-09 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4925?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA reopened HDFS-4925:
-

> TestQuorumJournalManager.testPurgeLogs intermittently Fails 
> assertNoThreadsMatching
> ---
>
> Key: HDFS-4925
> URL: https://issues.apache.org/jira/browse/HDFS-4925
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: qjm, test
>Affects Versions: 3.0.0
>Reporter: Tao Luo
>Priority: Minor
> Attachments: TestQuorumJournalManager.rtf
>
>
> On Jenkins, I saw this assert fail in the @After shutdown() for 
> TestQuorumJournalManager.testPurgeLogs():
> {code}// Should not leak clients between tests -- this can cause flaky 
> tests.
> // (See HDFS-4643)
> GenericTestUtils.assertNoThreadsMatching(".*IPC Client.*");
> {code}
> https://builds.apache.org/job/PreCommit-HDFS-Build/4546//testReport/
> I could not reproduce locally.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HDFS-4925) TestQuorumJournalManager.testPurgeLogs intermittently Fails assertNoThreadsMatching

2016-02-09 Thread Wei-Chiu Chuang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4925?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wei-Chiu Chuang resolved HDFS-4925.
---
   Resolution: Fixed
Fix Version/s: 2.6.5
   2.7.3
   2.8.0

I am going to resolve this as fixed. The assertion failure was fixed via 
HDFS-9347.

> TestQuorumJournalManager.testPurgeLogs intermittently Fails 
> assertNoThreadsMatching
> ---
>
> Key: HDFS-4925
> URL: https://issues.apache.org/jira/browse/HDFS-4925
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: qjm, test
>Affects Versions: 3.0.0
>Reporter: Tao Luo
>Priority: Minor
> Fix For: 2.8.0, 2.7.3, 2.6.5
>
> Attachments: TestQuorumJournalManager.rtf
>
>
> On Jenkins, I saw this assert fail in the @After shutdown() for 
> TestQuorumJournalManager.testPurgeLogs():
> {code}// Should not leak clients between tests -- this can cause flaky 
> tests.
> // (See HDFS-4643)
> GenericTestUtils.assertNoThreadsMatching(".*IPC Client.*");
> {code}
> https://builds.apache.org/job/PreCommit-HDFS-Build/4546//testReport/
> I could not reproduce locally.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Guocui Mi (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140396#comment-15140396
 ] 

Guocui Mi commented on HDFS-9787:
-

Agree with you on this, should create one new workitem for this. 
1> Non-primary SNNs check ANN checkpoint version via RPC;
2> Check if checkpoint period beyond threshold 
   Yes : try to upload checkpoint to ANN. become primary SNN if success.
3> repeat 1> 2>.

This patch is just a simple bug fix to make it work as expected first.

> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to 
> false.
> ---
>
> Key: HDFS-9787
> URL: https://issues.apache.org/jira/browse/HDFS-9787
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 3.0.0
>Reporter: Guocui Mi
>Assignee: Guocui Mi
> Attachments: HDFS-9786-v000.patch
>
>
> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
> Here is the logic to check if upload FSImage or not.
> In StandbyCheckpointer.java
> boolean sendRequest = isPrimaryCheckPointer || secsSinceLast >= 
> checkpointConf.getQuietPeriod();
> doCheckpoint(sendRequest);
> The sendRequest is always false if isPrimaryCheckPointer is false giving 
> secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
> (checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
> false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Jesse Yates (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140384#comment-15140384
 ] 

Jesse Yates commented on HDFS-9787:
---

The original solution was an attempting to catch the case where we don't flood 
the NN with checkpoint requests. Instead, maybe the better solution would be to 
do a small RPC to see when the latest image was uploaded. If it was uploaded 
the quietMultiplier beyond the checkpoint period, then we attempt to upload the 
checkpoint.

Its a bit more work, but I think this more clearly lays out the intentions in 
the code, rather than obtaining the same effect, but without the overhead of 
actually sending the checkpoint along each time we want to find out if its 
behind.

> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to 
> false.
> ---
>
> Key: HDFS-9787
> URL: https://issues.apache.org/jira/browse/HDFS-9787
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 3.0.0
>Reporter: Guocui Mi
>Assignee: Guocui Mi
> Attachments: HDFS-9786-v000.patch
>
>
> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
> Here is the logic to check if upload FSImage or not.
> In StandbyCheckpointer.java
> boolean sendRequest = isPrimaryCheckPointer || secsSinceLast >= 
> checkpointConf.getQuietPeriod();
> doCheckpoint(sendRequest);
> The sendRequest is always false if isPrimaryCheckPointer is false giving 
> secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
> (checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
> false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Guocui Mi (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140372#comment-15140372
 ] 

Guocui Mi commented on HDFS-9787:
-

lastCheckpointTime is also used to check if it is necessary to do another 
checkpoint. we can't mix up last checkpoint time and last upload time in order 
to honor dfs.namenode.checkpoint.period
FYI: else if (secsSinceLast >= checkpointConf.getPeriod()) {

> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to 
> false.
> ---
>
> Key: HDFS-9787
> URL: https://issues.apache.org/jira/browse/HDFS-9787
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 3.0.0
>Reporter: Guocui Mi
>Assignee: Guocui Mi
> Attachments: HDFS-9786-v000.patch
>
>
> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
> Here is the logic to check if upload FSImage or not.
> In StandbyCheckpointer.java
> boolean sendRequest = isPrimaryCheckPointer || secsSinceLast >= 
> checkpointConf.getQuietPeriod();
> doCheckpoint(sendRequest);
> The sendRequest is always false if isPrimaryCheckPointer is false giving 
> secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
> (checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
> false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Guocui Mi (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140367#comment-15140367
 ] 

Guocui Mi commented on HDFS-9787:
-

>>> this would imply that the non-primary SNN never sends a checkpoint after 
>>> the first time?
It is true according to my observation.
I am trying to add unittest to cover the scenario. Another two scenarios 
triggered in our cluster:
1) PrimaryCheckpoint uploading fsimage failure due to ANN not available 
temporarily.
2) Restart all NNs at same time.

I afraid the proposal you shared can't work.
1) set lastCheckpointTime before following code in doCheckpoint(): no 
difference between putting after each loop iteration.
2) after following code in doCheckpoint() :  Non-primary SNN will do checkpoint 
one by one continuously since lastCheckpointTime not get updated.
if(!sendCheckpoint){  return;}

> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to 
> false.
> ---
>
> Key: HDFS-9787
> URL: https://issues.apache.org/jira/browse/HDFS-9787
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 3.0.0
>Reporter: Guocui Mi
>Assignee: Guocui Mi
> Attachments: HDFS-9786-v000.patch
>
>
> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
> Here is the logic to check if upload FSImage or not.
> In StandbyCheckpointer.java
> boolean sendRequest = isPrimaryCheckPointer || secsSinceLast >= 
> checkpointConf.getQuietPeriod();
> doCheckpoint(sendRequest);
> The sendRequest is always false if isPrimaryCheckPointer is false giving 
> secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
> (checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
> false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9637) Add test for HADOOP-12702 and HADOOP-12759

2016-02-09 Thread Daniel Templeton (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9637?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Templeton updated HDFS-9637:
---
Attachment: HDFS-9637.006.patch

Here's one more little update.  I noticed some formatting that could be 
cleaner, and I moved the flag resets to the setup method.

> Add test for HADOOP-12702 and HADOOP-12759
> --
>
> Key: HDFS-9637
> URL: https://issues.apache.org/jira/browse/HDFS-9637
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: test
>Affects Versions: 2.7.1
>Reporter: Daniel Templeton
>Assignee: Daniel Templeton
> Attachments: HDFS-9637.001.patch, HDFS-9637.002.patch, 
> HDFS-9637.003.patch, HDFS-9637.004.patch, HDFS-9637.005.patch, 
> HDFS-9637.006.patch
>
>
> Per discussion on the dev list, the tests for the new FileSystemSink class 
> should be added to the HDFS project to avoid creating a dependency for the 
> common project on the HDFS project.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9780) RollingFileSystemSink doesn't work on secure clusters

2016-02-09 Thread Karthik Kambatla (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9780?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140277#comment-15140277
 ] 

Karthik Kambatla commented on HDFS-9780:


Quickly skimmed through the patch. 

Do we want to checkForProperty even if security is not enabled? 
{code}
// Validate config so that we don't get an NPE
checkForProperty(conf, KEYTAB_PROPERTY_KEY);
checkForProperty(conf, USERNAME_PROPERTY_KEY);
{code}

Once the patch is updated based on HDFS-9637, will take a closer look. 

> RollingFileSystemSink doesn't work on secure clusters
> -
>
> Key: HDFS-9780
> URL: https://issues.apache.org/jira/browse/HDFS-9780
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.9.0
>Reporter: Daniel Templeton
>Assignee: Daniel Templeton
>Priority: Critical
> Attachments: HADOOP-12775.001.patch, HADOOP-12775.002.patch, 
> HADOOP-12775.003.patch, HDFS-9780.004.patch
>
>
> If HDFS has kerberos enabled, the sink cannot write its logs.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9780) RollingFileSystemSink doesn't work on secure clusters

2016-02-09 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9780?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla updated HDFS-9780:
---
Status: Open  (was: Patch Available)

Canceling patch until HDFS-9637 gets committed. 

> RollingFileSystemSink doesn't work on secure clusters
> -
>
> Key: HDFS-9780
> URL: https://issues.apache.org/jira/browse/HDFS-9780
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.9.0
>Reporter: Daniel Templeton
>Assignee: Daniel Templeton
>Priority: Critical
> Attachments: HADOOP-12775.001.patch, HADOOP-12775.002.patch, 
> HADOOP-12775.003.patch, HDFS-9780.004.patch
>
>
> If HDFS has kerberos enabled, the sink cannot write its logs.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140258#comment-15140258
 ] 

Hadoop QA commented on HDFS-9787:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 14s 
{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s 
{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s 
{color} | {color:red} The patch doesn't appear to include any new or modified 
tests. Please justify why no new tests are needed for this patch. Also please 
list what manual steps were performed to verify this patch. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 
34s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 40s 
{color} | {color:green} trunk passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 42s 
{color} | {color:green} trunk passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 
21s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 52s 
{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
13s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 
52s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 5s 
{color} | {color:green} trunk passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 46s 
{color} | {color:green} trunk passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 
46s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 37s 
{color} | {color:green} the patch passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 37s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 39s 
{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 39s 
{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red} 0m 18s 
{color} | {color:red} hadoop-hdfs-project/hadoop-hdfs: patch generated 1 new + 
26 unchanged - 1 fixed = 27 total (was 27) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 49s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
12s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 
0s {color} | {color:green} Patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 4s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 0s 
{color} | {color:green} the patch passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 43s 
{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 62m 54s {color} 
| {color:red} hadoop-hdfs in the patch failed with JDK v1.8.0_72. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 63m 41s {color} 
| {color:red} hadoop-hdfs in the patch failed with JDK v1.7.0_95. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 
20s {color} | {color:green} Patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 150m 22s {color} 
| {color:black} {color} |
\\
\\
|| Reason || Tests ||
| JDK v1.8.0_72 Failed junit tests | hadoop.fs.TestHdfsNativeCodeLoader |
| JDK v1.7.0_95 Failed junit tests | hadoop.fs.TestHdfsNativeCodeLoader |
|   | hadoop.hdfs.server.namenode.TestFileTruncate |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:0ca8df7 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12787193/HDFS-9786-v000.patch |
| JIRA Issue | HDFS-9787 |
| Optional Tests |  asflicense  compile  javac  javad

[jira] [Commented] (HDFS-9637) Add test for HADOOP-12702 and HADOOP-12759

2016-02-09 Thread Karthik Kambatla (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9637?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140252#comment-15140252
 ] 

Karthik Kambatla commented on HDFS-9637:


Looks good. +1. Will check this in tomorrow morning if I don't hear any 
objections. 

> Add test for HADOOP-12702 and HADOOP-12759
> --
>
> Key: HDFS-9637
> URL: https://issues.apache.org/jira/browse/HDFS-9637
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: test
>Affects Versions: 2.7.1
>Reporter: Daniel Templeton
>Assignee: Daniel Templeton
> Attachments: HDFS-9637.001.patch, HDFS-9637.002.patch, 
> HDFS-9637.003.patch, HDFS-9637.004.patch, HDFS-9637.005.patch
>
>
> Per discussion on the dev list, the tests for the new FileSystemSink class 
> should be added to the HDFS project to avoid creating a dependency for the 
> common project on the HDFS project.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9637) Add test for HADOOP-12702 and HADOOP-12759

2016-02-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9637?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140247#comment-15140247
 ] 

Hadoop QA commented on HDFS-9637:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 10s 
{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s 
{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 
0s {color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 
44s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 41s 
{color} | {color:green} trunk passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 42s 
{color} | {color:green} trunk passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 
21s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 51s 
{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
14s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 
55s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 7s 
{color} | {color:green} trunk passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 47s 
{color} | {color:green} trunk passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 
46s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 34s 
{color} | {color:green} the patch passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 34s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 37s 
{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 37s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 
17s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 49s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
11s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 
0s {color} | {color:green} Patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 3s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 59s 
{color} | {color:green} the patch passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 43s 
{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 55m 23s {color} 
| {color:red} hadoop-hdfs in the patch failed with JDK v1.8.0_72. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 50m 5s {color} 
| {color:red} hadoop-hdfs in the patch failed with JDK v1.7.0_95. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 
21s {color} | {color:green} Patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 129m 21s {color} 
| {color:black} {color} |
\\
\\
|| Reason || Tests ||
| JDK v1.8.0_72 Failed junit tests | hadoop.hdfs.TestFileAppend |
|   | hadoop.hdfs.server.datanode.TestBlockScanner |
|   | hadoop.fs.TestHdfsNativeCodeLoader |
| JDK v1.7.0_95 Failed junit tests | hadoop.fs.TestHdfsNativeCodeLoader |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:0ca8df7 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12787194/HDFS-9637.005.patch |
| JIRA Issue | HDFS-9637 |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux 18ad0c04c7cb 3.13.0-36-lowlatency #63-Ubuntu SMP PREEMPT Wed 
Sep 3 21:56:12 UTC 2014 x86_64 x86_6

[jira] [Updated] (HDFS-9703) DiskBalancer : getBandwidth implementation

2016-02-09 Thread Anu Engineer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9703?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anu Engineer updated HDFS-9703:
---
Attachment: HDFS-9703-HDFS-1312.001.patch

> DiskBalancer : getBandwidth implementation
> --
>
> Key: HDFS-9703
> URL: https://issues.apache.org/jira/browse/HDFS-9703
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: HDFS-1312
>Reporter: Anu Engineer
>Assignee: Anu Engineer
> Fix For: HDFS-1312
>
> Attachments: HDFS-9703-HDFS-1312.001.patch
>
>
> Add getBandwidth call



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9703) DiskBalancer : getBandwidth implementation

2016-02-09 Thread Anu Engineer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9703?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anu Engineer updated HDFS-9703:
---
Attachment: (was: HDFS-9703-HDFS-1312.001.patch)

> DiskBalancer : getBandwidth implementation
> --
>
> Key: HDFS-9703
> URL: https://issues.apache.org/jira/browse/HDFS-9703
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: HDFS-1312
>Reporter: Anu Engineer
>Assignee: Anu Engineer
> Fix For: HDFS-1312
>
> Attachments: HDFS-9703-HDFS-1312.001.patch
>
>
> Add getBandwidth call



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9779) TestReplicationPolicyWithNodeGroup NODE variable picks wrong rack value

2016-02-09 Thread Kuhu Shukla (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9779?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140221#comment-15140221
 ] 

Kuhu Shukla commented on HDFS-9779:
---

Thank you Uma!

> TestReplicationPolicyWithNodeGroup NODE variable picks wrong rack value
> ---
>
> Key: HDFS-9779
> URL: https://issues.apache.org/jira/browse/HDFS-9779
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.7.2
>Reporter: Kuhu Shukla
>Assignee: Kuhu Shukla
>Priority: Minor
> Fix For: 2.8.0
>
> Attachments: HDFS-9779.001.patch
>
>
> The DatanodeDescriptor object for {{NODE}} calls the wrong constructor on an 
> already created DatanodeDescriptor object. This overwrites the 
> networkLocation value to "default-rack" instead of expected value of 
> "/d2/r4/n7". I didn't find any other such malformed calls to 
> {{DFSTestUtil.getDatanodeDescriptor}}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Inigo Goiri (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140220#comment-15140220
 ] 

Inigo Goiri commented on HDFS-9787:
---

My bad, this is right.
We will test resetting the {{lastCheckpointTime}} in every loop.

> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to 
> false.
> ---
>
> Key: HDFS-9787
> URL: https://issues.apache.org/jira/browse/HDFS-9787
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 3.0.0
>Reporter: Guocui Mi
>Assignee: Guocui Mi
> Attachments: HDFS-9786-v000.patch
>
>
> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
> Here is the logic to check if upload FSImage or not.
> In StandbyCheckpointer.java
> boolean sendRequest = isPrimaryCheckPointer || secsSinceLast >= 
> checkpointConf.getQuietPeriod();
> doCheckpoint(sendRequest);
> The sendRequest is always false if isPrimaryCheckPointer is false giving 
> secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
> (checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
> false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Inigo Goiri (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140218#comment-15140218
 ] 

Inigo Goiri commented on HDFS-9787:
---

Now that I see this, can it be the problem that we do:
{code}
// Reset checkpoint time so that we don't always checkpoint
// on startup.
lastCheckpointTime = monotonicNow();
{code}

And then:
{code}
lastCheckpointTime = now;
{code}

We'll do the test in our cluster anyway.

> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to 
> false.
> ---
>
> Key: HDFS-9787
> URL: https://issues.apache.org/jira/browse/HDFS-9787
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 3.0.0
>Reporter: Guocui Mi
>Assignee: Guocui Mi
> Attachments: HDFS-9786-v000.patch
>
>
> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
> Here is the logic to check if upload FSImage or not.
> In StandbyCheckpointer.java
> boolean sendRequest = isPrimaryCheckPointer || secsSinceLast >= 
> checkpointConf.getQuietPeriod();
> doCheckpoint(sendRequest);
> The sendRequest is always false if isPrimaryCheckPointer is false giving 
> secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
> (checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
> false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9711) Integrate CSRF prevention filter in WebHDFS.

2016-02-09 Thread Larry McCay (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9711?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140195#comment-15140195
 ] 

Larry McCay commented on HDFS-9711:
---

Hi [~cnauroth] - Looks great! 
The effort to add a filter for webhdfs is greater than I anticipated.

a couple quick things:

* I like the refactoring for an isRequestAllowed method on the filter - I 
actually meant to go back and do that earlier
* I notice that you have to return your own error message in the channelRead0 
method of RestCsrfPreventionFilterHandler. Perhaps, we should provide a 
constant for that in the filter too. As it stands now, the message you return 
is slightly different and a bit more ambiguous then what is returned by the 
filter itself (which is why I changed it).
* I'd also like to understand why the typical filter processing isn't being 
used in this code path. Not because I think it should but I'd like to 
understand the usecase here.

> Integrate CSRF prevention filter in WebHDFS.
> 
>
> Key: HDFS-9711
> URL: https://issues.apache.org/jira/browse/HDFS-9711
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: datanode, namenode, webhdfs
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-9711.001.patch, HDFS-9711.002.patch, 
> HDFS-9711.003.patch
>
>
> HADOOP-12691 introduced a filter in Hadoop Common to help REST APIs guard 
> against cross-site request forgery attacks.  This issue tracks integration of 
> that filter in WebHDFS.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Jesse Yates (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140176#comment-15140176
 ] 

Jesse Yates commented on HDFS-9787:
---

I don't think you need a new variable - lastCheckpointTime is never used 
outside of the class and only used for for this check

> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to 
> false.
> ---
>
> Key: HDFS-9787
> URL: https://issues.apache.org/jira/browse/HDFS-9787
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 3.0.0
>Reporter: Guocui Mi
>Assignee: Guocui Mi
> Attachments: HDFS-9786-v000.patch
>
>
> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
> Here is the logic to check if upload FSImage or not.
> In StandbyCheckpointer.java
> boolean sendRequest = isPrimaryCheckPointer || secsSinceLast >= 
> checkpointConf.getQuietPeriod();
> doCheckpoint(sendRequest);
> The sendRequest is always false if isPrimaryCheckPointer is false giving 
> secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
> (checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
> false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Jesse Yates (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140175#comment-15140175
 ] 

Jesse Yates commented on HDFS-9787:
---

Taking a quick look, this would imply that the non-primary SNN never sends a 
checkpoint after the first time? A good test to ensure that this is the case is 
to start the NNs, wait until there primary SNN is selected and then remove it 
from the cluster. Are any more checkpoints sent to the ANN?

My inclination is that you are correct, no (unless it takes a long time to 
build the checkpoint), but I'd like to hear if that's actually the case. I 
think the fix is to just set lastCheckpointTime in doCheckpoint() rather than 
after each loop iteration.

> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to 
> false.
> ---
>
> Key: HDFS-9787
> URL: https://issues.apache.org/jira/browse/HDFS-9787
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 3.0.0
>Reporter: Guocui Mi
>Assignee: Guocui Mi
> Attachments: HDFS-9786-v000.patch
>
>
> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
> Here is the logic to check if upload FSImage or not.
> In StandbyCheckpointer.java
> boolean sendRequest = isPrimaryCheckPointer || secsSinceLast >= 
> checkpointConf.getQuietPeriod();
> doCheckpoint(sendRequest);
> The sendRequest is always false if isPrimaryCheckPointer is false giving 
> secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
> (checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
> false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Inigo Goiri (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Inigo Goiri updated HDFS-9787:
--
Status: Patch Available  (was: Open)

> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to 
> false.
> ---
>
> Key: HDFS-9787
> URL: https://issues.apache.org/jira/browse/HDFS-9787
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 3.0.0
>Reporter: Guocui Mi
>Assignee: Guocui Mi
> Attachments: HDFS-9786-v000.patch
>
>
> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
> Here is the logic to check if upload FSImage or not.
> In StandbyCheckpointer.java
> boolean sendRequest = isPrimaryCheckPointer || secsSinceLast >= 
> checkpointConf.getQuietPeriod();
> doCheckpoint(sendRequest);
> The sendRequest is always false if isPrimaryCheckPointer is false giving 
> secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
> (checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
> false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Inigo Goiri (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140162#comment-15140162
 ] 

Inigo Goiri commented on HDFS-9787:
---

Thanks!

> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to 
> false.
> ---
>
> Key: HDFS-9787
> URL: https://issues.apache.org/jira/browse/HDFS-9787
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 3.0.0
>Reporter: Guocui Mi
>Assignee: Guocui Mi
> Attachments: HDFS-9786-v000.patch
>
>
> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
> Here is the logic to check if upload FSImage or not.
> In StandbyCheckpointer.java
> boolean sendRequest = isPrimaryCheckPointer || secsSinceLast >= 
> checkpointConf.getQuietPeriod();
> doCheckpoint(sendRequest);
> The sendRequest is always false if isPrimaryCheckPointer is false giving 
> secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
> (checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
> false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9637) Add test for HADOOP-12702 and HADOOP-12759

2016-02-09 Thread Daniel Templeton (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9637?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Templeton updated HDFS-9637:
---
Attachment: HDFS-9637.005.patch

Huh.  Good point, [~kasha].  At one time there was a reason for not using 
{{\@Before}} and {{\@After}}, but clearly that's no longer the case...  This 
new patch removes the try-finally blocks.

> Add test for HADOOP-12702 and HADOOP-12759
> --
>
> Key: HDFS-9637
> URL: https://issues.apache.org/jira/browse/HDFS-9637
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: test
>Affects Versions: 2.7.1
>Reporter: Daniel Templeton
>Assignee: Daniel Templeton
> Attachments: HDFS-9637.001.patch, HDFS-9637.002.patch, 
> HDFS-9637.003.patch, HDFS-9637.004.patch, HDFS-9637.005.patch
>
>
> Per discussion on the dev list, the tests for the new FileSystemSink class 
> should be added to the HDFS project to avoid creating a dependency for the 
> common project on the HDFS project.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Guocui Mi (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Guocui Mi updated HDFS-9787:

Attachment: HDFS-9786-v000.patch

check last upload time instead of last check point time.

> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to 
> false.
> ---
>
> Key: HDFS-9787
> URL: https://issues.apache.org/jira/browse/HDFS-9787
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 3.0.0
>Reporter: Guocui Mi
>Assignee: Guocui Mi
> Attachments: HDFS-9786-v000.patch
>
>
> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
> Here is the logic to check if upload FSImage or not.
> In StandbyCheckpointer.java
> boolean sendRequest = isPrimaryCheckPointer || secsSinceLastUpload >= 
> checkpointConf.getQuietPeriod();
> doCheckpoint(sendRequest);
> The sendRequest is always false if isPrimaryCheckPointer is false giving 
> secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
> (checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
> false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Guocui Mi (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Guocui Mi updated HDFS-9787:

Description: 
SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
Here is the logic to check if upload FSImage or not.
In StandbyCheckpointer.java
boolean sendRequest = isPrimaryCheckPointer || secsSinceLast >= 
checkpointConf.getQuietPeriod();
doCheckpoint(sendRequest);

The sendRequest is always false if isPrimaryCheckPointer is false giving 
secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
(checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
false.

  was:
SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
Here is the logic to check if upload FSImage or not.
In StandbyCheckpointer.java
boolean sendRequest = isPrimaryCheckPointer || secsSinceLastUpload >= 
checkpointConf.getQuietPeriod();
doCheckpoint(sendRequest);

The sendRequest is always false if isPrimaryCheckPointer is false giving 
secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
(checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
false.


> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to 
> false.
> ---
>
> Key: HDFS-9787
> URL: https://issues.apache.org/jira/browse/HDFS-9787
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 3.0.0
>Reporter: Guocui Mi
>Assignee: Guocui Mi
> Attachments: HDFS-9786-v000.patch
>
>
> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
> Here is the logic to check if upload FSImage or not.
> In StandbyCheckpointer.java
> boolean sendRequest = isPrimaryCheckPointer || secsSinceLast >= 
> checkpointConf.getQuietPeriod();
> doCheckpoint(sendRequest);
> The sendRequest is always false if isPrimaryCheckPointer is false giving 
> secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
> (checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
> false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Jesse Yates (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140148#comment-15140148
 ] 

Jesse Yates commented on HDFS-9787:
---

I think I'll have some time tomorrow - adding it to my list

> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to 
> false.
> ---
>
> Key: HDFS-9787
> URL: https://issues.apache.org/jira/browse/HDFS-9787
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 3.0.0
>Reporter: Guocui Mi
>Assignee: Guocui Mi
>
> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
> Here is the logic to check if upload FSImage or not.
> In StandbyCheckpointer.java
> boolean sendRequest = isPrimaryCheckPointer || secsSinceLastUpload >= 
> checkpointConf.getQuietPeriod();
> doCheckpoint(sendRequest);
> The sendRequest is always false if isPrimaryCheckPointer is false giving 
> secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
> (checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
> false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Inigo Goiri (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140146#comment-15140146
 ] 

Inigo Goiri commented on HDFS-9787:
---

[~jesse_yates], do you mind taking a look at this?

> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to 
> false.
> ---
>
> Key: HDFS-9787
> URL: https://issues.apache.org/jira/browse/HDFS-9787
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 3.0.0
>Reporter: Guocui Mi
>Assignee: Guocui Mi
>
> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
> Here is the logic to check if upload FSImage or not.
> In StandbyCheckpointer.java
> boolean sendRequest = isPrimaryCheckPointer || secsSinceLastUpload >= 
> checkpointConf.getQuietPeriod();
> doCheckpoint(sendRequest);
> The sendRequest is always false if isPrimaryCheckPointer is false giving 
> secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
> (checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
> false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Guocui Mi (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Guocui Mi updated HDFS-9787:

Status: Open  (was: Patch Available)

> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to 
> false.
> ---
>
> Key: HDFS-9787
> URL: https://issues.apache.org/jira/browse/HDFS-9787
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 3.0.0
>Reporter: Guocui Mi
>Assignee: Guocui Mi
>
> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
> Here is the logic to check if upload FSImage or not.
> In StandbyCheckpointer.java
> boolean sendRequest = isPrimaryCheckPointer || secsSinceLastUpload >= 
> checkpointConf.getQuietPeriod();
> doCheckpoint(sendRequest);
> The sendRequest is always false if isPrimaryCheckPointer is false giving 
> secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
> (checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
> false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Guocui Mi (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Guocui Mi updated HDFS-9787:

Assignee: Guocui Mi
  Status: Patch Available  (was: Open)

check time since last upload instead of time since last checkpoint.

> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to 
> false.
> ---
>
> Key: HDFS-9787
> URL: https://issues.apache.org/jira/browse/HDFS-9787
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 3.0.0
>Reporter: Guocui Mi
>Assignee: Guocui Mi
>
> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
> Here is the logic to check if upload FSImage or not.
> In StandbyCheckpointer.java
> boolean sendRequest = isPrimaryCheckPointer || secsSinceLastUpload >= 
> checkpointConf.getQuietPeriod();
> doCheckpoint(sendRequest);
> The sendRequest is always false if isPrimaryCheckPointer is false giving 
> secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
> (checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
> false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Inigo Goiri (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9787?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Inigo Goiri updated HDFS-9787:
--
Affects Version/s: (was: 2.6.2)
   3.0.0

> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to 
> false.
> ---
>
> Key: HDFS-9787
> URL: https://issues.apache.org/jira/browse/HDFS-9787
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 3.0.0
>Reporter: Guocui Mi
>
> SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
> Here is the logic to check if upload FSImage or not.
> In StandbyCheckpointer.java
> boolean sendRequest = isPrimaryCheckPointer || secsSinceLastUpload >= 
> checkpointConf.getQuietPeriod();
> doCheckpoint(sendRequest);
> The sendRequest is always false if isPrimaryCheckPointer is false giving 
> secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
> (checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
> false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9711) Integrate CSRF prevention filter in WebHDFS.

2016-02-09 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9711?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140135#comment-15140135
 ] 

Chris Nauroth commented on HDFS-9711:
-

The last round of JDK 8 test failures are unrelated.  I cannot repro.

> Integrate CSRF prevention filter in WebHDFS.
> 
>
> Key: HDFS-9711
> URL: https://issues.apache.org/jira/browse/HDFS-9711
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: datanode, namenode, webhdfs
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-9711.001.patch, HDFS-9711.002.patch, 
> HDFS-9711.003.patch
>
>
> HADOOP-12691 introduced a filter in Hadoop Common to help REST APIs guard 
> against cross-site request forgery attacks.  This issue tracks integration of 
> that filter in WebHDFS.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9637) Add test for HADOOP-12702 and HADOOP-12759

2016-02-09 Thread Karthik Kambatla (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9637?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140132#comment-15140132
 ] 

Karthik Kambatla commented on HDFS-9637:


Reviewing the tests here, following up from the feature added in Common. 

The patch looks pretty good. Nice tests and very happy to see the detailed 
documentation. My only comment would be: use @Before and @After methods to do 
the setup and cleanup, that way we don't have to do try-finally blocks. 

> Add test for HADOOP-12702 and HADOOP-12759
> --
>
> Key: HDFS-9637
> URL: https://issues.apache.org/jira/browse/HDFS-9637
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: test
>Affects Versions: 2.7.1
>Reporter: Daniel Templeton
>Assignee: Daniel Templeton
> Attachments: HDFS-9637.001.patch, HDFS-9637.002.patch, 
> HDFS-9637.003.patch, HDFS-9637.004.patch
>
>
> Per discussion on the dev list, the tests for the new FileSystemSink class 
> should be added to the HDFS project to avoid creating a dependency for the 
> common project on the HDFS project.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-9787) SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer changed to false.

2016-02-09 Thread Guocui Mi (JIRA)
Guocui Mi created HDFS-9787:
---

 Summary: SNNs stop uploading FSImage to ANN once 
isPrimaryCheckPointer changed to false.
 Key: HDFS-9787
 URL: https://issues.apache.org/jira/browse/HDFS-9787
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: ha
Affects Versions: 2.6.2
Reporter: Guocui Mi


SNNs stop uploading FSImage to ANN once isPrimaryCheckPointer become false. 
Here is the logic to check if upload FSImage or not.
In StandbyCheckpointer.java
boolean sendRequest = isPrimaryCheckPointer || secsSinceLastUpload >= 
checkpointConf.getQuietPeriod();
doCheckpoint(sendRequest);

The sendRequest is always false if isPrimaryCheckPointer is false giving 
secsSinceLast (~checkpointPeriod) >= checkpointConf.getQuietPeriod() 
(checkpointPeriod * this.quietMultiplier(default value 1.5)) always returns 
false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9711) Integrate CSRF prevention filter in WebHDFS.

2016-02-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9711?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140056#comment-15140056
 ] 

Hadoop QA commented on HDFS-9711:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 16s 
{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s 
{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 
0s {color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 15s 
{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 
52s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 11s 
{color} | {color:green} trunk passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 42s 
{color} | {color:green} trunk passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 
3s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 28s 
{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
41s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 5m 2s 
{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 18s 
{color} | {color:green} trunk passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 3m 10s 
{color} | {color:green} trunk passed with JDK v1.7.0_95 {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 16s 
{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 
57s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 5m 55s 
{color} | {color:green} the patch passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 5m 55s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 51s 
{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 6m 51s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 
10s {color} | {color:green} root: patch generated 0 new + 107 unchanged - 5 
fixed = 107 total (was 112) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 53s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
47s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 
0s {color} | {color:green} Patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 0s 
{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 6m 
29s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 49s 
{color} | {color:green} the patch passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 3m 40s 
{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 8m 0s 
{color} | {color:green} hadoop-common in the patch passed with JDK v1.8.0_72. 
{color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 0m 55s 
{color} | {color:green} hadoop-hdfs-client in the patch passed with JDK 
v1.8.0_72. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 75m 26s {color} 
| {color:red} hadoop-hdfs in the patch failed with JDK v1.8.0_72. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 7m 21s 
{color} | {color:green} hadoop-common in the patch passed with JDK v1.7.0_95. 
{color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 1m 3s 
{color} | {color:green} hadoop-hdfs-clie

[jira] [Commented] (HDFS-9779) TestReplicationPolicyWithNodeGroup NODE variable picks wrong rack value

2016-02-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9779?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15140030#comment-15140030
 ] 

Hudson commented on HDFS-9779:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #9271 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/9271/])
HDFS-9779 . TestReplicationPolicyWithNodeGroup NODE variable picks wrong 
(uma.gangumalla: rev a7fce9ab415fcc834ca1d91f913157ffd103fa5f)
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestReplicationPolicyWithNodeGroup.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


> TestReplicationPolicyWithNodeGroup NODE variable picks wrong rack value
> ---
>
> Key: HDFS-9779
> URL: https://issues.apache.org/jira/browse/HDFS-9779
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.7.2
>Reporter: Kuhu Shukla
>Assignee: Kuhu Shukla
>Priority: Minor
> Fix For: 2.8.0
>
> Attachments: HDFS-9779.001.patch
>
>
> The DatanodeDescriptor object for {{NODE}} calls the wrong constructor on an 
> already created DatanodeDescriptor object. This overwrites the 
> networkLocation value to "default-rack" instead of expected value of 
> "/d2/r4/n7". I didn't find any other such malformed calls to 
> {{DFSTestUtil.getDatanodeDescriptor}}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7955) Improve naming of classes, methods, and variables related to block replication and recovery

2016-02-09 Thread Uma Maheswara Rao G (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7955?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15139998#comment-15139998
 ] 

Uma Maheswara Rao G commented on HDFS-7955:
---

Thanks [~rakeshr] for the work. I think we can close this now. Do you have any 
other sub tasks planned?

> Improve naming of classes, methods, and variables related to block 
> replication and recovery
> ---
>
> Key: HDFS-7955
> URL: https://issues.apache.org/jira/browse/HDFS-7955
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: erasure-coding
>Reporter: Zhe Zhang
>Assignee: Rakesh R
> Attachments: HDFS-7955-001.patch, HDFS-7955-002.patch, 
> HDFS-7955-003.patch, HDFS-7955-004.patch, HDFS-7955-5.patch
>
>
> Many existing names should be revised to avoid confusion when blocks can be 
> both replicated and erasure coded. This JIRA aims to solicit opinions on 
> making those names more consistent and intuitive.
> # In current HDFS _block recovery_ refers to the process of finalizing the 
> last block of a file, triggered by _lease recovery_. It is different from the 
> intuitive meaning of _recovering a lost block_. To avoid confusion, I can 
> think of 2 options:
> #* Rename this process as _block finalization_ or _block completion_. I 
> prefer this option because this is literally not a recovery.
> #* If we want to keep existing terms unchanged we can name all EC recovery 
> and re-replication logics as _reconstruction_.  
> # As Kai [suggested | 
> https://issues.apache.org/jira/browse/HDFS-7369?focusedCommentId=14361131&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14361131]
>  under HDFS-7369, several replication-based names should be made more generic:
> #* {{UnderReplicatedBlocks}} and {{neededReplications}}. E.g. we can use 
> {{LowRedundancyBlocks}}/{{AtRiskBlocks}}, and 
> {{neededRecovery}}/{{neededReconstruction}}.
> #* {{PendingReplicationBlocks}}
> #* {{ReplicationMonitor}}
> I'm sure the above list is incomplete; discussions and comments are very 
> welcome.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9779) TestReplicationPolicyWithNodeGroup NODE variable picks wrong rack value

2016-02-09 Thread Uma Maheswara Rao G (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9779?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uma Maheswara Rao G updated HDFS-9779:
--
   Resolution: Fixed
 Hadoop Flags: Reviewed
Fix Version/s: 2.8.0
   Status: Resolved  (was: Patch Available)

Thanks for the contribution [~kshukla]. Thanks for the reviews [~shahrs87]

+1, I have just pushed it to trunk, branch-2 and 2.8 

> TestReplicationPolicyWithNodeGroup NODE variable picks wrong rack value
> ---
>
> Key: HDFS-9779
> URL: https://issues.apache.org/jira/browse/HDFS-9779
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.7.2
>Reporter: Kuhu Shukla
>Assignee: Kuhu Shukla
>Priority: Minor
> Fix For: 2.8.0
>
> Attachments: HDFS-9779.001.patch
>
>
> The DatanodeDescriptor object for {{NODE}} calls the wrong constructor on an 
> already created DatanodeDescriptor object. This overwrites the 
> networkLocation value to "default-rack" instead of expected value of 
> "/d2/r4/n7". I didn't find any other such malformed calls to 
> {{DFSTestUtil.getDatanodeDescriptor}}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9775) Erasure Coding : Rename BlockRecoveryWork to BlockReconstructionWork

2016-02-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9775?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15139971#comment-15139971
 ] 

Hudson commented on HDFS-9775:
--

FAILURE: Integrated in Hadoop-trunk-Commit #9270 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/9270/])
HDFS-9775. Erasure Coding : Rename BlockRecoveryWork to (zhz: rev 
a0fb2eff9b71e2e2c0e53262773b34bed82585d4)
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/ReplicationWork.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockReconstructionWork.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockManager.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/ErasureCodingWork.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManagerTestUtil.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeDescriptor.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockRecoveryWork.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java


> Erasure Coding : Rename BlockRecoveryWork to BlockReconstructionWork
> 
>
> Key: HDFS-9775
> URL: https://issues.apache.org/jira/browse/HDFS-9775
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Rakesh R
>Assignee: Rakesh R
> Fix For: 3.0.0
>
> Attachments: HDFS-9775-01.patch
>
>
> This sub-task is to visit the block recovery work and make the logic as 
> reconstruction. ie, rename "recovery" to "reconstruction"



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9702) DiskBalancer : getVolumeMap implementation

2016-02-09 Thread Anu Engineer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9702?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anu Engineer updated HDFS-9702:
---
Attachment: HDFS-9702-HDFS-1312.001.patch

> DiskBalancer : getVolumeMap implementation
> --
>
> Key: HDFS-9702
> URL: https://issues.apache.org/jira/browse/HDFS-9702
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: HDFS-1312
>Reporter: Anu Engineer
>Assignee: Anu Engineer
> Fix For: HDFS-1312
>
> Attachments: HDFS-9702-HDFS-1312.001.patch
>
>
> Add get volume map 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-9786) HttpFS doesn't write the proxyuser information in logfile

2016-02-09 Thread HeeSoo Kim (JIRA)
HeeSoo Kim created HDFS-9786:


 Summary: HttpFS doesn't write the proxyuser information in logfile
 Key: HDFS-9786
 URL: https://issues.apache.org/jira/browse/HDFS-9786
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: HeeSoo Kim
Assignee: HeeSoo Kim


According to the httpfs-log4j.properties, the log pattern indicates that
{code}
log4j.appender.httpfsaudit.layout.ConversionPattern=%d{ISO8601} %5p 
[%X{hostname}][%X{user}:%X{doAs}] %X{op} %m%n
{code}

However, the httpfsaudit doesn't write right information for user and proxyuser 
information. It is better to write ugi on audit log in HttpFS GW.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9702) DiskBalancer : getVolumeMap implementation

2016-02-09 Thread Anu Engineer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9702?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anu Engineer updated HDFS-9702:
---
Attachment: (was: HDFS-9702-HDFS-1312.001.patch)

> DiskBalancer : getVolumeMap implementation
> --
>
> Key: HDFS-9702
> URL: https://issues.apache.org/jira/browse/HDFS-9702
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: HDFS-1312
>Reporter: Anu Engineer
>Assignee: Anu Engineer
> Fix For: HDFS-1312
>
> Attachments: HDFS-9702-HDFS-1312.001.patch
>
>
> Add get volume map 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9775) Erasure Coding : Rename BlockRecoveryWork to BlockReconstructionWork

2016-02-09 Thread Zhe Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9775?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhe Zhang updated HDFS-9775:

   Resolution: Fixed
 Hadoop Flags: Reviewed
Fix Version/s: 3.0.0
   Status: Resolved  (was: Patch Available)

> Erasure Coding : Rename BlockRecoveryWork to BlockReconstructionWork
> 
>
> Key: HDFS-9775
> URL: https://issues.apache.org/jira/browse/HDFS-9775
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Rakesh R
>Assignee: Rakesh R
> Fix For: 3.0.0
>
> Attachments: HDFS-9775-01.patch
>
>
> This sub-task is to visit the block recovery work and make the logic as 
> reconstruction. ie, rename "recovery" to "reconstruction"



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9775) Erasure Coding : Rename BlockRecoveryWork to BlockReconstructionWork

2016-02-09 Thread Zhe Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9775?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15139925#comment-15139925
 ] 

Zhe Zhang commented on HDFS-9775:
-

+1 on the patch. I just committed it to trunk. Thanks Rakesh for the work!

> Erasure Coding : Rename BlockRecoveryWork to BlockReconstructionWork
> 
>
> Key: HDFS-9775
> URL: https://issues.apache.org/jira/browse/HDFS-9775
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Rakesh R
>Assignee: Rakesh R
> Fix For: 3.0.0
>
> Attachments: HDFS-9775-01.patch
>
>
> This sub-task is to visit the block recovery work and make the logic as 
> reconstruction. ie, rename "recovery" to "reconstruction"



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9760) WebHDFS AuthFilter cannot be configured with custom AltKerberos auth handler

2016-02-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9760?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15139896#comment-15139896
 ] 

Hudson commented on HDFS-9760:
--

FAILURE: Integrated in Hadoop-trunk-Commit #9269 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/9269/])
HDFS-9760. WebHDFS AuthFilter cannot be configured with custom (aw: rev 
401ae4ecdb64e1ae2730976f96f7949831305c07)
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestAuthFilter.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/AuthFilter.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNodeHttpServer.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


> WebHDFS AuthFilter cannot be configured with custom AltKerberos auth handler
> 
>
> Key: HDFS-9760
> URL: https://issues.apache.org/jira/browse/HDFS-9760
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Reporter: Ryan Sasson
>Assignee: Ryan Sasson
> Fix For: 2.8.0
>
> Attachments: HDFS-9760.patch
>
>
> Currently the WebHDFS AuthFilter selects its authentication type based on a 
> call to UserGroupInformation.isSecurityEnabled() with only two choices, 
> KerberosAuthentication or PsuedoAuthentication. Thus there is no condition 
> where the WebHDFS server can be configured with a custom AltKerberos 
> authentication handler.
> Additionally, at the time the WebHDFS AuthFilter is initialized the method 
> getAuthFilterParams(conf) is called in NameNodeHttpServer which picks and 
> chooses a certain few configurations with the prefix 
> 'dfs.web.authentication'. The issue is this method strips away the 
> configuration that could set the authentication type AND additional 
> configurations that are specific to the custom auth handler (using the prefix 
> 'dfs.web.authentication.alt-kerberos').
> The consequence of this lack of configurability is that a user that makes 
> authenticated access to the namenode web UI (through a custom authentication 
> handler) will not be able to access the namenode file browser (because it is 
> making ajax calls to WebHDFS that has a different authentication type). 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9760) WebHDFS AuthFilter cannot be configured with custom AltKerberos auth handler

2016-02-09 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9760?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HDFS-9760:
---
   Resolution: Fixed
Fix Version/s: 2.8.0
   Status: Resolved  (was: Patch Available)

+1

committed to 2.8

thanks!

> WebHDFS AuthFilter cannot be configured with custom AltKerberos auth handler
> 
>
> Key: HDFS-9760
> URL: https://issues.apache.org/jira/browse/HDFS-9760
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Reporter: Ryan Sasson
>Assignee: Ryan Sasson
> Fix For: 2.8.0
>
> Attachments: HDFS-9760.patch
>
>
> Currently the WebHDFS AuthFilter selects its authentication type based on a 
> call to UserGroupInformation.isSecurityEnabled() with only two choices, 
> KerberosAuthentication or PsuedoAuthentication. Thus there is no condition 
> where the WebHDFS server can be configured with a custom AltKerberos 
> authentication handler.
> Additionally, at the time the WebHDFS AuthFilter is initialized the method 
> getAuthFilterParams(conf) is called in NameNodeHttpServer which picks and 
> chooses a certain few configurations with the prefix 
> 'dfs.web.authentication'. The issue is this method strips away the 
> configuration that could set the authentication type AND additional 
> configurations that are specific to the custom auth handler (using the prefix 
> 'dfs.web.authentication.alt-kerberos').
> The consequence of this lack of configurability is that a user that makes 
> authenticated access to the namenode web UI (through a custom authentication 
> handler) will not be able to access the namenode file browser (because it is 
> making ajax calls to WebHDFS that has a different authentication type). 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9311) Support optional offload of NameNode HA service health checks to a separate RPC server.

2016-02-09 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9311?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-9311:

Release Note: There is now support for offloading HA health check RPC 
activity to a separate RPC server endpoint running within the NameNode process. 
 This may improve reliability of HA health checks and prevent spurious 
failovers in highly overloaded conditions.  For more details, please refer to 
the hdfs-default.xml documentation for properties 
dfs.namenode.lifeline.rpc-address, dfs.namenode.lifeline.rpc-bind-host and 
dfs.namenode.lifeline.handler.count.

> Support optional offload of NameNode HA service health checks to a separate 
> RPC server.
> ---
>
> Key: HDFS-9311
> URL: https://issues.apache.org/jira/browse/HDFS-9311
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: ha, namenode
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Fix For: 2.8.0
>
> Attachments: HDFS-9311.001.patch, HDFS-9311.002.patch, 
> HDFS-9311.003.patch
>
>
> When a NameNode is overwhelmed with load, it can lead to resource exhaustion 
> of the RPC handler pools (both client-facing and service-facing).  
> Eventually, this blocks the health check RPC issued from ZKFC, which triggers 
> a failover.  Depending on fencing configuration, the former active NameNode 
> may be killed.  In an overloaded situation, the new active NameNode is likely 
> to suffer the same fate, because client load patterns don't change after the 
> failover.  This can degenerate into flapping between the 2 NameNodes without 
> real recovery.  If a NameNode had been killed by fencing, then it would have 
> to transition through safe mode, further delaying time to recovery.
> This issue proposes a separate, optional RPC server at the NameNode for 
> isolating the HA health checks.  These health checks are lightweight 
> operations that do not suffer from contention issues on the namesystem lock 
> or other shared resources.  Isolating the RPC handlers is sufficient to avoid 
> this situation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9683) DiskBalancer : Add cancelPlan implementation

2016-02-09 Thread Anu Engineer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9683?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anu Engineer updated HDFS-9683:
---
Attachment: HDFS-9683-HDFS-1312.001.patch

> DiskBalancer : Add cancelPlan implementation
> 
>
> Key: HDFS-9683
> URL: https://issues.apache.org/jira/browse/HDFS-9683
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: HDFS-1312
>Reporter: Anu Engineer
>Assignee: Anu Engineer
> Fix For: HDFS-1312
>
> Attachments: HDFS-9683-HDFS-1312.001.patch
>
>
> Add datanode side code for Cancel Plan



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9683) DiskBalancer : Add cancelPlan implementation

2016-02-09 Thread Anu Engineer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9683?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anu Engineer updated HDFS-9683:
---
Attachment: (was: HDFS-9683-HDFS-1312.001.patch)

> DiskBalancer : Add cancelPlan implementation
> 
>
> Key: HDFS-9683
> URL: https://issues.apache.org/jira/browse/HDFS-9683
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: HDFS-1312
>Reporter: Anu Engineer
>Assignee: Anu Engineer
> Fix For: HDFS-1312
>
>
> Add datanode side code for Cancel Plan



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9711) Integrate CSRF prevention filter in WebHDFS.

2016-02-09 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-9711:

Attachment: HDFS-9711.003.patch

Here is patch v003.  I updated this in response HADOOP-12758, where [~lmccay] 
added support to the filter for checking the User-Agent.  With consideration of 
Larry's recent comments here, I have kept the configurability of the header 
name.

> Integrate CSRF prevention filter in WebHDFS.
> 
>
> Key: HDFS-9711
> URL: https://issues.apache.org/jira/browse/HDFS-9711
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: datanode, namenode, webhdfs
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-9711.001.patch, HDFS-9711.002.patch, 
> HDFS-9711.003.patch
>
>
> HADOOP-12691 introduced a filter in Hadoop Common to help REST APIs guard 
> against cross-site request forgery attacks.  This issue tracks integration of 
> that filter in WebHDFS.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9785) Remove unused TrashPolicy#getInstance and initiate code

2016-02-09 Thread Zhe Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9785?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15139667#comment-15139667
 ] 

Zhe Zhang commented on HDFS-9785:
-

Actually {{TrashPolicy}} could be directly used by applications. Marking as 
incompatible and we should consider it for 3.0.

> Remove unused TrashPolicy#getInstance and initiate code
> ---
>
> Key: HDFS-9785
> URL: https://issues.apache.org/jira/browse/HDFS-9785
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Zhe Zhang
>Priority: Minor
>  Labels: incompatibleChange
>
> A follow-on from HDFS-8831: now the {{getInstance}} and {{initiate}} APIs 
> with Path is not used anymore.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9785) Remove unused TrashPolicy#getInstance and initiate code

2016-02-09 Thread Zhe Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9785?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhe Zhang updated HDFS-9785:

Labels: incompatibleChange  (was: )

> Remove unused TrashPolicy#getInstance and initiate code
> ---
>
> Key: HDFS-9785
> URL: https://issues.apache.org/jira/browse/HDFS-9785
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Zhe Zhang
>Priority: Minor
>  Labels: incompatibleChange
>
> A follow-on from HDFS-8831: now the {{getInstance}} and {{initiate}} APIs 
> with Path is not used anymore.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9785) Remove unused TrashPolicy#getInstance and initiate code

2016-02-09 Thread Zhe Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9785?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhe Zhang updated HDFS-9785:

Target Version/s: 3.0.0  (was: 2.8.0)

> Remove unused TrashPolicy#getInstance and initiate code
> ---
>
> Key: HDFS-9785
> URL: https://issues.apache.org/jira/browse/HDFS-9785
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Zhe Zhang
>Priority: Minor
>  Labels: incompatibleChange
>
> A follow-on from HDFS-8831: now the {{getInstance}} and {{initiate}} APIs 
> with Path is not used anymore.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9785) Remove unused TrashPolicy#getInstance and initiate code

2016-02-09 Thread Zhe Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9785?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhe Zhang updated HDFS-9785:

Description: A follow-on from HDFS-8831: now the {{getInstance}} and 
{{initiate}} APIs with Path is not used anymore.  (was: A follow-on from 
HDFS-8831: now the {{getInstance}} API with Path is not used anymore.)

> Remove unused TrashPolicy#getInstance and initiate code
> ---
>
> Key: HDFS-9785
> URL: https://issues.apache.org/jira/browse/HDFS-9785
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Zhe Zhang
>Priority: Minor
>
> A follow-on from HDFS-8831: now the {{getInstance}} and {{initiate}} APIs 
> with Path is not used anymore.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9785) Remove unused TrashPolicy#getInstance and initiate code

2016-02-09 Thread Zhe Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9785?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhe Zhang updated HDFS-9785:

Summary: Remove unused TrashPolicy#getInstance and initiate code  (was: 
Remove unused TrashPolicy#getInstance code)

> Remove unused TrashPolicy#getInstance and initiate code
> ---
>
> Key: HDFS-9785
> URL: https://issues.apache.org/jira/browse/HDFS-9785
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Zhe Zhang
>Priority: Minor
>
> A follow-on from HDFS-8831: now the {{getInstance}} API with Path is not used 
> anymore.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-9785) Remove unused TrashPolicy#getInstance code

2016-02-09 Thread Zhe Zhang (JIRA)
Zhe Zhang created HDFS-9785:
---

 Summary: Remove unused TrashPolicy#getInstance code
 Key: HDFS-9785
 URL: https://issues.apache.org/jira/browse/HDFS-9785
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Zhe Zhang
Priority: Minor


A follow-on from HDFS-8831: now the {{getInstance}} API with Path is not used 
anymore.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9784) Example usage is not correct in Transparent Encryption document

2016-02-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9784?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15139402#comment-15139402
 ] 

Hudson commented on HDFS-9784:
--

FAILURE: Integrated in Hadoop-trunk-Commit #9268 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/9268/])
HDFS-9784. Example usage is not correct in Transparent Encryption (aajisaka: 
rev 60d2011b7c0fe55b8bc44a141660d3e2df37a68d)
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/TransparentEncryption.md


> Example usage is not correct in Transparent Encryption document
> ---
>
> Key: HDFS-9784
> URL: https://issues.apache.org/jira/browse/HDFS-9784
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 3.0.0, 2.7.2
>Reporter: Takashi Ohnishi
>Assignee: Takashi Ohnishi
> Fix For: 2.8.0, 2.7.3
>
> Attachments: HADOOP-12779.1.patch
>
>
> It says
> {code}
> # As the normal user, create a new encryption key
> hadoop key create myKey
> {code}
> But, this actually fails with the below error.
> {code}
> $ hadoop key create myKey
> java.lang.IllegalArgumentException: Uppercase key names are unsupported: myKey
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
> at 
> org.apache.hadoop.util.HttpExceptionUtils.validateResponse(HttpExceptionUtils.java:157)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.call(KMSClientProvider.java:546)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.call(KMSClientProvider.java:504)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.createKeyInternal(KMSClientProvider.java:677)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.createKey(KMSClientProvider.java:685)
> at 
> org.apache.hadoop.crypto.key.KeyShell$CreateCommand.execute(KeyShell.java:483)
> at org.apache.hadoop.crypto.key.KeyShell.run(KeyShell.java:79)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> at org.apache.hadoop.crypto.key.KeyShell.main(KeyShell.java:515)
> {code}
> Though I'm not sure why it is so, I think the document should be fixed to use 
> only lowercase in key names.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9752) Permanent write failures may happen to slow writers during datanode rolling upgrades

2016-02-09 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9752?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-9752:

   Resolution: Fixed
 Hadoop Flags: Reviewed
Fix Version/s: 2.6.5
   2.7.3
   2.8.0
   Status: Resolved  (was: Patch Available)

Committed the branch-2.7 patch. Thanks [~walter.k.su] for the contribution.

> Permanent write failures may happen to slow writers during datanode rolling 
> upgrades
> 
>
> Key: HDFS-9752
> URL: https://issues.apache.org/jira/browse/HDFS-9752
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Kihwal Lee
>Assignee: Walter Su
>Priority: Critical
> Fix For: 2.8.0, 2.7.3, 2.6.5
>
> Attachments: HDFS-9752-branch-2.6.03.patch, 
> HDFS-9752-branch-2.7.03.patch, HDFS-9752.01.patch, HDFS-9752.02.patch, 
> HDFS-9752.03.patch, HdfsWriter.java
>
>
> When datanodes are being upgraded, an out-of-band ack is sent upstream and 
> the client does a pipeline recovery. The client may hit this multiple times 
> as more nodes get upgraded.  This normally does not cause any issue, but if 
> the client is holding the stream open without writing any data during this 
> time, a permanent write failure can occur.
> This is because there is a limit of 5 recovery trials for the same packet, 
> which is tracked by "last acked sequence number". Since the empty heartbeat 
> packets for an idle output stream does not increment the sequence number, the 
> write will fail after it seeing 5 pipeline breakages by datanode upgrades.
> This check/limit was added to avoid spinning until running out of nodes in 
> the cluster due to a corruption or any other irrecoverable conditions.  The 
> datanode upgrade-restart  should be excluded from the count.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9752) Permanent write failures may happen to slow writers during datanode rolling upgrades

2016-02-09 Thread Akira AJISAKA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9752?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15139393#comment-15139393
 ] 

Akira AJISAKA commented on HDFS-9752:
-

+1 for the branch-2.7 patch. I'll commit it shortly.

> Permanent write failures may happen to slow writers during datanode rolling 
> upgrades
> 
>
> Key: HDFS-9752
> URL: https://issues.apache.org/jira/browse/HDFS-9752
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Kihwal Lee
>Assignee: Walter Su
>Priority: Critical
> Attachments: HDFS-9752-branch-2.6.03.patch, 
> HDFS-9752-branch-2.7.03.patch, HDFS-9752.01.patch, HDFS-9752.02.patch, 
> HDFS-9752.03.patch, HdfsWriter.java
>
>
> When datanodes are being upgraded, an out-of-band ack is sent upstream and 
> the client does a pipeline recovery. The client may hit this multiple times 
> as more nodes get upgraded.  This normally does not cause any issue, but if 
> the client is holding the stream open without writing any data during this 
> time, a permanent write failure can occur.
> This is because there is a limit of 5 recovery trials for the same packet, 
> which is tracked by "last acked sequence number". Since the empty heartbeat 
> packets for an idle output stream does not increment the sequence number, the 
> write will fail after it seeing 5 pipeline breakages by datanode upgrades.
> This check/limit was added to avoid spinning until running out of nodes in 
> the cluster due to a corruption or any other irrecoverable conditions.  The 
> datanode upgrade-restart  should be excluded from the count.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HDFS-9784) Example usage is not correct in Transparent Encryption document

2016-02-09 Thread Akira AJISAKA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9784?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15139378#comment-15139378
 ] 

Akira AJISAKA edited comment on HDFS-9784 at 2/9/16 6:43 PM:
-

Filed HADOOP-12786 for documenting "hadoop key" command usage.


was (Author: ajisakaa):
Filed HADOOP-12876 for documenting "hadoop key" command usage.

> Example usage is not correct in Transparent Encryption document
> ---
>
> Key: HDFS-9784
> URL: https://issues.apache.org/jira/browse/HDFS-9784
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 3.0.0, 2.7.2
>Reporter: Takashi Ohnishi
>Assignee: Takashi Ohnishi
> Fix For: 2.8.0, 2.7.3
>
> Attachments: HADOOP-12779.1.patch
>
>
> It says
> {code}
> # As the normal user, create a new encryption key
> hadoop key create myKey
> {code}
> But, this actually fails with the below error.
> {code}
> $ hadoop key create myKey
> java.lang.IllegalArgumentException: Uppercase key names are unsupported: myKey
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
> at 
> org.apache.hadoop.util.HttpExceptionUtils.validateResponse(HttpExceptionUtils.java:157)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.call(KMSClientProvider.java:546)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.call(KMSClientProvider.java:504)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.createKeyInternal(KMSClientProvider.java:677)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.createKey(KMSClientProvider.java:685)
> at 
> org.apache.hadoop.crypto.key.KeyShell$CreateCommand.execute(KeyShell.java:483)
> at org.apache.hadoop.crypto.key.KeyShell.run(KeyShell.java:79)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> at org.apache.hadoop.crypto.key.KeyShell.main(KeyShell.java:515)
> {code}
> Though I'm not sure why it is so, I think the document should be fixed to use 
> only lowercase in key names.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9784) Example usage is not correct in Transparent Encryption document

2016-02-09 Thread Akira AJISAKA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9784?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15139378#comment-15139378
 ] 

Akira AJISAKA commented on HDFS-9784:
-

Filed HADOOP-12876 for documenting "hadoop key" command usage.

> Example usage is not correct in Transparent Encryption document
> ---
>
> Key: HDFS-9784
> URL: https://issues.apache.org/jira/browse/HDFS-9784
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 3.0.0, 2.7.2
>Reporter: Takashi Ohnishi
>Assignee: Takashi Ohnishi
> Fix For: 2.8.0, 2.7.3
>
> Attachments: HADOOP-12779.1.patch
>
>
> It says
> {code}
> # As the normal user, create a new encryption key
> hadoop key create myKey
> {code}
> But, this actually fails with the below error.
> {code}
> $ hadoop key create myKey
> java.lang.IllegalArgumentException: Uppercase key names are unsupported: myKey
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
> at 
> org.apache.hadoop.util.HttpExceptionUtils.validateResponse(HttpExceptionUtils.java:157)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.call(KMSClientProvider.java:546)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.call(KMSClientProvider.java:504)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.createKeyInternal(KMSClientProvider.java:677)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.createKey(KMSClientProvider.java:685)
> at 
> org.apache.hadoop.crypto.key.KeyShell$CreateCommand.execute(KeyShell.java:483)
> at org.apache.hadoop.crypto.key.KeyShell.run(KeyShell.java:79)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> at org.apache.hadoop.crypto.key.KeyShell.main(KeyShell.java:515)
> {code}
> Though I'm not sure why it is so, I think the document should be fixed to use 
> only lowercase in key names.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9784) Example usage is not correct in Transparent Encryption document

2016-02-09 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9784?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-9784:

Assignee: Takashi Ohnishi

> Example usage is not correct in Transparent Encryption document
> ---
>
> Key: HDFS-9784
> URL: https://issues.apache.org/jira/browse/HDFS-9784
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 3.0.0, 2.7.2
>Reporter: Takashi Ohnishi
>Assignee: Takashi Ohnishi
> Fix For: 2.8.0, 2.7.3
>
> Attachments: HADOOP-12779.1.patch
>
>
> It says
> {code}
> # As the normal user, create a new encryption key
> hadoop key create myKey
> {code}
> But, this actually fails with the below error.
> {code}
> $ hadoop key create myKey
> java.lang.IllegalArgumentException: Uppercase key names are unsupported: myKey
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
> at 
> org.apache.hadoop.util.HttpExceptionUtils.validateResponse(HttpExceptionUtils.java:157)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.call(KMSClientProvider.java:546)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.call(KMSClientProvider.java:504)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.createKeyInternal(KMSClientProvider.java:677)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.createKey(KMSClientProvider.java:685)
> at 
> org.apache.hadoop.crypto.key.KeyShell$CreateCommand.execute(KeyShell.java:483)
> at org.apache.hadoop.crypto.key.KeyShell.run(KeyShell.java:79)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> at org.apache.hadoop.crypto.key.KeyShell.main(KeyShell.java:515)
> {code}
> Though I'm not sure why it is so, I think the document should be fixed to use 
> only lowercase in key names.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9784) Example usage is not correct in Transparent Encryption document

2016-02-09 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9784?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-9784:

   Resolution: Fixed
Fix Version/s: 2.7.3
   2.8.0
   Status: Resolved  (was: Patch Available)

Committed this to trunk, branch-2, branch-2.8, and branch-2.7. Thanks 
[~bwtakacy] for the contribution.

> Example usage is not correct in Transparent Encryption document
> ---
>
> Key: HDFS-9784
> URL: https://issues.apache.org/jira/browse/HDFS-9784
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 3.0.0, 2.7.2
>Reporter: Takashi Ohnishi
> Fix For: 2.8.0, 2.7.3
>
> Attachments: HADOOP-12779.1.patch
>
>
> It says
> {code}
> # As the normal user, create a new encryption key
> hadoop key create myKey
> {code}
> But, this actually fails with the below error.
> {code}
> $ hadoop key create myKey
> java.lang.IllegalArgumentException: Uppercase key names are unsupported: myKey
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
> at 
> org.apache.hadoop.util.HttpExceptionUtils.validateResponse(HttpExceptionUtils.java:157)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.call(KMSClientProvider.java:546)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.call(KMSClientProvider.java:504)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.createKeyInternal(KMSClientProvider.java:677)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.createKey(KMSClientProvider.java:685)
> at 
> org.apache.hadoop.crypto.key.KeyShell$CreateCommand.execute(KeyShell.java:483)
> at org.apache.hadoop.crypto.key.KeyShell.run(KeyShell.java:79)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> at org.apache.hadoop.crypto.key.KeyShell.main(KeyShell.java:515)
> {code}
> Though I'm not sure why it is so, I think the document should be fixed to use 
> only lowercase in key names.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9784) Example usage is not correct in Transparent Encryption document

2016-02-09 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9784?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-9784:

Hadoop Flags: Reviewed
 Summary: Example usage is not correct in Transparent Encryption 
document  (was: Example usage is not correct in "Transparent Encryption in 
HDFS" document.)

> Example usage is not correct in Transparent Encryption document
> ---
>
> Key: HDFS-9784
> URL: https://issues.apache.org/jira/browse/HDFS-9784
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 3.0.0, 2.7.2
>Reporter: Takashi Ohnishi
> Attachments: HADOOP-12779.1.patch
>
>
> It says
> {code}
> # As the normal user, create a new encryption key
> hadoop key create myKey
> {code}
> But, this actually fails with the below error.
> {code}
> $ hadoop key create myKey
> java.lang.IllegalArgumentException: Uppercase key names are unsupported: myKey
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
> at 
> org.apache.hadoop.util.HttpExceptionUtils.validateResponse(HttpExceptionUtils.java:157)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.call(KMSClientProvider.java:546)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.call(KMSClientProvider.java:504)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.createKeyInternal(KMSClientProvider.java:677)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.createKey(KMSClientProvider.java:685)
> at 
> org.apache.hadoop.crypto.key.KeyShell$CreateCommand.execute(KeyShell.java:483)
> at org.apache.hadoop.crypto.key.KeyShell.run(KeyShell.java:79)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> at org.apache.hadoop.crypto.key.KeyShell.main(KeyShell.java:515)
> {code}
> Though I'm not sure why it is so, I think the document should be fixed to use 
> only lowercase in key names.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Moved] (HDFS-9784) Example usage is not correct in "Transparent Encryption in HDFS" document.

2016-02-09 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9784?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA moved HADOOP-12779 to HDFS-9784:
--

Affects Version/s: (was: 2.7.2)
   (was: 3.0.0)
   3.0.0
   2.7.2
  Component/s: (was: documentation)
   documentation
  Key: HDFS-9784  (was: HADOOP-12779)
  Project: Hadoop HDFS  (was: Hadoop Common)

> Example usage is not correct in "Transparent Encryption in HDFS" document.
> --
>
> Key: HDFS-9784
> URL: https://issues.apache.org/jira/browse/HDFS-9784
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 2.7.2, 3.0.0
>Reporter: Takashi Ohnishi
> Attachments: HADOOP-12779.1.patch
>
>
> It says
> {code}
> # As the normal user, create a new encryption key
> hadoop key create myKey
> {code}
> But, this actually fails with the below error.
> {code}
> $ hadoop key create myKey
> java.lang.IllegalArgumentException: Uppercase key names are unsupported: myKey
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
> at 
> org.apache.hadoop.util.HttpExceptionUtils.validateResponse(HttpExceptionUtils.java:157)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.call(KMSClientProvider.java:546)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.call(KMSClientProvider.java:504)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.createKeyInternal(KMSClientProvider.java:677)
> at 
> org.apache.hadoop.crypto.key.kms.KMSClientProvider.createKey(KMSClientProvider.java:685)
> at 
> org.apache.hadoop.crypto.key.KeyShell$CreateCommand.execute(KeyShell.java:483)
> at org.apache.hadoop.crypto.key.KeyShell.run(KeyShell.java:79)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> at org.apache.hadoop.crypto.key.KeyShell.main(KeyShell.java:515)
> {code}
> Though I'm not sure why it is so, I think the document should be fixed to use 
> only lowercase in key names.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8959) Provide an iterator-based API for listing all the snapshottable directories

2016-02-09 Thread Rakesh R (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8959?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15139311#comment-15139311
 ] 

Rakesh R commented on HDFS-8959:


I've rebased the patch in the latest trunk code. I'm planning to use this API 
in the {{lsSnapshottableDir}} command. Reference:  
[comment|https://issues.apache.org/jira/browse/HDFS-8643?focusedCommentId=14658800&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14658800].
 [~jingzhao], [~umamaheswararao]  could you please check the patch when you get 
a chance. Thanks!

> Provide an iterator-based API for listing all the snapshottable directories
> ---
>
> Key: HDFS-8959
> URL: https://issues.apache.org/jira/browse/HDFS-8959
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: hdfs-client
>Reporter: Rakesh R
>Assignee: Rakesh R
> Attachments: HDFS-8959-00.patch, HDFS-8959-01.patch, 
> HDFS-8959-02.patch
>
>
> Presently {{DistributedFileSystem#getSnapshottableDirListing()}} is sending 
> all the {{SnapshottableDirectoryStatus[]}} array to the clients. Now the 
> client should have enough space to hold it in memory. There could be chance 
> that the client JVMs running out of memory because of this. Also, some time 
> back there was a 
> [comment|https://issues.apache.org/jira/browse/HDFS-8643?focusedCommentId=14658800&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14658800]
>  about RPC packet limitation and a large number of snapshot list can again 
> cause issues.
> I believe iterator based {{DistributedFileSystem#listSnapshottableDirs()}} 
> API would be a good addition!



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9752) Permanent write failures may happen to slow writers during datanode rolling upgrades

2016-02-09 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9752?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15139280#comment-15139280
 ] 

Arpit Agarwal commented on HDFS-9752:
-

+1 for the branch-2.6 patch. I committed it to branch-2.6. Thanks for taking 
care of this backport [~walter.k.su].

> Permanent write failures may happen to slow writers during datanode rolling 
> upgrades
> 
>
> Key: HDFS-9752
> URL: https://issues.apache.org/jira/browse/HDFS-9752
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Kihwal Lee
>Assignee: Walter Su
>Priority: Critical
> Attachments: HDFS-9752-branch-2.6.03.patch, 
> HDFS-9752-branch-2.7.03.patch, HDFS-9752.01.patch, HDFS-9752.02.patch, 
> HDFS-9752.03.patch, HdfsWriter.java
>
>
> When datanodes are being upgraded, an out-of-band ack is sent upstream and 
> the client does a pipeline recovery. The client may hit this multiple times 
> as more nodes get upgraded.  This normally does not cause any issue, but if 
> the client is holding the stream open without writing any data during this 
> time, a permanent write failure can occur.
> This is because there is a limit of 5 recovery trials for the same packet, 
> which is tracked by "last acked sequence number". Since the empty heartbeat 
> packets for an idle output stream does not increment the sequence number, the 
> write will fail after it seeing 5 pipeline breakages by datanode upgrades.
> This check/limit was added to avoid spinning until running out of nodes in 
> the cluster due to a corruption or any other irrecoverable conditions.  The 
> datanode upgrade-restart  should be excluded from the count.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HDFS-8701) WebHDFS write commands fail when running multiple DataNodes on one machine

2016-02-09 Thread Anthony Hsu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8701?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anthony Hsu resolved HDFS-8701.
---
Resolution: Cannot Reproduce

> WebHDFS write commands fail when running multiple DataNodes on one machine
> --
>
> Key: HDFS-8701
> URL: https://issues.apache.org/jira/browse/HDFS-8701
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Anthony Hsu
>
> For testing purposes, we are running multiple DataNodes per machine. {{hadoop 
> fs}} commands work fine when using the {{hdfs://}} protocol, but when using 
> {{webhdfs://}}, any command that writes to HDFS (e.g.: {{-put}} or 
> {{-touchz}}) fails:
> {code}
> $ hadoop fs -put test.txt webhdfs://:/user/foo
> put: -dn-4
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9781) FsDatasetImpl#getBlockReports can occasionally throw NullPointerException

2016-02-09 Thread Wei-Chiu Chuang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9781?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wei-Chiu Chuang updated HDFS-9781:
--
Description: 
FsDatasetImpl#getBlockReports occasionally throws NPE. The NPE caused 
TestFsDatasetImpl#testRemoveVolumeBeingWritten to time out, because the test 
waits for the call to FsDatasetImpl#getBlockReports to complete without 
exceptions.

Additionally, the test should be updated to identify an expected exception, 
using {{GenericTestUtils.assertExceptionContains()}}

{noformat}
Exception in thread "Thread-20" java.lang.NullPointerException
at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.getBlockReports(FsDatasetImpl.java:1709)
at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestFsDatasetImpl$1BlockReportThread.run(TestFsDatasetImpl.java:587)
2016-02-08 15:47:30,379 [Thread-21] WARN  impl.TestFsDatasetImpl 
(TestFsDatasetImpl.java:run(606)) - Exception caught. This should not affect 
the test
java.io.IOException: Failed to move meta file for ReplicaBeingWritten, blk_0_0, 
RBW
  getNumBytes() = 0
  getBytesOnDisk()  = 0
  getVisibleLength()= 0
  getVolume()   = 
/home/weichiu/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/Nmi6rYndvr/data0/current
  getBlockFile()= 
/home/weichiu/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/Nmi6rYndvr/data0/current/bpid-0/current/rbw/blk_0
  bytesAcked=0
  bytesOnDisk=0 from 
/home/weichiu/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/Nmi6rYndvr/data0/current/bpid-0/current/rbw/blk_0_0.meta
 to 
/home/weichiu/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/Nmi6rYndvr/data0/current/bpid-0/current/finalized/subdir0/subdir0/blk_0_0.meta
at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.moveBlockFiles(FsDatasetImpl.java:857)
at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.BlockPoolSlice.addFinalizedBlock(BlockPoolSlice.java:295)
at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsVolumeImpl.addFinalizedBlock(FsVolumeImpl.java:819)
at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.finalizeReplica(FsDatasetImpl.java:1620)
at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.finalizeBlock(FsDatasetImpl.java:1601)
at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestFsDatasetImpl$1ResponderThread.run(TestFsDatasetImpl.java:603)
Caused by: java.io.IOException: 
renameTo(src=/home/weichiu/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/Nmi6rYndvr/data0/current/bpid-0/current/rbw/blk_0_0.meta,
 
dst=/home/weichiu/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/Nmi6rYndvr/data0/current/bpid-0/current/finalized/subdir0/subdir0/blk_0_0.meta)
 failed.
at org.apache.hadoop.io.nativeio.NativeIO.renameTo(NativeIO.java:873)
at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.moveBlockFiles(FsDatasetImpl.java:855)
... 5 more
2016-02-08 15:47:34,381 [Thread-19] INFO  impl.FsDatasetImpl 
(FsVolumeList.java:waitVolumeRemoved(287)) - Volume reference is released.
2016-02-08 15:47:34,384 [Thread-19] INFO  impl.TestFsDatasetImpl 
(TestFsDatasetImpl.java:testRemoveVolumeBeingWritten(622)) - Volumes removed
{noformat}

  was:
FsDatasetImpl#getBlockReports occasionally throws NPE. The NPE caused 
TestFsDatasetImpl#testRemoveVolumeBeingWritten to fail, because the test waits 
for the call to FsDatasetImpl#getBlockReports to complete without exceptions.

Additionally, the test should be updated to identify an expected exception, 
using {{GenericTestUtils.assertExceptionContains()}}

{noformat}
Exception in thread "Thread-20" java.lang.NullPointerException
at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.getBlockReports(FsDatasetImpl.java:1709)
at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestFsDatasetImpl$1BlockReportThread.run(TestFsDatasetImpl.java:587)
2016-02-08 15:47:30,379 [Thread-21] WARN  impl.TestFsDatasetImpl 
(TestFsDatasetImpl.java:run(606)) - Exception caught. This should not affect 
the test
java.io.IOException: Failed to move meta file for ReplicaBeingWritten, blk_0_0, 
RBW
  getNumBytes() = 0
  getBytesOnDisk()  = 0
  getVisibleLength()= 0
  getVolume()   = 
/home/weichiu/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/Nmi6rYndvr/data0/current
  getBlockFile()= 
/home/weichiu/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/Nmi6rYndvr/data0/current/bpid-0/current/rbw/blk_0
  bytesAcked=0
  bytesOnDisk=0 from 
/home/weichiu/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/Nmi6rYndvr/data0/current/bpid-0/current/rbw/blk_0_0.meta
 to 
/home/weichiu/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/Nmi6rYndvr/data0/current/bpid-0/current/finalized/subdir0/subdir0/blk_0_0.meta
at 
org.apache.hadoop

[jira] [Commented] (HDFS-9781) FsDatasetImpl#getBlockReports can occasionally throw NullPointerException

2016-02-09 Thread Wei-Chiu Chuang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9781?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15139140#comment-15139140
 ] 

Wei-Chiu Chuang commented on HDFS-9781:
---

This is related to HDFS-9701, which created 
{TestFsDatasetImpl#testRemoveVolumeBeingWritten}. 

The NPE is caused by accessing a volume after it is removed in the test. So the 
test case should be improved, and FsDatasetImpl should also be improved to 
handle a potential NPE.

> FsDatasetImpl#getBlockReports can occasionally throw NullPointerException
> -
>
> Key: HDFS-9781
> URL: https://issues.apache.org/jira/browse/HDFS-9781
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 3.0.0
> Environment: Jenkins
>Reporter: Wei-Chiu Chuang
>Assignee: Wei-Chiu Chuang
>
> FsDatasetImpl#getBlockReports occasionally throws NPE. The NPE caused 
> TestFsDatasetImpl#testRemoveVolumeBeingWritten to fail, because the test 
> waits for the call to FsDatasetImpl#getBlockReports to complete without 
> exceptions.
> Additionally, the test should be updated to identify an expected exception, 
> using {{GenericTestUtils.assertExceptionContains()}}
> {noformat}
> Exception in thread "Thread-20" java.lang.NullPointerException
> at 
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.getBlockReports(FsDatasetImpl.java:1709)
> at 
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestFsDatasetImpl$1BlockReportThread.run(TestFsDatasetImpl.java:587)
> 2016-02-08 15:47:30,379 [Thread-21] WARN  impl.TestFsDatasetImpl 
> (TestFsDatasetImpl.java:run(606)) - Exception caught. This should not affect 
> the test
> java.io.IOException: Failed to move meta file for ReplicaBeingWritten, 
> blk_0_0, RBW
>   getNumBytes() = 0
>   getBytesOnDisk()  = 0
>   getVisibleLength()= 0
>   getVolume()   = 
> /home/weichiu/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/Nmi6rYndvr/data0/current
>   getBlockFile()= 
> /home/weichiu/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/Nmi6rYndvr/data0/current/bpid-0/current/rbw/blk_0
>   bytesAcked=0
>   bytesOnDisk=0 from 
> /home/weichiu/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/Nmi6rYndvr/data0/current/bpid-0/current/rbw/blk_0_0.meta
>  to 
> /home/weichiu/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/Nmi6rYndvr/data0/current/bpid-0/current/finalized/subdir0/subdir0/blk_0_0.meta
> at 
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.moveBlockFiles(FsDatasetImpl.java:857)
> at 
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.BlockPoolSlice.addFinalizedBlock(BlockPoolSlice.java:295)
> at 
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsVolumeImpl.addFinalizedBlock(FsVolumeImpl.java:819)
> at 
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.finalizeReplica(FsDatasetImpl.java:1620)
> at 
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.finalizeBlock(FsDatasetImpl.java:1601)
> at 
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestFsDatasetImpl$1ResponderThread.run(TestFsDatasetImpl.java:603)
> Caused by: java.io.IOException: 
> renameTo(src=/home/weichiu/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/Nmi6rYndvr/data0/current/bpid-0/current/rbw/blk_0_0.meta,
>  
> dst=/home/weichiu/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/Nmi6rYndvr/data0/current/bpid-0/current/finalized/subdir0/subdir0/blk_0_0.meta)
>  failed.
> at org.apache.hadoop.io.nativeio.NativeIO.renameTo(NativeIO.java:873)
> at 
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.moveBlockFiles(FsDatasetImpl.java:855)
> ... 5 more
> 2016-02-08 15:47:34,381 [Thread-19] INFO  impl.FsDatasetImpl 
> (FsVolumeList.java:waitVolumeRemoved(287)) - Volume reference is released.
> 2016-02-08 15:47:34,384 [Thread-19] INFO  impl.TestFsDatasetImpl 
> (TestFsDatasetImpl.java:testRemoveVolumeBeingWritten(622)) - Volumes removed
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HDFS-9783) DataNode deadlocks after running hdfs mover

2016-02-09 Thread Vinayakumar B (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vinayakumar B resolved HDFS-9783.
-
Resolution: Duplicate

Resolving as duplicate.
Feel free to re-open if fix in HDFS-9661 doesnt work for you.

> DataNode deadlocks after running hdfs mover
> ---
>
> Key: HDFS-9783
> URL: https://issues.apache.org/jira/browse/HDFS-9783
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 2.6.0
>Reporter: David Watzke
> Attachments: datanode.stack.txt
>
>
> We're running CDH 5.4.4 (with Hadoop 2.6.0). Recently we've added 800T of 
> ARCHIVE storage, marked some data (16T * 2 repl. factor) as "COLD" and ran 
> the hdfs mover to enforce the storage policy.
> After a few minutes, the mover hangs because one or more datanodes hang as 
> well. Please check out the deadlock revealed by jstack. Also, here's what 
> appeared in DN log:
> 2016-02-09 15:58:15,676 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
> Not able to copy block 1157230686 to /5.45.60.142:40144 because threads quota 
> is exceeded.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9783) DataNode deadlocks after running hdfs mover

2016-02-09 Thread Vinayakumar B (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9783?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15139135#comment-15139135
 ] 

Vinayakumar B commented on HDFS-9783:
-

This issue has been already fixed in HDFS-9661 which is available in branch-2.7 
( 2.7.3)
And the stacktrace shown is not part of hadoop-2.6 code. 
{{FsDataSetImpl#moveBlockAcrossStorage(..)}} not present in 2.6 code.

So IMO this would be a duplicate.

> DataNode deadlocks after running hdfs mover
> ---
>
> Key: HDFS-9783
> URL: https://issues.apache.org/jira/browse/HDFS-9783
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 2.6.0
>Reporter: David Watzke
> Attachments: datanode.stack.txt
>
>
> We're running CDH 5.4.4 (with Hadoop 2.6.0). Recently we've added 800T of 
> ARCHIVE storage, marked some data (16T * 2 repl. factor) as "COLD" and ran 
> the hdfs mover to enforce the storage policy.
> After a few minutes, the mover hangs because one or more datanodes hang as 
> well. Please check out the deadlock revealed by jstack. Also, here's what 
> appeared in DN log:
> 2016-02-09 15:58:15,676 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
> Not able to copy block 1157230686 to /5.45.60.142:40144 because threads quota 
> is exceeded.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9637) Add test for HADOOP-12702 and HADOOP-12759

2016-02-09 Thread Daniel Templeton (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9637?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15139120#comment-15139120
 ] 

Daniel Templeton commented on HDFS-9637:


Test failures are unrelated.

> Add test for HADOOP-12702 and HADOOP-12759
> --
>
> Key: HDFS-9637
> URL: https://issues.apache.org/jira/browse/HDFS-9637
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: test
>Affects Versions: 2.7.1
>Reporter: Daniel Templeton
>Assignee: Daniel Templeton
> Attachments: HDFS-9637.001.patch, HDFS-9637.002.patch, 
> HDFS-9637.003.patch, HDFS-9637.004.patch
>
>
> Per discussion on the dev list, the tests for the new FileSystemSink class 
> should be added to the HDFS project to avoid creating a dependency for the 
> common project on the HDFS project.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9783) DataNode deadlocks after running hdfs mover

2016-02-09 Thread David Watzke (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Watzke updated HDFS-9783:
---
Description: 
We're running CDH 5.4.4 (with Hadoop 2.6.0). Recently we've added 800T of 
ARCHIVE storage, marked some data (16T * 2 repl. factor) as "COLD" and ran the 
hdfs mover to enforce the storage policy.

After a few minutes, the mover hangs because one or more datanodes hang as 
well. Please check out the deadlock revealed by jstack. Also, here's what 
appeared in DN log:

2016-02-09 15:58:15,676 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Not able to copy block 1157230686 to /5.45.60.142:40144 because threads quota 
is exceeded.

  was:
We're running CDH 5.4.4 (with Hadoop 2.6.0). Recently we've added 800T of 
ARCHIVE storage, marked some data (16T * 2 repl. factor) as "COLD" and ran the 
hdfs mover to enforce the storage policy.

After a few minutes, the mover hangs because one or more datanodes hang as 
well. Please check out the deadlock revealed by jstack .


> DataNode deadlocks after running hdfs mover
> ---
>
> Key: HDFS-9783
> URL: https://issues.apache.org/jira/browse/HDFS-9783
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 2.6.0
>Reporter: David Watzke
> Attachments: datanode.stack.txt
>
>
> We're running CDH 5.4.4 (with Hadoop 2.6.0). Recently we've added 800T of 
> ARCHIVE storage, marked some data (16T * 2 repl. factor) as "COLD" and ran 
> the hdfs mover to enforce the storage policy.
> After a few minutes, the mover hangs because one or more datanodes hang as 
> well. Please check out the deadlock revealed by jstack. Also, here's what 
> appeared in DN log:
> 2016-02-09 15:58:15,676 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
> Not able to copy block 1157230686 to /5.45.60.142:40144 because threads quota 
> is exceeded.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9783) DataNode deadlocks after running hdfs mover

2016-02-09 Thread David Watzke (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Watzke updated HDFS-9783:
---
Description: 
We're running CDH 5.4.4 (with Hadoop 2.6.0). Recently we've added 800T of 
ARCHIVE storage, marked some data (16T * 2 repl. factor) as "COLD" and ran the 
hdfs mover to enforce the storage policy.

After a few minutes, the mover hangs because one or more datanodes hang as 
well. Please check out the deadlock revealed by jstack .

  was:
We're running CDH 5.4.4 (with Hadoop 2.6.0). Recently we've added 800T of 
ARCHIVE storage, marked some data (16T * 2 repl. factor) as "COLD" and ran the 
hdfs mover to enforce the storage policy.

After a few minutes, the mover hangs and a few datanodes hang as well.


> DataNode deadlocks after running hdfs mover
> ---
>
> Key: HDFS-9783
> URL: https://issues.apache.org/jira/browse/HDFS-9783
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 2.6.0
>Reporter: David Watzke
> Attachments: datanode.stack.txt
>
>
> We're running CDH 5.4.4 (with Hadoop 2.6.0). Recently we've added 800T of 
> ARCHIVE storage, marked some data (16T * 2 repl. factor) as "COLD" and ran 
> the hdfs mover to enforce the storage policy.
> After a few minutes, the mover hangs because one or more datanodes hang as 
> well. Please check out the deadlock revealed by jstack .



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9783) DataNode deadlocks after running hdfs mover

2016-02-09 Thread David Watzke (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Watzke updated HDFS-9783:
---
Attachment: datanode.stack.txt

deadlock revealed by jstack

> DataNode deadlocks after running hdfs mover
> ---
>
> Key: HDFS-9783
> URL: https://issues.apache.org/jira/browse/HDFS-9783
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 2.6.0
>Reporter: David Watzke
> Attachments: datanode.stack.txt
>
>
> We're running CDH 5.4.4 (with Hadoop 2.6.0). Recently we've added 800T of 
> ARCHIVE storage, marked some data (16T * 2 repl. factor) as "COLD" and ran 
> the hdfs mover to enforce the storage policy.
> After a few minutes, the mover hangs and a few datanodes hang as well.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-9783) DataNode deadlocks after running hdfs mover

2016-02-09 Thread David Watzke (JIRA)
David Watzke created HDFS-9783:
--

 Summary: DataNode deadlocks after running hdfs mover
 Key: HDFS-9783
 URL: https://issues.apache.org/jira/browse/HDFS-9783
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode
Affects Versions: 2.6.0
Reporter: David Watzke


We're running CDH 5.4.4 (with Hadoop 2.6.0). Recently we've added 800T of 
ARCHIVE storage, marked some data (16T * 2 repl. factor) as "COLD" and ran the 
hdfs mover to enforce the storage policy.

After a few minutes, the mover hangs and a few datanodes hang as well.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9395) HDFS operations vary widely in which failures they put in the audit log and which they leave out

2016-02-09 Thread Kuhu Shukla (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9395?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15139027#comment-15139027
 ] 

Kuhu Shukla commented on HDFS-9395:
---

Thank you Vinayakumar. While making this change I also noticed the 
{{checkAccess}} method which logs only when there is an ACE and not for 
successful attempts. My concern is from an older JIRA, HDFS-6570.
{quote}
getAclStatus is probably the simplest method to look at for an example that 
does everything. Do you think we need to write to the audit log for this 
method? I'm thinking that we shouldn't, because the purpose of this method is 
to query whether or not the user has access.
{quote}

[~cnauroth], Request for some comments on the methods that dont log successful 
attempts like checkAccess and getAclStatus. Thanks a lot!

> HDFS operations vary widely in which failures they put in the audit log and 
> which they leave out
> 
>
> Key: HDFS-9395
> URL: https://issues.apache.org/jira/browse/HDFS-9395
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Kihwal Lee
>Assignee: Kuhu Shukla
> Attachments: HDFS-9395.001.patch, HDFS-9395.002.patch, 
> HDFS-9395.003.patch, HDFS-9395.004.patch, HDFS-9395.005.patch
>
>
> So, the big question here is what should go in the audit log? All failures, 
> or just "permission denied" failures? Or, to put it a different way, if 
> someone attempts to do something and it fails because a file doesn't exist, 
> is that worth an audit log entry?
> We are currently inconsistent on this point. For example, concat, 
> getContentSummary, addCacheDirective, and setErasureEncodingPolicy create an 
> audit log entry for all failures, but setOwner, delete, and setAclEntries 
> attempt to only create an entry for AccessControlException-based failures. 
> There are a few operations, like allowSnapshot, disallowSnapshot, and 
> startRollingUpgrade that never create audit log failure entries at all. They 
> simply log nothing for any failure, and log success for a successful 
> operation.
> So to summarize, different HDFS operations currently fall into 3 categories:
> 1. audit-log all failures
> 2. audit-log only AccessControlException failures
> 3. never audit-log failures
> Which category is right?  And how can we fix the inconsistency



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9637) Add test for HADOOP-12702 and HADOOP-12759

2016-02-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9637?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15138924#comment-15138924
 ] 

Hadoop QA commented on HDFS-9637:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 13s 
{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s 
{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 
0s {color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 9m 
36s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 7s 
{color} | {color:green} trunk passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 56s 
{color} | {color:green} trunk passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 
29s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 11s 
{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
19s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 
35s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 41s 
{color} | {color:green} trunk passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 28s 
{color} | {color:green} trunk passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 
5s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 5s 
{color} | {color:green} the patch passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 5s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 55s 
{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 55s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 
24s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 10s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
16s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 
0s {color} | {color:green} Patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 
53s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 32s 
{color} | {color:green} the patch passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 29s 
{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 83m 52s {color} 
| {color:red} hadoop-hdfs in the patch failed with JDK v1.8.0_72. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 80m 53s {color} 
| {color:red} hadoop-hdfs in the patch failed with JDK v1.7.0_95. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 
29s {color} | {color:green} Patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 199m 3s {color} 
| {color:black} {color} |
\\
\\
|| Reason || Tests ||
| JDK v1.8.0_72 Failed junit tests | hadoop.fs.TestHdfsNativeCodeLoader |
|   | hadoop.hdfs.server.namenode.TestFSImageWithSnapshot |
|   | hadoop.hdfs.TestFileAppend |
|   | hadoop.hdfs.server.datanode.TestBlockScanner |
|   | hadoop.hdfs.server.datanode.TestDirectoryScanner |
|   | hadoop.hdfs.security.TestDelegationTokenForProxyUser |
| JDK v1.7.0_95 Failed junit tests | hadoop.fs.TestHdfsNativeCodeLoader |
|   | hadoop.hdfs.server.datanode.TestBlockScanner |
|   | hadoop.hdfs.server.datanode.TestDirectoryScanner |
| JDK v1.7.0_95 Timed out junit tests | 
org.apache.hadoop.hdfs.TestErasureCodeBenchmarkThroughput |
\\
\\
|| Subsystem || Report/Notes ||
| Do

[jira] [Commented] (HDFS-8959) Provide an iterator-based API for listing all the snapshottable directories

2016-02-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8959?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15138869#comment-15138869
 ] 

Hadoop QA commented on HDFS-8959:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 17s 
{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s 
{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 
0s {color} | {color:green} The patch appears to include 2 new or modified test 
files. {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 29s 
{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 9m 
11s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 2m 17s 
{color} | {color:green} trunk passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 40s 
{color} | {color:green} trunk passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 
40s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 53s 
{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
31s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 4m 
44s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 14s 
{color} | {color:green} trunk passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 3m 8s 
{color} | {color:green} trunk passed with JDK v1.7.0_95 {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 11s 
{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 
44s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 2m 8s 
{color} | {color:green} the patch passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} cc {color} | {color:green} 2m 8s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 2m 8s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 29s 
{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} cc {color} | {color:green} 1m 29s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 29s 
{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red} 0m 32s 
{color} | {color:red} hadoop-hdfs-project: patch generated 1 new + 797 
unchanged - 0 fixed = 798 total (was 797) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 29s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
23s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 
0s {color} | {color:green} Patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 4m 
25s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 47s 
{color} | {color:green} the patch passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 25s 
{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 1m 5s 
{color} | {color:green} hadoop-hdfs-client in the patch passed with JDK 
v1.8.0_72. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 84m 34s {color} 
| {color:red} hadoop-hdfs in the patch failed with JDK v1.8.0_72. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 1m 11s 
{color} | {color:green} hadoop-hdfs-client in the patch passed with JDK 
v1.7.0_95. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 68m 18s {color} 
| {color:red} hadoop-hdfs in the patch failed with JDK v1.7.0_95. {color} |
| {color:gr

[jira] [Commented] (HDFS-9579) Provide bytes-read-by-network-distance metrics at FileSystem.Statistics level

2016-02-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9579?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15138779#comment-15138779
 ] 

Hadoop QA commented on HDFS-9579:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 16s 
{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s 
{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 
0s {color} | {color:green} The patch appears to include 2 new or modified test 
files. {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 45s 
{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 
35s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 9m 0s 
{color} | {color:green} trunk passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 8m 16s 
{color} | {color:green} trunk passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 
11s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 47s 
{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
44s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 5m 
27s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 48s 
{color} | {color:green} trunk passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 3m 41s 
{color} | {color:green} trunk passed with JDK v1.7.0_95 {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 17s 
{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 
15s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 8m 28s 
{color} | {color:green} the patch passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 8m 28s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 7m 48s 
{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 7m 48s 
{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red} 1m 11s 
{color} | {color:red} root: patch generated 32 new + 402 unchanged - 0 fixed = 
434 total (was 402) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 45s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
42s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 
0s {color} | {color:green} Patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 6m 3s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 37s 
{color} | {color:green} the patch passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 3m 34s 
{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 9m 47s 
{color} | {color:green} hadoop-common in the patch passed with JDK v1.8.0_72. 
{color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 1m 4s 
{color} | {color:green} hadoop-hdfs-client in the patch passed with JDK 
v1.8.0_72. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 80m 53s {color} 
| {color:red} hadoop-hdfs in the patch failed with JDK v1.8.0_72. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 9m 27s 
{color} | {color:green} hadoop-common in the patch passed with JDK v1.7.0_95. 
{color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 1m 2s 
{color} | {color:green} hadoop-hdfs-client in the patch passed with JDK 
v1.7.0_95. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 102m 14s 
{color} | {color:red} hadoo

[jira] [Updated] (HDFS-9637) Add test for HADOOP-12702 and HADOOP-12759

2016-02-09 Thread Daniel Templeton (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9637?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Templeton updated HDFS-9637:
---
Attachment: HDFS-9637.004.patch

Reposting patch 4 as it wasn't the current code.

> Add test for HADOOP-12702 and HADOOP-12759
> --
>
> Key: HDFS-9637
> URL: https://issues.apache.org/jira/browse/HDFS-9637
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: test
>Affects Versions: 2.7.1
>Reporter: Daniel Templeton
>Assignee: Daniel Templeton
> Attachments: HDFS-9637.001.patch, HDFS-9637.002.patch, 
> HDFS-9637.003.patch, HDFS-9637.004.patch
>
>
> Per discussion on the dev list, the tests for the new FileSystemSink class 
> should be added to the HDFS project to avoid creating a dependency for the 
> common project on the HDFS project.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9637) Add test for HADOOP-12702 and HADOOP-12759

2016-02-09 Thread Daniel Templeton (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9637?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Templeton updated HDFS-9637:
---
Attachment: (was: HDFS-9637.004.patch)

> Add test for HADOOP-12702 and HADOOP-12759
> --
>
> Key: HDFS-9637
> URL: https://issues.apache.org/jira/browse/HDFS-9637
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: test
>Affects Versions: 2.7.1
>Reporter: Daniel Templeton
>Assignee: Daniel Templeton
> Attachments: HDFS-9637.001.patch, HDFS-9637.002.patch, 
> HDFS-9637.003.patch
>
>
> Per discussion on the dev list, the tests for the new FileSystemSink class 
> should be added to the HDFS project to avoid creating a dependency for the 
> common project on the HDFS project.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9780) RollingFileSystemSink doesn't work on secure clusters

2016-02-09 Thread Daniel Templeton (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9780?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15138646#comment-15138646
 ] 

Daniel Templeton commented on HDFS-9780:


HDFS-9782 has been created to make the interval configurable.

> RollingFileSystemSink doesn't work on secure clusters
> -
>
> Key: HDFS-9780
> URL: https://issues.apache.org/jira/browse/HDFS-9780
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.9.0
>Reporter: Daniel Templeton
>Assignee: Daniel Templeton
>Priority: Critical
> Attachments: HADOOP-12775.001.patch, HADOOP-12775.002.patch, 
> HADOOP-12775.003.patch, HDFS-9780.004.patch
>
>
> If HDFS has kerberos enabled, the sink cannot write its logs.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9782) RollingFileSystemSink should have configurable roll interval

2016-02-09 Thread Daniel Templeton (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9782?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Templeton updated HDFS-9782:
---
Attachment: HDFS-9782.001.patch

Here's an initial sketch of the plan.

> RollingFileSystemSink should have configurable roll interval
> 
>
> Key: HDFS-9782
> URL: https://issues.apache.org/jira/browse/HDFS-9782
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Daniel Templeton
>Assignee: Daniel Templeton
> Attachments: HDFS-9782.001.patch
>
>
> Right now it defaults to rolling at the top of every hour.  Instead that 
> interval should be configurable.  The interval should also allow for some 
> play so that all hosts don't try to flush their files simultaneously.
> I'm filing this in HDFS because I suspect it will involve touching the HDFS 
> tests.  If it turns out not to, I'll move it into common instead.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8959) Provide an iterator-based API for listing all the snapshottable directories

2016-02-09 Thread Rakesh R (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8959?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rakesh R updated HDFS-8959:
---
Attachment: HDFS-8959-02.patch

> Provide an iterator-based API for listing all the snapshottable directories
> ---
>
> Key: HDFS-8959
> URL: https://issues.apache.org/jira/browse/HDFS-8959
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: hdfs-client
>Reporter: Rakesh R
>Assignee: Rakesh R
> Attachments: HDFS-8959-00.patch, HDFS-8959-01.patch, 
> HDFS-8959-02.patch
>
>
> Presently {{DistributedFileSystem#getSnapshottableDirListing()}} is sending 
> all the {{SnapshottableDirectoryStatus[]}} array to the clients. Now the 
> client should have enough space to hold it in memory. There could be chance 
> that the client JVMs running out of memory because of this. Also, some time 
> back there was a 
> [comment|https://issues.apache.org/jira/browse/HDFS-8643?focusedCommentId=14658800&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14658800]
>  about RPC packet limitation and a large number of snapshot list can again 
> cause issues.
> I believe iterator based {{DistributedFileSystem#listSnapshottableDirs()}} 
> API would be a good addition!



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9637) Add test for HADOOP-12702 and HADOOP-12759

2016-02-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9637?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15138523#comment-15138523
 ] 

Hadoop QA commented on HDFS-9637:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 11s 
{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s 
{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 
0s {color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 
34s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 38s 
{color} | {color:green} trunk passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 39s 
{color} | {color:green} trunk passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 
20s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 51s 
{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
14s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 
50s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 4s 
{color} | {color:green} trunk passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 44s 
{color} | {color:green} trunk passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 
44s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 34s 
{color} | {color:green} the patch passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 34s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 37s 
{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 37s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 
18s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 48s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
11s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 
0s {color} | {color:green} Patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 8s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 1s 
{color} | {color:green} the patch passed with JDK v1.8.0_72 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 43s 
{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 51m 44s {color} 
| {color:red} hadoop-hdfs in the patch failed with JDK v1.8.0_72. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 50m 11s {color} 
| {color:red} hadoop-hdfs in the patch failed with JDK v1.7.0_95. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 
21s {color} | {color:green} Patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 125m 26s {color} 
| {color:black} {color} |
\\
\\
|| Reason || Tests ||
| JDK v1.8.0_72 Failed junit tests | hadoop.hdfs.web.TestWebHDFS |
|   | hadoop.hdfs.TestDFSUpgradeFromImage |
|   | hadoop.metrics2.sink.TestRollingFileSystemSinkWithHdfs |
|   | hadoop.fs.TestHdfsNativeCodeLoader |
| JDK v1.7.0_95 Failed junit tests | 
hadoop.metrics2.sink.TestRollingFileSystemSinkWithHdfs |
|   | hadoop.fs.TestHdfsNativeCodeLoader |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:0ca8df7 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12786998/HDFS-9637.004.patch |
| JIRA Issue | HDFS-9637 |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyl