[jira] [Comment Edited] (HADOOP-15621) s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store

2018-09-21 Thread Gabor Bota (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15621?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16624144#comment-16624144
 ] 

Gabor Bota edited comment on HADOOP-15621 at 9/21/18 8:51 PM:
--

Just a note: There is no call from {{S3AFileSytem}} to {{MetadataStore#put}} 
directly, only from {{S3Guard}} (with static methods). 
 Both {{S3Guard}} AND {{S3AFileSytem}} should be analyzed more closely how it 
interacts with {{MetadataStore}}, and all [get, listChildren, put, move] 
interactions should be modified to handle TTL.

Also note: we have usages from S3GuardTool as well, directly calling 
MetadataStore#put. We need to handle those as well.


was (Author: gabor.bota):
Just a note: There is no call from {{S3AFileSytem}} to {{MetadataStore#put}} 
directly, only from {{S3Guard}} (with static methods). 
Both {{S3Guard}} AND {{S3AFileSytem}} should be analyzed more closely how it 
interacts with {{MetadataStore}}, and all [get, listChildren, put, move] 
interactions should be modified to handle TTL.

> s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store
> --
>
> Key: HADOOP-15621
> URL: https://issues.apache.org/jira/browse/HADOOP-15621
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.0.0-beta1
>Reporter: Aaron Fabbri
>Assignee: Gabor Bota
>Priority: Minor
> Attachments: HADOOP-15621.001.patch
>
>
> Similar to HADOOP-13649, I think we should add a TTL (time to live) feature 
> to the Dynamo metadata store (MS) for S3Guard.
> This is a similar concept to an "online algorithm" version of the CLI prune() 
> function, which is the "offline algorithm".
> Why: 
>  1. Self healing (soft state): since we do not implement transactions around 
> modification of the two systems (s3 and metadata store), certain failures can 
> lead to inconsistency between S3 and the metadata store (MS) state. Having a 
> time to live (TTL) on each entry in S3Guard means that any inconsistencies 
> will be time bound. Thus "wait and restart your job" becomes a valid, if 
> ugly, way to get around any issues with FS client failure leaving things in a 
> bad state.
>  2. We could make manual invocation of `hadoop s3guard prune ...` 
> unnecessary, depending on the implementation.
>  3. Makes it possible to fix the problem that dynamo MS prune() doesn't prune 
> directories due to the lack of true modification time.
> How:
>  I think we need a new column in the dynamo table "entry last written time". 
> This is updated each time the entry is written to dynamo.
>  After that we can either
>  1. Have the client simply ignore / elide any entries that are older than the 
> configured TTL.
>  2. Have the client delete entries older than the TTL.
> The issue with #2 is it will increase latency if done inline in the context 
> of an FS operation. We could mitigate this some by using an async helper 
> thread, or probabilistically doing it "some times" to amortize the expense of 
> deleting stale entries (allowing some batching as well).
> Caveats:
>  - Clock synchronization as usual is a concern. Many clusters already keep 
> clocks close enough via NTP. We should at least document the requirement 
> along with the configuration knob that enables the feature.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-15621) s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store

2018-09-21 Thread Gabor Bota (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15621?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16624144#comment-16624144
 ] 

Gabor Bota edited comment on HADOOP-15621 at 9/21/18 8:51 PM:
--

There is no call from {{S3AFileSytem}} to {{MetadataStore#put}} directly, only 
from {{S3Guard}} (with static methods). 
 Both {{S3Guard}} AND {{S3AFileSytem}} should be analyzed more closely how it 
interacts with {{MetadataStore}}, and all [get, listChildren, put, move] 
interactions should be modified to handle TTL.

Also note: we have usages from S3GuardTool as well, directly calling 
MetadataStore#put. We need to handle those as well.


was (Author: gabor.bota):
Just a note: There is no call from {{S3AFileSytem}} to {{MetadataStore#put}} 
directly, only from {{S3Guard}} (with static methods). 
 Both {{S3Guard}} AND {{S3AFileSytem}} should be analyzed more closely how it 
interacts with {{MetadataStore}}, and all [get, listChildren, put, move] 
interactions should be modified to handle TTL.

Also note: we have usages from S3GuardTool as well, directly calling 
MetadataStore#put. We need to handle those as well.

> s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store
> --
>
> Key: HADOOP-15621
> URL: https://issues.apache.org/jira/browse/HADOOP-15621
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.0.0-beta1
>Reporter: Aaron Fabbri
>Assignee: Gabor Bota
>Priority: Minor
> Attachments: HADOOP-15621.001.patch
>
>
> Similar to HADOOP-13649, I think we should add a TTL (time to live) feature 
> to the Dynamo metadata store (MS) for S3Guard.
> This is a similar concept to an "online algorithm" version of the CLI prune() 
> function, which is the "offline algorithm".
> Why: 
>  1. Self healing (soft state): since we do not implement transactions around 
> modification of the two systems (s3 and metadata store), certain failures can 
> lead to inconsistency between S3 and the metadata store (MS) state. Having a 
> time to live (TTL) on each entry in S3Guard means that any inconsistencies 
> will be time bound. Thus "wait and restart your job" becomes a valid, if 
> ugly, way to get around any issues with FS client failure leaving things in a 
> bad state.
>  2. We could make manual invocation of `hadoop s3guard prune ...` 
> unnecessary, depending on the implementation.
>  3. Makes it possible to fix the problem that dynamo MS prune() doesn't prune 
> directories due to the lack of true modification time.
> How:
>  I think we need a new column in the dynamo table "entry last written time". 
> This is updated each time the entry is written to dynamo.
>  After that we can either
>  1. Have the client simply ignore / elide any entries that are older than the 
> configured TTL.
>  2. Have the client delete entries older than the TTL.
> The issue with #2 is it will increase latency if done inline in the context 
> of an FS operation. We could mitigate this some by using an async helper 
> thread, or probabilistically doing it "some times" to amortize the expense of 
> deleting stale entries (allowing some batching as well).
> Caveats:
>  - Clock synchronization as usual is a concern. Many clusters already keep 
> clocks close enough via NTP. We should at least document the requirement 
> along with the configuration knob that enables the feature.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-15621) s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store

2018-09-21 Thread Gabor Bota (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15621?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16624144#comment-16624144
 ] 

Gabor Bota edited comment on HADOOP-15621 at 9/21/18 8:44 PM:
--

Just a note: There is no call from {{S3AFileSytem}} to {{MetadataStore#put}} 
directly, only from {{S3Guard}} (with static methods). 
Both {{S3Guard}} AND {{S3AFileSytem}} should be analyzed more closely how it 
interacts with {{MetadataStore}}, and all {get, listChildren, put, move} 
interactions should be modified to handle TTL.


was (Author: gabor.bota):
Just a note: There is no call from {{S3AFileSytem}} to {{MetadataStore#put}} 
directly, only from {{S3Guard}} (with static methods). 
Both {{S3Guard}} AND {{S3AFileSytem}} should be analyzed more closely how it 
interacts with {{MetadataStore}}, and all interactions should be modified to 
handle TTL.

> s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store
> --
>
> Key: HADOOP-15621
> URL: https://issues.apache.org/jira/browse/HADOOP-15621
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.0.0-beta1
>Reporter: Aaron Fabbri
>Assignee: Gabor Bota
>Priority: Minor
> Attachments: HADOOP-15621.001.patch
>
>
> Similar to HADOOP-13649, I think we should add a TTL (time to live) feature 
> to the Dynamo metadata store (MS) for S3Guard.
> This is a similar concept to an "online algorithm" version of the CLI prune() 
> function, which is the "offline algorithm".
> Why: 
>  1. Self healing (soft state): since we do not implement transactions around 
> modification of the two systems (s3 and metadata store), certain failures can 
> lead to inconsistency between S3 and the metadata store (MS) state. Having a 
> time to live (TTL) on each entry in S3Guard means that any inconsistencies 
> will be time bound. Thus "wait and restart your job" becomes a valid, if 
> ugly, way to get around any issues with FS client failure leaving things in a 
> bad state.
>  2. We could make manual invocation of `hadoop s3guard prune ...` 
> unnecessary, depending on the implementation.
>  3. Makes it possible to fix the problem that dynamo MS prune() doesn't prune 
> directories due to the lack of true modification time.
> How:
>  I think we need a new column in the dynamo table "entry last written time". 
> This is updated each time the entry is written to dynamo.
>  After that we can either
>  1. Have the client simply ignore / elide any entries that are older than the 
> configured TTL.
>  2. Have the client delete entries older than the TTL.
> The issue with #2 is it will increase latency if done inline in the context 
> of an FS operation. We could mitigate this some by using an async helper 
> thread, or probabilistically doing it "some times" to amortize the expense of 
> deleting stale entries (allowing some batching as well).
> Caveats:
>  - Clock synchronization as usual is a concern. Many clusters already keep 
> clocks close enough via NTP. We should at least document the requirement 
> along with the configuration knob that enables the feature.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-15621) s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store

2018-09-21 Thread Gabor Bota (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15621?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16624144#comment-16624144
 ] 

Gabor Bota edited comment on HADOOP-15621 at 9/21/18 8:44 PM:
--

Just a note: There is no call from {{S3AFileSytem}} to {{MetadataStore#put}} 
directly, only from {{S3Guard}} (with static methods). 
Both {{S3Guard}} AND {{S3AFileSytem}} should be analyzed more closely how it 
interacts with {{MetadataStore}}, and all [get, listChildren, put, move] 
interactions should be modified to handle TTL.


was (Author: gabor.bota):
Just a note: There is no call from {{S3AFileSytem}} to {{MetadataStore#put}} 
directly, only from {{S3Guard}} (with static methods). 
Both {{S3Guard}} AND {{S3AFileSytem}} should be analyzed more closely how it 
interacts with {{MetadataStore}}, and all {get, listChildren, put, move} 
interactions should be modified to handle TTL.

> s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store
> --
>
> Key: HADOOP-15621
> URL: https://issues.apache.org/jira/browse/HADOOP-15621
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.0.0-beta1
>Reporter: Aaron Fabbri
>Assignee: Gabor Bota
>Priority: Minor
> Attachments: HADOOP-15621.001.patch
>
>
> Similar to HADOOP-13649, I think we should add a TTL (time to live) feature 
> to the Dynamo metadata store (MS) for S3Guard.
> This is a similar concept to an "online algorithm" version of the CLI prune() 
> function, which is the "offline algorithm".
> Why: 
>  1. Self healing (soft state): since we do not implement transactions around 
> modification of the two systems (s3 and metadata store), certain failures can 
> lead to inconsistency between S3 and the metadata store (MS) state. Having a 
> time to live (TTL) on each entry in S3Guard means that any inconsistencies 
> will be time bound. Thus "wait and restart your job" becomes a valid, if 
> ugly, way to get around any issues with FS client failure leaving things in a 
> bad state.
>  2. We could make manual invocation of `hadoop s3guard prune ...` 
> unnecessary, depending on the implementation.
>  3. Makes it possible to fix the problem that dynamo MS prune() doesn't prune 
> directories due to the lack of true modification time.
> How:
>  I think we need a new column in the dynamo table "entry last written time". 
> This is updated each time the entry is written to dynamo.
>  After that we can either
>  1. Have the client simply ignore / elide any entries that are older than the 
> configured TTL.
>  2. Have the client delete entries older than the TTL.
> The issue with #2 is it will increase latency if done inline in the context 
> of an FS operation. We could mitigate this some by using an async helper 
> thread, or probabilistically doing it "some times" to amortize the expense of 
> deleting stale entries (allowing some batching as well).
> Caveats:
>  - Clock synchronization as usual is a concern. Many clusters already keep 
> clocks close enough via NTP. We should at least document the requirement 
> along with the configuration knob that enables the feature.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-15621) s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store

2018-09-21 Thread Gabor Bota (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15621?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16624144#comment-16624144
 ] 

Gabor Bota edited comment on HADOOP-15621 at 9/21/18 8:40 PM:
--

Just a note: There is no call from {{S3AFileSytem}} to {{MetadataStore#put}} 
directly, only from {{S3Guard}} (with static methods). 
Both {{S3Guard}} AND {{S3AFileSytem}} should be analyzed more closely how it 
interacts with {{MetadataStore}}, and all interactions should be modified to 
handle TTL.


was (Author: gabor.bota):
Just a note: There is no call from {{S3AFileSytem}} to {{MetadataStore#put}} 
directly, only from {{S3Guard}} (with static methods). 
{{S3Guard}} should be analyzed more closely how it interacts with 
{{MetadataStore}}, and all interactions should be modified to handle TTL.

> s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store
> --
>
> Key: HADOOP-15621
> URL: https://issues.apache.org/jira/browse/HADOOP-15621
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.0.0-beta1
>Reporter: Aaron Fabbri
>Assignee: Gabor Bota
>Priority: Minor
> Attachments: HADOOP-15621.001.patch
>
>
> Similar to HADOOP-13649, I think we should add a TTL (time to live) feature 
> to the Dynamo metadata store (MS) for S3Guard.
> This is a similar concept to an "online algorithm" version of the CLI prune() 
> function, which is the "offline algorithm".
> Why: 
>  1. Self healing (soft state): since we do not implement transactions around 
> modification of the two systems (s3 and metadata store), certain failures can 
> lead to inconsistency between S3 and the metadata store (MS) state. Having a 
> time to live (TTL) on each entry in S3Guard means that any inconsistencies 
> will be time bound. Thus "wait and restart your job" becomes a valid, if 
> ugly, way to get around any issues with FS client failure leaving things in a 
> bad state.
>  2. We could make manual invocation of `hadoop s3guard prune ...` 
> unnecessary, depending on the implementation.
>  3. Makes it possible to fix the problem that dynamo MS prune() doesn't prune 
> directories due to the lack of true modification time.
> How:
>  I think we need a new column in the dynamo table "entry last written time". 
> This is updated each time the entry is written to dynamo.
>  After that we can either
>  1. Have the client simply ignore / elide any entries that are older than the 
> configured TTL.
>  2. Have the client delete entries older than the TTL.
> The issue with #2 is it will increase latency if done inline in the context 
> of an FS operation. We could mitigate this some by using an async helper 
> thread, or probabilistically doing it "some times" to amortize the expense of 
> deleting stale entries (allowing some batching as well).
> Caveats:
>  - Clock synchronization as usual is a concern. Many clusters already keep 
> clocks close enough via NTP. We should at least document the requirement 
> along with the configuration knob that enables the feature.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15621) s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store

2018-09-21 Thread Gabor Bota (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15621?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16624144#comment-16624144
 ] 

Gabor Bota commented on HADOOP-15621:
-

Just a note: There is no call from {{S3AFileSytem}} to {{MetadataStore#put}} 
directly, only from {{S3Guard}} (with static methods). 
{{S3Guard}} should be analyzed more closely how it interacts with 
{{MetadataStore}}, and all interactions should be modified to handle TTL.

> s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store
> --
>
> Key: HADOOP-15621
> URL: https://issues.apache.org/jira/browse/HADOOP-15621
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.0.0-beta1
>Reporter: Aaron Fabbri
>Assignee: Gabor Bota
>Priority: Minor
> Attachments: HADOOP-15621.001.patch
>
>
> Similar to HADOOP-13649, I think we should add a TTL (time to live) feature 
> to the Dynamo metadata store (MS) for S3Guard.
> This is a similar concept to an "online algorithm" version of the CLI prune() 
> function, which is the "offline algorithm".
> Why: 
>  1. Self healing (soft state): since we do not implement transactions around 
> modification of the two systems (s3 and metadata store), certain failures can 
> lead to inconsistency between S3 and the metadata store (MS) state. Having a 
> time to live (TTL) on each entry in S3Guard means that any inconsistencies 
> will be time bound. Thus "wait and restart your job" becomes a valid, if 
> ugly, way to get around any issues with FS client failure leaving things in a 
> bad state.
>  2. We could make manual invocation of `hadoop s3guard prune ...` 
> unnecessary, depending on the implementation.
>  3. Makes it possible to fix the problem that dynamo MS prune() doesn't prune 
> directories due to the lack of true modification time.
> How:
>  I think we need a new column in the dynamo table "entry last written time". 
> This is updated each time the entry is written to dynamo.
>  After that we can either
>  1. Have the client simply ignore / elide any entries that are older than the 
> configured TTL.
>  2. Have the client delete entries older than the TTL.
> The issue with #2 is it will increase latency if done inline in the context 
> of an FS operation. We could mitigate this some by using an async helper 
> thread, or probabilistically doing it "some times" to amortize the expense of 
> deleting stale entries (allowing some batching as well).
> Caveats:
>  - Clock synchronization as usual is a concern. Many clusters already keep 
> clocks close enough via NTP. We should at least document the requirement 
> along with the configuration knob that enables the feature.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-15621) s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store

2018-09-21 Thread Gabor Bota (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15621?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16607040#comment-16607040
 ] 

Gabor Bota edited comment on HADOOP-15621 at 9/21/18 8:31 PM:
--

Based on your comments [~fabbri] I think we could rename this jira. It's more 
like a TTL for authoritative directory listings and files for all metadata 
store rather than a ttl expiry for dynamodb.

 


was (Author: gabor.bota):
Based on your comments [~fabbri] I think we could rename this jira. It's more 
like a TTL for authoritative directory listings for all metadata store rather 
than a ttl expiry for dynamodb.

 

> s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store
> --
>
> Key: HADOOP-15621
> URL: https://issues.apache.org/jira/browse/HADOOP-15621
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.0.0-beta1
>Reporter: Aaron Fabbri
>Assignee: Gabor Bota
>Priority: Minor
> Attachments: HADOOP-15621.001.patch
>
>
> Similar to HADOOP-13649, I think we should add a TTL (time to live) feature 
> to the Dynamo metadata store (MS) for S3Guard.
> This is a similar concept to an "online algorithm" version of the CLI prune() 
> function, which is the "offline algorithm".
> Why: 
>  1. Self healing (soft state): since we do not implement transactions around 
> modification of the two systems (s3 and metadata store), certain failures can 
> lead to inconsistency between S3 and the metadata store (MS) state. Having a 
> time to live (TTL) on each entry in S3Guard means that any inconsistencies 
> will be time bound. Thus "wait and restart your job" becomes a valid, if 
> ugly, way to get around any issues with FS client failure leaving things in a 
> bad state.
>  2. We could make manual invocation of `hadoop s3guard prune ...` 
> unnecessary, depending on the implementation.
>  3. Makes it possible to fix the problem that dynamo MS prune() doesn't prune 
> directories due to the lack of true modification time.
> How:
>  I think we need a new column in the dynamo table "entry last written time". 
> This is updated each time the entry is written to dynamo.
>  After that we can either
>  1. Have the client simply ignore / elide any entries that are older than the 
> configured TTL.
>  2. Have the client delete entries older than the TTL.
> The issue with #2 is it will increase latency if done inline in the context 
> of an FS operation. We could mitigate this some by using an async helper 
> thread, or probabilistically doing it "some times" to amortize the expense of 
> deleting stale entries (allowing some batching as well).
> Caveats:
>  - Clock synchronization as usual is a concern. Many clusters already keep 
> clocks close enough via NTP. We should at least document the requirement 
> along with the configuration knob that enables the feature.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15774) Discovery of HA servers

2018-09-21 Thread Anu Engineer (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15774?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16624003#comment-16624003
 ] 

Anu Engineer commented on HADOOP-15774:
---

{quote}[~anu] any thoughts on the general approach? I'm not very familiar with 
the OZone architecture; do you guys also need to specify servers manually?
{quote}
[~elgoiri] Thanks for posting the design docs. My apologies for the delayed 
response, been busy with Ozone Alpha release.

I have proposed the same idea a couple of times internally. Each time it got 
shot down because of the issue that [~ste...@apache.org] is referring to.
{quote}Being part of the YARN may bring some weird dependencies (HDFS->YARN), 
any ideas here? Should it be move to a separate place? Maybe parts?
{quote}
The most important concern was always HDFS taking a dependency on YARN. It 
creates a cyclical dependency.

Ozone needs and plans to build something very similar to this, but our thought 
has been to create something like this in SCM.

For Hadoop and also from Ozone perspective, The consideration for such a 
service should be slightly different. It is possible for someone to run a 
cluster without HDFS or even without YARN. A Discovery service should be 
independent of these specific services.

If we are planning to make this a Hadoop level discovery service we should 
probably build this independent of all other services. If possible, even 
independent of ZooKeeper.

Then this service can be the bootup service, and all other services including 
HDFS, YARN and even Zookeeper can use this service to discover endpoints. Many 
services need to find the address of zookeeper too.

If this service can be built independent of all other services, Ozone can 
certainly use this, and we don't need to reinvent discovery.

Another critical issue that Ozone struggles with this the issue of config 
changes – for example, in the HA case if a server fails and a new server is 
added Datanodes need to be told of that.

Ozone solves it by reading these changes from SCM, not only config via 
Heartbeat.

But this problem is more generic, in the sense that there is a class of 
configuration changes that need to be pushed to all entities in the cluster. We 
have been thinking about building a discovery and config store for Ozone.

> Discovery of HA servers
> ---
>
> Key: HADOOP-15774
> URL: https://issues.apache.org/jira/browse/HADOOP-15774
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Íñigo Goiri
>Priority: Major
> Attachments: Discovery Service.pdf
>
>
> Currently, Hadoop relies on configuration files to specify the servers.
> This requires maintaining these configuration files and propagating the 
> changes.
> Hadoop should have a framework to provide discovery.
> For example, in HDFS, we could define the Namenodes in a shared location and 
> the DNs would use the framework to find the Namenodes.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15775) [JDK9] Add missing javax.activation-api dependency

2018-09-21 Thread Andrew Purtell (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15775?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623975#comment-16623975
 ] 

Andrew Purtell commented on HADOOP-15775:
-

FWIW the 003 patch doesn't address some cases, like 
TestWebHDFSStoragePolicyCommands. You need to specify this dependency in the 
hadoop-hdfs-project POM too.

Cause is:
{noformat}
Caused by: java.lang.ClassNotFoundException: javax.activation.DataSource
at 
java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:582)
at 
java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:521)
{noformat}

This is Java 11+28


> [JDK9] Add missing javax.activation-api dependency
> --
>
> Key: HADOOP-15775
> URL: https://issues.apache.org/jira/browse/HADOOP-15775
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: test
>Reporter: Akira Ajisaka
>Assignee: Akira Ajisaka
>Priority: Critical
> Attachments: HADOOP-15775.01.patch, HADOOP-15775.02.patch, 
> HADOOP-15775.03.patch
>
>
> Many unit tests fail due to missing java.activation module. This failure can 
> be fixed by adding javax.activation-api as third-party dependency.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-15775) [JDK9] Add missing javax.activation-api dependency

2018-09-21 Thread Andrew Purtell (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15775?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623975#comment-16623975
 ] 

Andrew Purtell edited comment on HADOOP-15775 at 9/21/18 6:02 PM:
--

FWIW the 003 patch doesn't address some cases, like 
TestWebHDFSStoragePolicyCommands. You need to specify this dependency in the 
hadoop-hdfs-project POM too.

Cause for TestWebHDFSStoragePolicyCommands failure (and others) is:
{noformat}
Caused by: java.lang.ClassNotFoundException: javax.activation.DataSource
at 
java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:582)
at 
java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:521)
{noformat}

This is Java 11+28



was (Author: apurtell):
FWIW the 003 patch doesn't address some cases, like 
TestWebHDFSStoragePolicyCommands. You need to specify this dependency in the 
hadoop-hdfs-project POM too.

Cause is:
{noformat}
Caused by: java.lang.ClassNotFoundException: javax.activation.DataSource
at 
java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:582)
at 
java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:521)
{noformat}

This is Java 11+28


> [JDK9] Add missing javax.activation-api dependency
> --
>
> Key: HADOOP-15775
> URL: https://issues.apache.org/jira/browse/HADOOP-15775
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: test
>Reporter: Akira Ajisaka
>Assignee: Akira Ajisaka
>Priority: Critical
> Attachments: HADOOP-15775.01.patch, HADOOP-15775.02.patch, 
> HADOOP-15775.03.patch
>
>
> Many unit tests fail due to missing java.activation module. This failure can 
> be fixed by adding javax.activation-api as third-party dependency.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15711) Fix branch-2 builds

2018-09-21 Thread Konstantin Shvachko (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15711?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623935#comment-16623935
 ] 

Konstantin Shvachko commented on HADOOP-15711:
--

[~aw] thanks a lot for your insight, this helps understanding the problem and 
differences between the branches.

Looked at the PreCommit runs with the attached patch more closely. Definitely 
some cheating there, as many tests were not executed at all (Just as [~aw] 
commented). This explains fast running time.
 I also made Jenkins [execute branch-3.1 
build|https://builds.apache.org/view/H-L/view/Hadoop/job/hadoop-qbt-branch2-java7-linux-x86-jhung/12/console],
 which completed in 12 hours.

I don't see any shortcut solutions to the branch-2 build problem at this point. 
We should probably take the long path and understand the fixes that were done 
to unit tests on branch-3, that were not backported to branch-2.

> Fix branch-2 builds
> ---
>
> Key: HADOOP-15711
> URL: https://issues.apache.org/jira/browse/HADOOP-15711
> Project: Hadoop Common
>  Issue Type: Task
>Reporter: Jonathan Hung
>Priority: Critical
> Attachments: HADOOP-15711.001.branch-2.patch
>
>
> Branch-2 builds have been disabled for a while: 
> https://builds.apache.org/view/H-L/view/Hadoop/job/hadoop-qbt-branch2-java7-linux-x86/
> A test run here causes hdfs tests to hang: 
> https://builds.apache.org/view/H-L/view/Hadoop/job/hadoop-qbt-branch2-java7-linux-x86-jhung/4/
> Running hadoop-hdfs tests locally reveal some errors such 
> as:{noformat}[ERROR] 
> testComplexAppend2(org.apache.hadoop.hdfs.TestFileAppend2)  Time elapsed: 
> 0.059 s  <<< ERROR!
> java.lang.OutOfMemoryError: unable to create new native thread
> at java.lang.Thread.start0(Native Method)
> at java.lang.Thread.start(Thread.java:714)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1164)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1128)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:174)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1172)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:403)
> at 
> org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:234)
> at 
> org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:1080)
> at 
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:883)
> at 
> org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514)
> at 
> org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473)
> at 
> org.apache.hadoop.hdfs.TestFileAppend2.testComplexAppend(TestFileAppend2.java:489)
> at 
> org.apache.hadoop.hdfs.TestFileAppend2.testComplexAppend2(TestFileAppend2.java:543)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){noformat}
> I was able to get more tests passing locally by increasing the max user 
> process count on my machine. But the error suggests that there's an issue in 
> the tests themselves. Not sure if the error seen locally is the same reason 
> as why jenkins builds are failing, I wasn't able to confirm based on the 
> jenkins builds' lack of output.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15723) ABFS: Ranger Support

2018-09-21 Thread Da Zhou (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15723?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623926#comment-16623926
 ] 

Da Zhou commented on HADOOP-15723:
--

The error indicates the reason: "URI scheme must be *abfss* when authenticating 
using Oauth", 
 "ITestAbfsFileSystemContractDistCp" is only used to test in secure mode 
(abfs://XXX), so when using Oauth (Oauth only support abfss://), 
"ITestAbfsFileSystemContractDistCp" should be skipped. 
 In future, we should merge "ITestAbfsFileSystemContractDistCp" and 
"ITestAbfsFileSystemContractSecureDistCp" since they are running same tests but 
in different mode.
 But for now, you can just skip this test set by checking if it is using Oauth 
Type.

> ABFS: Ranger Support
> 
>
> Key: HADOOP-15723
> URL: https://issues.apache.org/jira/browse/HADOOP-15723
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/azure
>Affects Versions: HADOOP-15407
>Reporter: Thomas Marquardt
>Assignee: Yuan Gao
>Priority: Major
> Attachments: HADOOP-15273-HADOOP-15407-001.patch, 
> HADOOP-15273-HADOOP-15407-002.patch, HADOOP-15273-HADOOP-15407-003.patch, 
> HADOOP-15273-HADOOP-15407-004.patch
>
>
> Add support for Ranger



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-15723) ABFS: Ranger Support

2018-09-21 Thread Da Zhou (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15723?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623926#comment-16623926
 ] 

Da Zhou edited comment on HADOOP-15723 at 9/21/18 5:35 PM:
---

The error indicates the reason: "URI scheme must be *abfss* when authenticating 
using Oauth", 
 "ITestAbfsFileSystemContractDistCp" is only used to test in secure mode 
(abfs://XXX), so when using Oauth (Oauth only support abfss://), 
"ITestAbfsFileSystemContractDistCp" should be skipped. 
 In future, we should merge "ITestAbfsFileSystemContractDistCp" and 
"ITestAbfsFileSystemContractSecureDistCp" since they are running same tests but 
in different mode.
 But for now, you can just skip this test set by checking if it is using Oauth 
authentication.


was (Author: danielzhou):
The error indicates the reason: "URI scheme must be *abfss* when authenticating 
using Oauth", 
 "ITestAbfsFileSystemContractDistCp" is only used to test in secure mode 
(abfs://XXX), so when using Oauth (Oauth only support abfss://), 
"ITestAbfsFileSystemContractDistCp" should be skipped. 
 In future, we should merge "ITestAbfsFileSystemContractDistCp" and 
"ITestAbfsFileSystemContractSecureDistCp" since they are running same tests but 
in different mode.
 But for now, you can just skip this test set by checking if it is using Oauth 
Type.

> ABFS: Ranger Support
> 
>
> Key: HADOOP-15723
> URL: https://issues.apache.org/jira/browse/HADOOP-15723
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/azure
>Affects Versions: HADOOP-15407
>Reporter: Thomas Marquardt
>Assignee: Yuan Gao
>Priority: Major
> Attachments: HADOOP-15273-HADOOP-15407-001.patch, 
> HADOOP-15273-HADOOP-15407-002.patch, HADOOP-15273-HADOOP-15407-003.patch, 
> HADOOP-15273-HADOOP-15407-004.patch
>
>
> Add support for Ranger



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-15723) ABFS: Ranger Support

2018-09-21 Thread Yuan Gao (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15723?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623725#comment-16623725
 ] 

Yuan Gao edited comment on HADOOP-15723 at 9/21/18 5:14 PM:


patch 004:
 - called makeQualified() before authz checks
 - moved configuration and factory method to AbfsConfiguration
 - moved the authorization interface to org.apache.hadoop.fs.azurebfs.extensions
 - renamed from adl to abfs
 - added test cases for authz checks in ITestAzureBlobFileSystemAuthorization

I was able to run:

{{mvn -T 1C -Dparallel-tests -Dscale -DtestsThreadCount=8 clean verify}}
 {{[INFO] Results:}}
 {{[WARNING] Tests run: 277, Failures: 0, Errors: 0, Skipped: 11}}

But there are errors when running:

{{mvn -T 1C -Dparallel-tests=abfs -Dscale -DtestsThreadCount=8 clean verify}}

{{[ERROR] Tests run: 6, Failures: 0, Errors: 6, Skipped: 0, Time elapsed: 5.955 
s <<< FAILURE! - in 
org.apache.hadoop.fs.azurebfs.contract.ITestAbfsFileSystemContractDistCp}}
 {{[ERROR] 
testUpdateDeepDirectoryStructureNoChange(org.apache.hadoop.fs.azurebfs.contract.ITestAbfsFileSystemContractDistCp)
 Time elapsed: 3.038 s <<< ERROR!}}
 {{java.io.IOException: Unable to initialize filesystem, URI scheme must be 
abfss when authenticating using Oauth.}}


was (Author: kowon2008):
patch 004:
 - called makeQualified() before authz checks
 - moved configuration and factory method to AbfsConfiguration
 - moved the authorization interface to org.apache.hadoop.fs.azurebfs.extensions
 - renamed from adl to abfs
 - added test cases for authz checks in ITestAzureBlobFileSystemAuthorization

I was able to run:

{{mvn -T 1C -Dparallel-tests -Dscale -DtestsThreadCount=8 clean verify}}
{{[INFO] Results:}}
{{[WARNING] Tests run: 277, Failures: 0, Errors: 0, Skipped: 11}}

But there are errors when running:

{{mvn -T 1C -Dparallel-tests=abfs -Dscale -DtestsThreadCount=8 clean verify}}

{{[ERROR] Tests run: 6, Failures: 0, Errors: 6, Skipped: 0, Time elapsed: 5.955 
s <<< FAILURE! - in 
org.apache.hadoop.fs.azurebfs.contract.ITestAbfsFileSystemContractDistCp}}
 {{[ERROR] 
testUpdateDeepDirectoryStructureNoChange(org.apache.hadoop.fs.azurebfs.contract.ITestAbfsFileSystemContractDistCp)
 Time elapsed: 3.038 s <<< ERROR!}}
 {{java.io.IOException: Unable to initialize filesystem 
abfs://dfsafast...@rajeevbstorageac.dfs.core.windows.net: 
java.lang.IllegalArgumentException: Incorrect URI 
abfs://dfsafast...@rajeevbstorageac.dfs.core.windows.net, URI scheme must be 
abfss when authenticating using Oauth.}}

> ABFS: Ranger Support
> 
>
> Key: HADOOP-15723
> URL: https://issues.apache.org/jira/browse/HADOOP-15723
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/azure
>Affects Versions: HADOOP-15407
>Reporter: Thomas Marquardt
>Assignee: Yuan Gao
>Priority: Major
> Attachments: HADOOP-15273-HADOOP-15407-001.patch, 
> HADOOP-15273-HADOOP-15407-002.patch, HADOOP-15273-HADOOP-15407-003.patch, 
> HADOOP-15273-HADOOP-15407-004.patch
>
>
> Add support for Ranger



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15765) Can not find login module class for IBM due to hard codes

2018-09-21 Thread Eric Yang (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15765?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623890#comment-16623890
 ] 

Eric Yang commented on HADOOP-15765:


[~jiangjianfei] How about IBM Java version detection to make decision which 
security class to use?  I believe that was the approach taken by existing 
developers.

> Can not find login module class for IBM due to hard codes
> -
>
> Key: HADOOP-15765
> URL: https://issues.apache.org/jira/browse/HADOOP-15765
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: security
>Affects Versions: 3.0.3
>Reporter: Jianfei Jiang
>Priority: Major
> Attachments: HADOOP-15765_000.patch
>
>
> As the differences between various versions of IBM, the login module class is 
> sometimes different. However, the class for specified jdk (no matter the 
> version) is hard coded in Hadoop code. We have faced the error like following:
> *javax.security.auth.login.LoginException: unable to find LoginModule class: 
> com.ibm.security.auth.module.LinuxLoginModule*
>  
> Should we set the value as a config which can be set by users?



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15691) Add PathCapabilities to FS and FC to complement StreamCapabilities

2018-09-21 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15691?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623885#comment-16623885
 ] 

Hadoop QA commented on HADOOP-15691:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
17s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 4 new or modified test 
files. {color} |
|| || || || {color:brown} trunk Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
58s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 22m 
29s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 18m 
37s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  3m 
53s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  4m 
42s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
21m 46s{color} | {color:green} branch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  7m 
11s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  3m 
28s{color} | {color:green} trunk passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
22s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  4m 
10s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 17m 
54s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 17m 
54s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
3m 30s{color} | {color:orange} root: The patch generated 4 new + 525 unchanged 
- 0 fixed = 529 total (was 525) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  4m 
38s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch has 12 line(s) that end in whitespace. Use 
git apply --whitespace=fix <>. Refer 
https://git-scm.com/docs/git-apply {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
11m 34s{color} | {color:green} patch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  7m  
2s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} javadoc {color} | {color:red}  0m 
58s{color} | {color:red} hadoop-common-project_hadoop-common generated 2 new + 
0 unchanged - 0 fixed = 2 total (was 0) {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  8m 
27s{color} | {color:green} hadoop-common in the patch passed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  1m 
40s{color} | {color:green} hadoop-hdfs-client in the patch passed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  3m 
58s{color} | {color:green} hadoop-hdfs-httpfs in the patch passed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  4m 
38s{color} | {color:green} hadoop-aws in the patch passed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  1m 
14s{color} | {color:green} hadoop-azure in the patch passed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  0m 
55s{color} | {color:green} hadoop-azure-datalake in the patch passed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
39s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}153m  6s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce 

[jira] [Commented] (HADOOP-15679) ShutdownHookManager shutdown time needs to be configurable & extended

2018-09-21 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15679?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623882#comment-16623882
 ] 

Hadoop QA commented on HADOOP-15679:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 16m  
3s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m  
0s{color} | {color:blue} Findbugs executables are not available. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
|| || || || {color:brown} branch-2 Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 14m 
47s{color} | {color:green} branch-2 passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 12m 
20s{color} | {color:green} branch-2 passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
36s{color} | {color:green} branch-2 passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
13s{color} | {color:green} branch-2 passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
3s{color} | {color:green} branch-2 passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
49s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 12m 
45s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 12m 
45s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
0m 36s{color} | {color:orange} hadoop-common-project/hadoop-common: The patch 
generated 1 new + 115 unchanged - 1 fixed = 116 total (was 116) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
15s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
0s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
4s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 75m 24s{color} 
| {color:red} hadoop-common in the patch failed. {color} |
| {color:red}-1{color} | {color:red} asflicense {color} | {color:red}  0m 
44s{color} | {color:red} The patch generated 1 ASF License warnings. {color} |
| {color:black}{color} | {color:black} {color} | {color:black}140m 22s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:a716388 |
| JIRA Issue | HADOOP-15679 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12940790/HADOOP-15679-branch-2-004.patch
 |
| Optional Tests |  dupname  asflicense  compile  javac  javadoc  mvninstall  
mvnsite  unit  shadedclient  findbugs  checkstyle  xml  |
| uname | Linux 1408c8a2646c 3.13.0-143-generic #192-Ubuntu SMP Tue Feb 27 
10:45:36 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/patchprocess/precommit/personality/provided.sh |
| git revision | branch-2 / 3a6ad9c |
| maven | version: Apache Maven 3.3.9 
(bb52d8502b132ec0a5a3f4c09453c07478323dc5; 2015-11-10T16:41:47+00:00) |
| Default Java | 1.7.0_181 |
| checkstyle | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/15249/artifact/out/diff-checkstyle-hadoop-common-project_hadoop-common.txt
 |
| unit | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/15249/artifact/out/patch-unit-hadoop-common-project_hadoop-common.txt
 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/15249/testReport/ |
| asflicense | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/15249/artifact/out/patch-asflicense-problems.txt
 |
| Max. process+thread count | 1361 (vs. ulimit of 1) |
| modules | C: hadoop-common-project/hadoop-common U: 
hadoop-common-project/hadoop-common |
| Console output | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/15249/console |
| Powered by | Apache Yetus 0.8.0   

[jira] [Commented] (HADOOP-15774) Discovery of HA servers

2018-09-21 Thread Eric Yang (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15774?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623860#comment-16623860
 ] 

Eric Yang commented on HADOOP-15774:


RegistryDNS is a tool to provide DNS resolution.  If namenode hostname follows 
certain convention, then it might be possible to provide multi-A record to 
provide information about all namenodes.  This is assuming that namdenode would 
know how to find out about location of ZooKeeper servers via configuration file 
to publish themselves.  If ZooKeeper quorum can only be discovered by 
configuration, then there is still no real service discovery.

It would be nice to move registryDNS from YARN to Hadoop common, but 
dynamically spawned ZooKeeper server is broken currently.  ZooKeeper needs to 
be fixed in the following area to meet the requirements for service discovery 
for Hadoop core system:

# ZooKeeper client session affinity.  If ZooKeeper server spawned somewhere 
else, existing client still try to connect to old IP.  ZOOKEEPER-2929
# ZooKeeper Kerberos security fails, but continue to allow ZooKeeper 
connections.  ZOOKEEPER-1634

YARN usage of ZooKeeper and RegistryDNS are less prone to ZooKeeper defects 
because assumption of ZooKeeper deployment is static, and curator handles some 
odd conditions for ZooKeeper connections.  Curator dependency might generate a 
wave of realignment for HDFS dependency on ZooKeeper to get subpar results.

Alternative approach is to use multicast DNS for daemon process that supports 
service discovery.  Similar to Apple products that can find out about devices 
in surrounding environment.  There is a implementation of service discovery in 
[this 
project|https://github.com/macroadster/HMS/blob/master/beacon/src/main/java/org/apache/hms/beacon/Beacon.java]
 for ZooKeeper. This was canned 6 years ago when [~aw] and others provided 
feedback that said multicast DNS is too expensive and possible vulnerabilities. 
 Many issues were addressed by multicast DNS implementations in the last 6 
years.  Not sure if it is worth while to revisit multicast DNS for service 
discovery purpose.

DNS or multicast DNS are most likely the best approach for providing discovery 
of HA servers.  However, the information are provided to be aware of potential 
pitfalls in the current implementations.



> Discovery of HA servers
> ---
>
> Key: HADOOP-15774
> URL: https://issues.apache.org/jira/browse/HADOOP-15774
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Íñigo Goiri
>Priority: Major
> Attachments: Discovery Service.pdf
>
>
> Currently, Hadoop relies on configuration files to specify the servers.
> This requires maintaining these configuration files and propagating the 
> changes.
> Hadoop should have a framework to provide discovery.
> For example, in HDFS, we could define the Namenodes in a shared location and 
> the DNs would use the framework to find the Namenodes.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15723) ABFS: Ranger Support

2018-09-21 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15723?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623805#comment-16623805
 ] 

Steve Loughran commented on HADOOP-15723:
-

h3. ITestAzureBlobFileSystemAuthorization

* needs ASF header (as do the mock classes)
* tests should close the (opened) files after opening/creating/appending them. 
you can simply go fs.open(path).close() for this
L129, L139: use delete(path, false) to eliminate deprecation warnings.

I don't think the unauthed tests are really stressing things properly, because 
a test like {{testSetAclUnauthorized}} is going to fail on the fs.create(), 
before the setAcl() call goes through.

I'd recommend
* having the normal setup() in place, with the FS created there being 
authorized for full R/W
* Creating a second the other filesystems (readOnly, writeOnly) filesystems in 
test setup too; closing them in teardown.
* in each test, creating all files in the r/w fs
* and only use the readonly FS for those FS operations where failures are 
expected.


{code}
getFileSystem().create(TEST_FILE_PATH)
intercept(AbfsAuthorizationException.class,
 ()-> {
   readOnlyFS.removeACL(TEST_FILE_PATH)
 });
 
 Why LambdaTestUtils.intercept()? it lets you restrict exactly where the auth 
exceptions will be expected.



> ABFS: Ranger Support
> 
>
> Key: HADOOP-15723
> URL: https://issues.apache.org/jira/browse/HADOOP-15723
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/azure
>Affects Versions: HADOOP-15407
>Reporter: Thomas Marquardt
>Assignee: Yuan Gao
>Priority: Major
> Attachments: HADOOP-15273-HADOOP-15407-001.patch, 
> HADOOP-15273-HADOOP-15407-002.patch, HADOOP-15273-HADOOP-15407-003.patch, 
> HADOOP-15273-HADOOP-15407-004.patch
>
>
> Add support for Ranger



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15760) Include Apache Commons Collections4

2018-09-21 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15760?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623771#comment-16623771
 ] 

Steve Loughran commented on HADOOP-15760:
-

LGTM. 
what's your overall migration plan?

> Include Apache Commons Collections4
> ---
>
> Key: HADOOP-15760
> URL: https://issues.apache.org/jira/browse/HADOOP-15760
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 2.10.0, 3.0.3
>Reporter: BELUGA BEHR
>Priority: Major
> Attachments: HADOOP-15760.1.patch
>
>
> Please allow for use of Apache Commons Collections 4 library with the end 
> goal of migrating from Apache Commons Collections 3.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15774) Discovery of HA servers

2018-09-21 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15774?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623767#comment-16623767
 ] 

Steve Loughran commented on HADOOP-15774:
-


bq. Would it make sense to make the registry a standalone service instead of 
running in the RM?
Being part of the YARN may bring some weird dependencies (HDFS->YARN), any 
ideas here? Should it be move to a separate place? Maybe parts?

It can run standalone —its just a subclass of 
org.apache.hadoop.utils.AbstractService with its normal lifecycle.

It's integrated in YARN so that
* the RM can set things up with its kerberos permissions.
* it can do cleanup after an app is shut down.

Provided HDFS is set up with the permissions to edit the /services part of the 
tree, it could run the ZK "manager" service internally, registering itself on 
startup. This'd be easier if the HDFS NN went over to using that same Hadoop 
util.Service lifecycle, but if not, it could still wire it up.

All you really need to keep running is ZK.


> Discovery of HA servers
> ---
>
> Key: HADOOP-15774
> URL: https://issues.apache.org/jira/browse/HADOOP-15774
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Íñigo Goiri
>Priority: Major
> Attachments: Discovery Service.pdf
>
>
> Currently, Hadoop relies on configuration files to specify the servers.
> This requires maintaining these configuration files and propagating the 
> changes.
> Hadoop should have a framework to provide discovery.
> For example, in HDFS, we could define the Namenodes in a shared location and 
> the DNs would use the framework to find the Namenodes.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15679) ShutdownHookManager shutdown time needs to be configurable & extended

2018-09-21 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15679?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623761#comment-16623761
 ] 

Steve Loughran commented on HADOOP-15679:
-

issue was that in the backport, the checkstyle warning about an unused import 
was wrong: the import was used in the javadocs. Remove the import and javadoc 
errors surface.

latest patch reinstates the import, essentially reverting to the first branch-2 
patch

> ShutdownHookManager shutdown time needs to be configurable & extended
> -
>
> Key: HADOOP-15679
> URL: https://issues.apache.org/jira/browse/HADOOP-15679
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: util
>Affects Versions: 2.8.0, 3.0.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15679-001.patch, HADOOP-15679-002.patch, 
> HADOOP-15679-002.patch, HADOOP-15679-003.patch, 
> HADOOP-15679-branch-2-001.patch, HADOOP-15679-branch-2-001.patch, 
> HADOOP-15679-branch-2-003.patch, HADOOP-15679-branch-2-003.patch, 
> HADOOP-15679-branch-2-004.patch
>
>
> HADOOP-12950 added a timeout on shutdowns to avoid problems with hanging 
> shutdowns. But the timeout is too short for applications where a large flush 
> of data is needed on shutdown.
> A key example of this is Spark apps which save their history to object 
> stores, where the file close() call triggers an upload of the final local 
> cached block of data (could be 32+MB), and then execute the final mutipart 
> commit.
> Proposed
> # make the default sleep time 30s, not 10s
> # make it configurable with a time duration property (with minimum time of 
> 1s.?)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15774) Discovery of HA servers

2018-09-21 Thread Billie Rinaldi (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15774?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623736#comment-16623736
 ] 

Billie Rinaldi commented on HADOOP-15774:
-

bq. Would it make sense to make the registry a standalone service instead of 
running in the RM?
There isn't really anything running in the RM. The registry is more of a 
convention about how things are stored in ZooKeeper under /registry. The 
Registry DNS server that hosts DNS records based on registry data runs as a 
separate daemon.

bq. Being part of the YARN may bring some weird dependencies
I think it would be possible to refactor the 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-registry module and move it to 
hadoop common if there are HDFS/YARN dependency issues.

> Discovery of HA servers
> ---
>
> Key: HADOOP-15774
> URL: https://issues.apache.org/jira/browse/HADOOP-15774
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Íñigo Goiri
>Priority: Major
> Attachments: Discovery Service.pdf
>
>
> Currently, Hadoop relies on configuration files to specify the servers.
> This requires maintaining these configuration files and propagating the 
> changes.
> Hadoop should have a framework to provide discovery.
> For example, in HDFS, we could define the Namenodes in a shared location and 
> the DNs would use the framework to find the Namenodes.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-15723) ABFS: Ranger Support

2018-09-21 Thread Yuan Gao (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15723?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623725#comment-16623725
 ] 

Yuan Gao edited comment on HADOOP-15723 at 9/21/18 2:45 PM:


patch 004:
 - called makeQualified() before authz checks
 - moved configuration and factory method to AbfsConfiguration
 - moved the authorization interface to org.apache.hadoop.fs.azurebfs.extensions
 - renamed from adl to abfs
 - added test cases for authz checks in ITestAzureBlobFileSystemAuthorization

I was able to run:

{{mvn -T 1C -Dparallel-tests -Dscale -DtestsThreadCount=8 clean verify}}
{{[INFO] Results:}}
{{[WARNING] Tests run: 277, Failures: 0, Errors: 0, Skipped: 11}}

But there are errors when running:

{{mvn -T 1C -Dparallel-tests=abfs -Dscale -DtestsThreadCount=8 clean verify}}

{{[ERROR] Tests run: 6, Failures: 0, Errors: 6, Skipped: 0, Time elapsed: 5.955 
s <<< FAILURE! - in 
org.apache.hadoop.fs.azurebfs.contract.ITestAbfsFileSystemContractDistCp}}
 {{[ERROR] 
testUpdateDeepDirectoryStructureNoChange(org.apache.hadoop.fs.azurebfs.contract.ITestAbfsFileSystemContractDistCp)
 Time elapsed: 3.038 s <<< ERROR!}}
 {{java.io.IOException: Unable to initialize filesystem 
abfs://dfsafast...@rajeevbstorageac.dfs.core.windows.net: 
java.lang.IllegalArgumentException: Incorrect URI 
abfs://dfsafast...@rajeevbstorageac.dfs.core.windows.net, URI scheme must be 
abfss when authenticating using Oauth.}}


was (Author: kowon2008):
patch 004:
 - called makeQualified() before authz checks
 - moved configuration and factory method to AbfsConfiguration
 - moved the authorization interface to org.apache.hadoop.fs.azurebfs.extensions
 - renamed from adl to abfs
 - added test cases for authz checks in ITestAzureBlobFileSystemAuthorization

I was able to run:

{{ mvn -T 1C -Dparallel-tests -Dscale -DtestsThreadCount=8 clean verify}}
{{ [INFO] Results:}}
{{ [WARNING] Tests run: 277, Failures: 0, Errors: 0, Skipped: 11}}

But there are errors when running:

{{mvn -T 1C -Dparallel-tests=abfs -Dscale -DtestsThreadCount=8 clean verify}}

{{[ERROR] Tests run: 6, Failures: 0, Errors: 6, Skipped: 0, Time elapsed: 5.955 
s <<< FAILURE! - in 
org.apache.hadoop.fs.azurebfs.contract.ITestAbfsFileSystemContractDistCp}}
{{[ERROR] 
testUpdateDeepDirectoryStructureNoChange(org.apache.hadoop.fs.azurebfs.contract.ITestAbfsFileSystemContractDistCp)
 Time elapsed: 3.038 s <<< ERROR!}}
{{java.io.IOException: Unable to initialize filesystem 
abfs://dfsafast...@rajeevbstorageac.dfs.core.windows.net: 
java.lang.IllegalArgumentException: Incorrect URI 
abfs://dfsafast...@rajeevbstorageac.dfs.core.windows.net, URI scheme must be 
abfss when authenticating using Oauth.}}

> ABFS: Ranger Support
> 
>
> Key: HADOOP-15723
> URL: https://issues.apache.org/jira/browse/HADOOP-15723
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/azure
>Affects Versions: HADOOP-15407
>Reporter: Thomas Marquardt
>Assignee: Yuan Gao
>Priority: Major
> Attachments: HADOOP-15273-HADOOP-15407-001.patch, 
> HADOOP-15273-HADOOP-15407-002.patch, HADOOP-15273-HADOOP-15407-003.patch, 
> HADOOP-15273-HADOOP-15407-004.patch
>
>
> Add support for Ranger



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15723) ABFS: Ranger Support

2018-09-21 Thread Yuan Gao (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15723?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623725#comment-16623725
 ] 

Yuan Gao commented on HADOOP-15723:
---

patch 004:
 - called makeQualified() before authz checks
 - moved configuration and factory method to AbfsConfiguration
 - moved the authorization interface to org.apache.hadoop.fs.azurebfs.extensions
 - renamed from adl to abfs
 - added test cases for authz checks in ITestAzureBlobFileSystemAuthorization

I was able to run:

{{ mvn -T 1C -Dparallel-tests -Dscale -DtestsThreadCount=8 clean verify}}
{{ [INFO] Results:}}
{{ [WARNING] Tests run: 277, Failures: 0, Errors: 0, Skipped: 11}}

But there are errors when running:

{{mvn -T 1C -Dparallel-tests=abfs -Dscale -DtestsThreadCount=8 clean verify}}

{{[ERROR] Tests run: 6, Failures: 0, Errors: 6, Skipped: 0, Time elapsed: 5.955 
s <<< FAILURE! - in 
org.apache.hadoop.fs.azurebfs.contract.ITestAbfsFileSystemContractDistCp}}
{{[ERROR] 
testUpdateDeepDirectoryStructureNoChange(org.apache.hadoop.fs.azurebfs.contract.ITestAbfsFileSystemContractDistCp)
 Time elapsed: 3.038 s <<< ERROR!}}
{{java.io.IOException: Unable to initialize filesystem 
abfs://dfsafast...@rajeevbstorageac.dfs.core.windows.net: 
java.lang.IllegalArgumentException: Incorrect URI 
abfs://dfsafast...@rajeevbstorageac.dfs.core.windows.net, URI scheme must be 
abfss when authenticating using Oauth.}}

> ABFS: Ranger Support
> 
>
> Key: HADOOP-15723
> URL: https://issues.apache.org/jira/browse/HADOOP-15723
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/azure
>Affects Versions: HADOOP-15407
>Reporter: Thomas Marquardt
>Assignee: Yuan Gao
>Priority: Major
> Attachments: HADOOP-15273-HADOOP-15407-001.patch, 
> HADOOP-15273-HADOOP-15407-002.patch, HADOOP-15273-HADOOP-15407-003.patch, 
> HADOOP-15273-HADOOP-15407-004.patch
>
>
> Add support for Ranger



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15679) ShutdownHookManager shutdown time needs to be configurable & extended

2018-09-21 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15679?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15679:

Attachment: HADOOP-15679-branch-2-004.patch

> ShutdownHookManager shutdown time needs to be configurable & extended
> -
>
> Key: HADOOP-15679
> URL: https://issues.apache.org/jira/browse/HADOOP-15679
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: util
>Affects Versions: 2.8.0, 3.0.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15679-001.patch, HADOOP-15679-002.patch, 
> HADOOP-15679-002.patch, HADOOP-15679-003.patch, 
> HADOOP-15679-branch-2-001.patch, HADOOP-15679-branch-2-001.patch, 
> HADOOP-15679-branch-2-003.patch, HADOOP-15679-branch-2-003.patch, 
> HADOOP-15679-branch-2-004.patch
>
>
> HADOOP-12950 added a timeout on shutdowns to avoid problems with hanging 
> shutdowns. But the timeout is too short for applications where a large flush 
> of data is needed on shutdown.
> A key example of this is Spark apps which save their history to object 
> stores, where the file close() call triggers an upload of the final local 
> cached block of data (could be 32+MB), and then execute the final mutipart 
> commit.
> Proposed
> # make the default sleep time 30s, not 10s
> # make it configurable with a time duration property (with minimum time of 
> 1s.?)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15679) ShutdownHookManager shutdown time needs to be configurable & extended

2018-09-21 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15679?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15679:

Target Version/s: 3.0.3, 2.10.0, 2.9.2, 2.8.5  (was: 2.10.0, 2.9.2, 3.0.3, 
2.8.5)
  Status: Patch Available  (was: Open)

> ShutdownHookManager shutdown time needs to be configurable & extended
> -
>
> Key: HADOOP-15679
> URL: https://issues.apache.org/jira/browse/HADOOP-15679
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: util
>Affects Versions: 3.0.0, 2.8.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15679-001.patch, HADOOP-15679-002.patch, 
> HADOOP-15679-002.patch, HADOOP-15679-003.patch, 
> HADOOP-15679-branch-2-001.patch, HADOOP-15679-branch-2-001.patch, 
> HADOOP-15679-branch-2-003.patch, HADOOP-15679-branch-2-003.patch, 
> HADOOP-15679-branch-2-004.patch
>
>
> HADOOP-12950 added a timeout on shutdowns to avoid problems with hanging 
> shutdowns. But the timeout is too short for applications where a large flush 
> of data is needed on shutdown.
> A key example of this is Spark apps which save their history to object 
> stores, where the file close() call triggers an upload of the final local 
> cached block of data (could be 32+MB), and then execute the final mutipart 
> commit.
> Proposed
> # make the default sleep time 30s, not 10s
> # make it configurable with a time duration property (with minimum time of 
> 1s.?)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15679) ShutdownHookManager shutdown time needs to be configurable & extended

2018-09-21 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15679?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15679:

Target Version/s: 3.0.3, 2.10.0, 2.9.2, 2.8.5  (was: 2.10.0, 2.9.2, 3.0.3, 
2.8.5)
  Status: Open  (was: Patch Available)

> ShutdownHookManager shutdown time needs to be configurable & extended
> -
>
> Key: HADOOP-15679
> URL: https://issues.apache.org/jira/browse/HADOOP-15679
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: util
>Affects Versions: 3.0.0, 2.8.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15679-001.patch, HADOOP-15679-002.patch, 
> HADOOP-15679-002.patch, HADOOP-15679-003.patch, 
> HADOOP-15679-branch-2-001.patch, HADOOP-15679-branch-2-001.patch, 
> HADOOP-15679-branch-2-003.patch, HADOOP-15679-branch-2-003.patch, 
> HADOOP-15679-branch-2-004.patch
>
>
> HADOOP-12950 added a timeout on shutdowns to avoid problems with hanging 
> shutdowns. But the timeout is too short for applications where a large flush 
> of data is needed on shutdown.
> A key example of this is Spark apps which save their history to object 
> stores, where the file close() call triggers an upload of the final local 
> cached block of data (could be 32+MB), and then execute the final mutipart 
> commit.
> Proposed
> # make the default sleep time 30s, not 10s
> # make it configurable with a time duration property (with minimum time of 
> 1s.?)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15679) ShutdownHookManager shutdown time needs to be configurable & extended

2018-09-21 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15679?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623692#comment-16623692
 ] 

Steve Loughran commented on HADOOP-15679:
-

{code}
[WARNING] 
/testptch/hadoop/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ShutdownHookManager.java:64:
 warning - Tag @link: reference not found: 
CommonConfigurationKeysPublic#SERVICE_SHUTDOWN_TIMEOUT
[WARNING] 
/testptch/hadoop/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ShutdownHookManager.java:64:
 warning - Tag @link: reference not found: 
CommonConfigurationKeysPublic#SERVICE_SHUTDOWN_TIMEOUT_DEFAULT
{code}

> ShutdownHookManager shutdown time needs to be configurable & extended
> -
>
> Key: HADOOP-15679
> URL: https://issues.apache.org/jira/browse/HADOOP-15679
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: util
>Affects Versions: 2.8.0, 3.0.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15679-001.patch, HADOOP-15679-002.patch, 
> HADOOP-15679-002.patch, HADOOP-15679-003.patch, 
> HADOOP-15679-branch-2-001.patch, HADOOP-15679-branch-2-001.patch, 
> HADOOP-15679-branch-2-003.patch, HADOOP-15679-branch-2-003.patch
>
>
> HADOOP-12950 added a timeout on shutdowns to avoid problems with hanging 
> shutdowns. But the timeout is too short for applications where a large flush 
> of data is needed on shutdown.
> A key example of this is Spark apps which save their history to object 
> stores, where the file close() call triggers an upload of the final local 
> cached block of data (could be 32+MB), and then execute the final mutipart 
> commit.
> Proposed
> # make the default sleep time 30s, not 10s
> # make it configurable with a time duration property (with minimum time of 
> 1s.?)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15691) Add PathCapabilities to FS and FC to complement StreamCapabilities

2018-09-21 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15691?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15691:

Description: 
Add a {{PathCapabilities}} interface to both FileSystem and FileContext to 
declare the capabilities under the path of an FS

This is needed for 
* HADOOP-14707: declare that a dest FS supports permissions
* object stores to declare that they offer PUT-in-place alongside (slow-rename)
* Anything else where the implementation semantics of an FS is so different 
caller apps would benefit from probing for the underlying semantics

I know, we want all filesystem to work *exactly* the same. But it doesn't hold, 
especially for object stores —and to efficiently use them, callers need to be 
able to ask for specific features.

  was:
Add a {{PathCapabilities}} interface to both FileSystem and FileContext to 
declare the capabilities under the path of an FS

This is needed for 
* HADOOP-15407: declare that a dest FS supports permissions
* object stores to declare that they offer PUT-in-place alongside (slow-rename)
* Anything else where the implementation semantics of an FS is so different 
caller apps would benefit from probing for the underlying semantics

I know, we want all filesystem to work *exactly* the same. But it doesn't hold, 
especially for object stores —and to efficiently use them, callers need to be 
able to ask for specific features.


> Add PathCapabilities to FS and FC to complement StreamCapabilities
> --
>
> Key: HADOOP-15691
> URL: https://issues.apache.org/jira/browse/HADOOP-15691
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15691-001.patch, HADOOP-15691-002.patch, 
> HADOOP-15691-003.patch
>
>
> Add a {{PathCapabilities}} interface to both FileSystem and FileContext to 
> declare the capabilities under the path of an FS
> This is needed for 
> * HADOOP-14707: declare that a dest FS supports permissions
> * object stores to declare that they offer PUT-in-place alongside 
> (slow-rename)
> * Anything else where the implementation semantics of an FS is so different 
> caller apps would benefit from probing for the underlying semantics
> I know, we want all filesystem to work *exactly* the same. But it doesn't 
> hold, especially for object stores —and to efficiently use them, callers need 
> to be able to ask for specific features.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15691) Add PathCapabilities to FS and FC to complement StreamCapabilities

2018-09-21 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15691?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15691:

Status: Patch Available  (was: Open)

> Add PathCapabilities to FS and FC to complement StreamCapabilities
> --
>
> Key: HADOOP-15691
> URL: https://issues.apache.org/jira/browse/HADOOP-15691
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15691-001.patch, HADOOP-15691-002.patch, 
> HADOOP-15691-003.patch
>
>
> Add a {{PathCapabilities}} interface to both FileSystem and FileContext to 
> declare the capabilities under the path of an FS
> This is needed for 
> * HADOOP-15407: declare that a dest FS supports permissions
> * object stores to declare that they offer PUT-in-place alongside 
> (slow-rename)
> * Anything else where the implementation semantics of an FS is so different 
> caller apps would benefit from probing for the underlying semantics
> I know, we want all filesystem to work *exactly* the same. But it doesn't 
> hold, especially for object stores —and to efficiently use them, callers need 
> to be able to ask for specific features.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15691) Add PathCapabilities to FS and FC to complement StreamCapabilities

2018-09-21 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15691?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623684#comment-16623684
 ] 

Steve Loughran commented on HADOOP-15691:
-

Patch 003

* revise based on feedback
* catch up with trunk (pathhandles)
* add option for an FS connector to say "We know we are read only" (http, har)
* add checksum capability, wire up so those filesystems with it (DFS, S3A when 
its enabled)
* symlink capability is based on {{symlinksEnabled}} flag in FileSystem class.
* expanded docs
* ~0 tests

I now realise that a bool "true/false" isn't quite right, as it doesn't cover 
the "we think we'd support it but you'd have to try". Symlinks being a key 
example. There's DFS: we know we have it, other stores "we know we don't", but 
localFS is either "known disabled" or "try it and see"

An enum seems the outcome here, something like
{code}
known-available: its a known feature and known to work in this instance (ex: 
symlinks on hdfs)
known-unavailable: we know of it, we know we don't support it (ex: append on 
s3a)
known-maybe-available: we know of it, but to see if it works, you will have to 
try the operation. (symlink on local)
unknown: we don't know of this feature.
{code}

the base FS would say "unknown feature" for ~everything; implementations would 
list what they know about and their interpretation, so that everything with 
rename would add "fs:rename"; posix-ish filesytems with atomic rename, 
"fs-rename-atomic"; the object stores would mark the latter as 
"known-unavailable", etc.

We'd also need to qualify everything with "depending on your permissions". An 
FS may know of append, but it might be a read-only mount. I've already added to 
filesystem.md the fact that the state may change depending on external events.



> Add PathCapabilities to FS and FC to complement StreamCapabilities
> --
>
> Key: HADOOP-15691
> URL: https://issues.apache.org/jira/browse/HADOOP-15691
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15691-001.patch, HADOOP-15691-002.patch, 
> HADOOP-15691-003.patch
>
>
> Add a {{PathCapabilities}} interface to both FileSystem and FileContext to 
> declare the capabilities under the path of an FS
> This is needed for 
> * HADOOP-15407: declare that a dest FS supports permissions
> * object stores to declare that they offer PUT-in-place alongside 
> (slow-rename)
> * Anything else where the implementation semantics of an FS is so different 
> caller apps would benefit from probing for the underlying semantics
> I know, we want all filesystem to work *exactly* the same. But it doesn't 
> hold, especially for object stores —and to efficiently use them, callers need 
> to be able to ask for specific features.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15691) Add PathCapabilities to FS and FC to complement StreamCapabilities

2018-09-21 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15691?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15691:

Attachment: HADOOP-15691-003.patch

> Add PathCapabilities to FS and FC to complement StreamCapabilities
> --
>
> Key: HADOOP-15691
> URL: https://issues.apache.org/jira/browse/HADOOP-15691
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15691-001.patch, HADOOP-15691-002.patch, 
> HADOOP-15691-003.patch
>
>
> Add a {{PathCapabilities}} interface to both FileSystem and FileContext to 
> declare the capabilities under the path of an FS
> This is needed for 
> * HADOOP-15407: declare that a dest FS supports permissions
> * object stores to declare that they offer PUT-in-place alongside 
> (slow-rename)
> * Anything else where the implementation semantics of an FS is so different 
> caller apps would benefit from probing for the underlying semantics
> I know, we want all filesystem to work *exactly* the same. But it doesn't 
> hold, especially for object stores —and to efficiently use them, callers need 
> to be able to ask for specific features.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15566) Remove HTrace support

2018-09-21 Thread Carlos Alberto Cortez (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15566?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623548#comment-16623548
 ] 

Carlos Alberto Cortez commented on HADOOP-15566:


Hi all,

I went ahead and did a Proof-Of-Concept migration from HTrace to OpenTracing 
(using Zipkin as the backend). You can inspect (and play with the code) here: 
[https://github.com/apache/hadoop/compare/trunk...carlosalberto:ot_initial_integration]

Some notes:

1. It creates by default a Tracer instance based on Zipkin running in localhost 
(for simplicity purposes).
2. It uses the notion of a GlobalTracer so create and register and use the 
Tracer from a single place.
3. As Wei-Chiu mentioned, it needed some small extra work to pass around the 
parent-child relationship (which is done trough `SpanId` in HTrace, and 
`SpanContext` in OpenTracing).
4. Added a new SpanContext field in the clases using protobuf to pass trace 
info.

As mentioned, this is a POC, but hope this can throw light into this (and I'm 
happy to answer questions or contribute with this as an actual migration ;) )

> Remove HTrace support
> -
>
> Key: HADOOP-15566
> URL: https://issues.apache.org/jira/browse/HADOOP-15566
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: metrics
>Affects Versions: 3.1.0
>Reporter: Todd Lipcon
>Priority: Major
>  Labels: security
> Attachments: Screen Shot 2018-06-29 at 11.59.16 AM.png, 
> ss-trace-s3a.png
>
>
> The HTrace incubator project has voted to retire itself and won't be making 
> further releases. The Hadoop project currently has various hooks with HTrace. 
> It seems in some cases (eg HDFS-13702) these hooks have had measurable 
> performance overhead. Given these two factors, I think we should consider 
> removing the HTrace integration. If there is someone willing to do the work, 
> replacing it with OpenTracing might be a better choice since there is an 
> active community.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-8977) multiple FsShell test failures on Windows

2018-09-21 Thread Vinayakumar B (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-8977?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623505#comment-16623505
 ] 

Vinayakumar B commented on HADOOP-8977:
---

HDFS-13911 seems to be broken by this change.

After this change, all local paths with special characters (ex: space) should 
be encoded first before passing as argument to command line.

> multiple FsShell test failures on Windows
> -
>
> Key: HADOOP-8977
> URL: https://issues.apache.org/jira/browse/HADOOP-8977
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Affects Versions: trunk-win
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
>Priority: Major
> Fix For: trunk-win
>
> Attachments: HADOOP-8977-branch-trunk-win.patch, 
> HADOOP-8977-branch-trunk-win.patch, HADOOP-8977-branch-trunk-win.patch, 
> HADOOP-8977.patch
>
>
> Multiple FsShell-related tests fail on Windows.  Commands are returning 
> non-zero exit status.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15775) [JDK9] Add missing javax.activation-api dependency

2018-09-21 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15775?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623491#comment-16623491
 ] 

Hadoop QA commented on HADOOP-15775:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
27s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:red}-1{color} | {color:red} test4tests {color} | {color:red}  0m  
0s{color} | {color:red} The patch doesn't appear to include any new or modified 
tests. Please justify why no new tests are needed for this patch. Also please 
list what manual steps were performed to verify this patch. {color} |
|| || || || {color:brown} trunk Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  5m 
28s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 17m 
30s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 14m 
34s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  2m  
3s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
49m 50s{color} | {color:green} branch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
35s{color} | {color:green} trunk passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
22s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  4m 
10s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 13m 
47s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 13m 
47s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  2m 
14s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
4s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
10m 26s{color} | {color:green} patch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
27s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 18m 58s{color} 
| {color:red} hadoop-yarn-server-nodemanager in the patch failed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 75m 
20s{color} | {color:green} hadoop-yarn-server-resourcemanager in the patch 
passed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  0m 
29s{color} | {color:green} hadoop-client-minicluster in the patch passed. 
{color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
41s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}181m 24s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | 
hadoop.yarn.server.nodemanager.containermanager.TestNMProxy |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:4b8c2b1 |
| JIRA Issue | HADOOP-15775 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12940749/HADOOP-15775.03.patch 
|
| Optional Tests |  dupname  asflicense  compile  javac  javadoc  mvninstall  
mvnsite  unit  shadedclient  xml  |
| uname | Linux f80156c6aac8 4.4.0-133-generic #159-Ubuntu SMP Fri Aug 10 
07:31:43 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/patchprocess/precommit/personality/provided.sh |
| git revision | trunk / 524f7cd |
| maven | version: Apache Maven 3.3.9 |
| Default Java | 1.8.0_181 |
| unit | 

[jira] [Commented] (HADOOP-15691) Add PathCapabilities to FS and FC to complement StreamCapabilities

2018-09-21 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15691?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623450#comment-16623450
 ] 

Steve Loughran commented on HADOOP-15691:
-

bq. Should we document a well defined capabilities? For example, we can say in 
Hadoop 3.3, we define the following list of path capabilities (like 
fs:snapshots) and file system implementations should support probing of these 
capabilities. Right now these capabilities exist in the code, not in user docs.

bq. Let's leave to the javadocs, so that its more likely to be in sync with the 
code. We can add a ref to it though in the FS docs.

bq. Instead of defining the capabilities in StreamCapabilities.java, why not 
create a PathCapabilities class for these capability definitions?

I'd just added them alongside the existing ones, but yes, it makes sense. Will 
do this.

bq. Are these capabilities well defined somewhere? I am not a file system guy 
so I am not sure if people interpret these capabilities differently.

Not yet, we can be specific in the javadocs


> Add PathCapabilities to FS and FC to complement StreamCapabilities
> --
>
> Key: HADOOP-15691
> URL: https://issues.apache.org/jira/browse/HADOOP-15691
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15691-001.patch, HADOOP-15691-002.patch
>
>
> Add a {{PathCapabilities}} interface to both FileSystem and FileContext to 
> declare the capabilities under the path of an FS
> This is needed for 
> * HADOOP-15407: declare that a dest FS supports permissions
> * object stores to declare that they offer PUT-in-place alongside 
> (slow-rename)
> * Anything else where the implementation semantics of an FS is so different 
> caller apps would benefit from probing for the underlying semantics
> I know, we want all filesystem to work *exactly* the same. But it doesn't 
> hold, especially for object stores —and to efficiently use them, callers need 
> to be able to ask for specific features.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15679) ShutdownHookManager shutdown time needs to be configurable & extended

2018-09-21 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15679?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623447#comment-16623447
 ] 

Hadoop QA commented on HADOOP-15679:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 15m 
31s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m  
1s{color} | {color:blue} Findbugs executables are not available. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
|| || || || {color:brown} branch-2 Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 14m 
 9s{color} | {color:green} branch-2 passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 12m  
2s{color} | {color:green} branch-2 passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
36s{color} | {color:green} branch-2 passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
12s{color} | {color:green} branch-2 passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
4s{color} | {color:green} branch-2 passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
47s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 11m 
10s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 11m 
10s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
35s{color} | {color:green} hadoop-common-project/hadoop-common: The patch 
generated 0 new + 115 unchanged - 1 fixed = 115 total (was 116) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
11s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
1s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:red}-1{color} | {color:red} javadoc {color} | {color:red}  1m  
1s{color} | {color:red} hadoop-common-project_hadoop-common generated 2 new + 
10 unchanged - 0 fixed = 12 total (was 10) {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  8m 
41s{color} | {color:green} hadoop-common in the patch passed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
42s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 70m 16s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:a716388 |
| JIRA Issue | HADOOP-15679 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12940760/HADOOP-15679-branch-2-003.patch
 |
| Optional Tests |  dupname  asflicense  compile  javac  javadoc  mvninstall  
mvnsite  unit  shadedclient  findbugs  checkstyle  xml  |
| uname | Linux e48cccf61331 3.13.0-139-generic #188-Ubuntu SMP Tue Jan 9 
14:43:09 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/patchprocess/precommit/personality/provided.sh |
| git revision | branch-2 / 3a6ad9c |
| maven | version: Apache Maven 3.3.9 
(bb52d8502b132ec0a5a3f4c09453c07478323dc5; 2015-11-10T16:41:47+00:00) |
| Default Java | 1.7.0_181 |
| javadoc | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/15247/artifact/out/diff-javadoc-javadoc-hadoop-common-project_hadoop-common.txt
 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/15247/testReport/ |
| Max. process+thread count | 1286 (vs. ulimit of 1) |
| modules | C: hadoop-common-project/hadoop-common U: 
hadoop-common-project/hadoop-common |
| Console output | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/15247/console |
| Powered by | Apache Yetus 0.8.0   http://yetus.apache.org |


This message was automatically generated.



> ShutdownHookManager shutdown time needs to be configurable & extended
> 

[jira] [Commented] (HADOOP-15775) [JDK9] Add missing javax.activation-api dependency

2018-09-21 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15775?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623443#comment-16623443
 ] 

Hadoop QA commented on HADOOP-15775:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
29s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:red}-1{color} | {color:red} test4tests {color} | {color:red}  0m  
0s{color} | {color:red} The patch doesn't appear to include any new or modified 
tests. Please justify why no new tests are needed for this patch. Also please 
list what manual steps were performed to verify this patch. {color} |
|| || || || {color:brown} trunk Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  3m  
7s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 25m 
21s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  3m 
12s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
40s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
44m 48s{color} | {color:green} branch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
1s{color} | {color:green} trunk passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
14s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
34s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  3m  
6s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  3m  
6s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
28s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
3s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:red}-1{color} | {color:red} shadedclient {color} | {color:red} 12m 
44s{color} | {color:red} patch has errors when building and testing our client 
artifacts. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
54s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 19m 49s{color} 
| {color:red} hadoop-yarn-server-nodemanager in the patch failed. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 81m 16s{color} 
| {color:red} hadoop-yarn-server-resourcemanager in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
28s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}168m 51s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | 
hadoop.yarn.server.nodemanager.containermanager.TestNMProxy |
|   | 
hadoop.yarn.server.resourcemanager.scheduler.capacity.TestQueueManagementDynamicEditPolicy
 |
|   | hadoop.yarn.server.resourcemanager.TestApplicationMasterService |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:4b8c2b1 |
| JIRA Issue | HADOOP-15775 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12940745/HADOOP-15775.02.patch 
|
| Optional Tests |  dupname  asflicense  compile  javac  javadoc  mvninstall  
mvnsite  unit  shadedclient  xml  |
| uname | Linux db4dfd446c98 4.4.0-133-generic #159-Ubuntu SMP Fri Aug 10 
07:31:43 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/patchprocess/precommit/personality/provided.sh |
| git revision | trunk / 524f7cd |
| maven | version: Apache Maven 3.3.9 |
| Default Java | 1.8.0_181 |
| unit | 

[jira] [Updated] (HADOOP-15662) ABFS: Better exception handling of DNS errors

2018-09-21 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15662?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15662:

Parent Issue: HADOOP-15763  (was: HADOOP-15407)

> ABFS: Better exception handling of DNS errors
> -
>
> Key: HADOOP-15662
> URL: https://issues.apache.org/jira/browse/HADOOP-15662
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/azure
>Reporter: Thomas Marquardt
>Assignee: Da Zhou
>Priority: Major
>
> DNS errors are common during testing due to typos or misconfiguration.  They 
> can also occur in production, as some transient DNS issues occur from time to 
> time. 
> 1) Let's investigate if we can distinguish between the two and fail fast for 
> the test issues, but continue to have retry logic for the transient DNS 
> issues in production.
> 2) Let's improve the error handling of DNS failures, so the user has an 
> actionable error message.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15778) ABFS: Fix client side throttling for read

2018-09-21 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15778?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15778:

   Resolution: Fixed
Fix Version/s: HADOOP-15407
   Status: Resolved  (was: Patch Available)

+1
committed. Thanks!

> ABFS: Fix client side throttling for read
> -
>
> Key: HADOOP-15778
> URL: https://issues.apache.org/jira/browse/HADOOP-15778
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Sneha Varma
>Assignee: Sneha Varma
>Priority: Minor
> Fix For: HADOOP-15407
>
> Attachments: HADOOP-15778-HADOOP-15407-001.patch, 
> HADOOP-15778-HADOOP-15407-002.patch
>
>
> 1. The content length for ReadFile in updateMetrics of 
> AbfsClientThrottlingIntercept is incorrect for cases when the request fails.
> It is currently equal to the number of bytes that are read whereas it should 
> be equal to the number of bytes requested.
> 2. sendingRequest of AbfsClientThrottlingIntercept at AbfsRestOperation 
> should be called irrespective of  whether the request has body.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15779) S3guard: add inconsistency detection metrics

2018-09-21 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15779?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623362#comment-16623362
 ] 

Steve Loughran commented on HADOOP-15779:
-

this'd be good. I could imagine pumping out (CSV? Avro) logs of the events to 
the yarn log dir, for aggregation and analysis

> S3guard: add inconsistency detection metrics
> 
>
> Key: HADOOP-15779
> URL: https://issues.apache.org/jira/browse/HADOOP-15779
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.2.0
>Reporter: Aaron Fabbri
>Priority: Major
>
> S3Guard uses a trailing log of metadata changes made to an S3 bucket to add 
> consistency to the eventually-consistent AWS S3 service. We should add some 
> metrics that are incremented when we detect inconsistency (eventual 
> consistency) in S3.
> I'm thinking at least two counters: (1) getFileStatus() (HEAD) inconsistency 
> detected, and (2) listing inconsistency detected. We may want to further 
> separate into categories (present / not present etc.)
> This is related to Auth. Mode and TTL work that is ongoing, so let me outline 
> how I think this should all evolve:
> This should happen after HADOOP-15621 (TTL for dynamo MetadataStore), since 
> that will change *when* we query both S3 and the MetadataStore (e.g. Dynamo) 
> for metadata. There I suggest that:
>  # Prune time is different than TTL. Prune time: "how long until 
> inconsistency is no longer a problem" . TTL time "how long a MetadataStore 
> entry is considered authoritative before refresh"
>  # Prune expired: delete entries (when hadoop CLI prune command is run). TTL 
> Expired: entries become non-authoritative.
>  #  Prune implemented in each MetadataStore, but TTL filtering happens in S3A.
> Once we have this, S3A will be consulting both S3 and MetadataStore depending 
> on configuration and/or age of the entry in the MetadataStore. Today 
> HEAD/getFileStatus() is always short-circuit (skips S3 if MetadataStore 
> returns results). I think S3A should consult both when TTL is stale, and 
> invoke a callback on inconsistency that increments the new metrics. For 
> listing, we already are comparing both sources of truth (except when S3A auth 
> mode is on and a directory is marked authoritative in MS), so it would be 
> pretty simple to invoke a callback on inconsistency and bump some metrics.
> Comments / suggestions / questions welcomed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15679) ShutdownHookManager shutdown time needs to be configurable & extended

2018-09-21 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15679?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15679:

Target Version/s: 3.0.3, 2.10.0, 2.9.2, 2.8.5  (was: 2.10.0, 2.9.2, 3.0.3, 
2.8.5)
  Status: Patch Available  (was: Open)

> ShutdownHookManager shutdown time needs to be configurable & extended
> -
>
> Key: HADOOP-15679
> URL: https://issues.apache.org/jira/browse/HADOOP-15679
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: util
>Affects Versions: 3.0.0, 2.8.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15679-001.patch, HADOOP-15679-002.patch, 
> HADOOP-15679-002.patch, HADOOP-15679-003.patch, 
> HADOOP-15679-branch-2-001.patch, HADOOP-15679-branch-2-001.patch, 
> HADOOP-15679-branch-2-003.patch, HADOOP-15679-branch-2-003.patch
>
>
> HADOOP-12950 added a timeout on shutdowns to avoid problems with hanging 
> shutdowns. But the timeout is too short for applications where a large flush 
> of data is needed on shutdown.
> A key example of this is Spark apps which save their history to object 
> stores, where the file close() call triggers an upload of the final local 
> cached block of data (could be 32+MB), and then execute the final mutipart 
> commit.
> Proposed
> # make the default sleep time 30s, not 10s
> # make it configurable with a time duration property (with minimum time of 
> 1s.?)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15679) ShutdownHookManager shutdown time needs to be configurable & extended

2018-09-21 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15679?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15679:

Target Version/s: 3.0.3, 2.10.0, 2.9.2, 2.8.5  (was: 2.10.0, 2.9.2, 3.0.3, 
2.8.5)
  Status: Open  (was: Patch Available)

> ShutdownHookManager shutdown time needs to be configurable & extended
> -
>
> Key: HADOOP-15679
> URL: https://issues.apache.org/jira/browse/HADOOP-15679
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: util
>Affects Versions: 3.0.0, 2.8.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15679-001.patch, HADOOP-15679-002.patch, 
> HADOOP-15679-002.patch, HADOOP-15679-003.patch, 
> HADOOP-15679-branch-2-001.patch, HADOOP-15679-branch-2-001.patch, 
> HADOOP-15679-branch-2-003.patch, HADOOP-15679-branch-2-003.patch
>
>
> HADOOP-12950 added a timeout on shutdowns to avoid problems with hanging 
> shutdowns. But the timeout is too short for applications where a large flush 
> of data is needed on shutdown.
> A key example of this is Spark apps which save their history to object 
> stores, where the file close() call triggers an upload of the final local 
> cached block of data (could be 32+MB), and then execute the final mutipart 
> commit.
> Proposed
> # make the default sleep time 30s, not 10s
> # make it configurable with a time duration property (with minimum time of 
> 1s.?)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15679) ShutdownHookManager shutdown time needs to be configurable & extended

2018-09-21 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15679?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15679:

Attachment: HADOOP-15679-branch-2-003.patch

> ShutdownHookManager shutdown time needs to be configurable & extended
> -
>
> Key: HADOOP-15679
> URL: https://issues.apache.org/jira/browse/HADOOP-15679
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: util
>Affects Versions: 2.8.0, 3.0.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15679-001.patch, HADOOP-15679-002.patch, 
> HADOOP-15679-002.patch, HADOOP-15679-003.patch, 
> HADOOP-15679-branch-2-001.patch, HADOOP-15679-branch-2-001.patch, 
> HADOOP-15679-branch-2-003.patch, HADOOP-15679-branch-2-003.patch
>
>
> HADOOP-12950 added a timeout on shutdowns to avoid problems with hanging 
> shutdowns. But the timeout is too short for applications where a large flush 
> of data is needed on shutdown.
> A key example of this is Spark apps which save their history to object 
> stores, where the file close() call triggers an upload of the final local 
> cached block of data (could be 32+MB), and then execute the final mutipart 
> commit.
> Proposed
> # make the default sleep time 30s, not 10s
> # make it configurable with a time duration property (with minimum time of 
> 1s.?)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15711) Fix branch-2 builds

2018-09-21 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15711?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623338#comment-16623338
 ] 

Steve Loughran commented on HADOOP-15711:
-

bq. My guess is something in hadoop-common (and FileSystem in particular) that 
doesn't work well on JDK7. 

something in the native code binding? As I can't imagine much else

> Fix branch-2 builds
> ---
>
> Key: HADOOP-15711
> URL: https://issues.apache.org/jira/browse/HADOOP-15711
> Project: Hadoop Common
>  Issue Type: Task
>Reporter: Jonathan Hung
>Priority: Critical
> Attachments: HADOOP-15711.001.branch-2.patch
>
>
> Branch-2 builds have been disabled for a while: 
> https://builds.apache.org/view/H-L/view/Hadoop/job/hadoop-qbt-branch2-java7-linux-x86/
> A test run here causes hdfs tests to hang: 
> https://builds.apache.org/view/H-L/view/Hadoop/job/hadoop-qbt-branch2-java7-linux-x86-jhung/4/
> Running hadoop-hdfs tests locally reveal some errors such 
> as:{noformat}[ERROR] 
> testComplexAppend2(org.apache.hadoop.hdfs.TestFileAppend2)  Time elapsed: 
> 0.059 s  <<< ERROR!
> java.lang.OutOfMemoryError: unable to create new native thread
> at java.lang.Thread.start0(Native Method)
> at java.lang.Thread.start(Thread.java:714)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1164)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1128)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:174)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1172)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:403)
> at 
> org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:234)
> at 
> org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:1080)
> at 
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:883)
> at 
> org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514)
> at 
> org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473)
> at 
> org.apache.hadoop.hdfs.TestFileAppend2.testComplexAppend(TestFileAppend2.java:489)
> at 
> org.apache.hadoop.hdfs.TestFileAppend2.testComplexAppend2(TestFileAppend2.java:543)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){noformat}
> I was able to get more tests passing locally by increasing the max user 
> process count on my machine. But the error suggests that there's an issue in 
> the tests themselves. Not sure if the error seen locally is the same reason 
> as why jenkins builds are failing, I wasn't able to confirm based on the 
> jenkins builds' lack of output.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15735) backport HADOOP-11687 for hadoop-aws 2.7.x

2018-09-21 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15735?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15735:

   Resolution: Fixed
Fix Version/s: 2.7.8
   Status: Resolved  (was: Patch Available)

thanks for doing this, especially the testing. Fix will be in 2.7.8

Note that spark rolled back from 2.7.7 to 2.7.6 because part of a security 
lockdown broke things...that will need to be addressed in the same release 
before they can move up

> backport HADOOP-11687 for hadoop-aws 2.7.x
> --
>
> Key: HADOOP-15735
> URL: https://issues.apache.org/jira/browse/HADOOP-15735
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.7.7
>Reporter: MunShik JOUNG
>Assignee: MunShik JOUNG
>Priority: Trivial
> Fix For: 2.7.8
>
> Attachments: 
> 0001-HADOOP-12970-Intermitent-SignatureDoesNotMatch-Excep.patch
>
>
> HADOOP-12970 Intermitent SignatureDoesNotMatch Exception is fixed by 
> HADOOP-11687. 
> And from hadoop-aws >= 2.8, HADOOP-11687 is applied. but Spark 2.3.1 is still 
> using hadoop-aws 2.7.7. 
> I was stumpled upon Intermitent SignatureDoesNotMatch Exception with s3 with 
> spark 2.3.1. So I guess backporting might benefit some spark users. In my 
> situation, SignatureDoesNotMatch was reliably reproducible (4 out of 8 
> attemp). And after using patched version symptom was not seen (0 out of 10)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Assigned] (HADOOP-15735) backport HADOOP-11687 for hadoop-aws 2.7.x

2018-09-21 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15735?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran reassigned HADOOP-15735:
---

Assignee: MunShik JOUNG

> backport HADOOP-11687 for hadoop-aws 2.7.x
> --
>
> Key: HADOOP-15735
> URL: https://issues.apache.org/jira/browse/HADOOP-15735
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.7.7
>Reporter: MunShik JOUNG
>Assignee: MunShik JOUNG
>Priority: Trivial
> Attachments: 
> 0001-HADOOP-12970-Intermitent-SignatureDoesNotMatch-Excep.patch
>
>
> HADOOP-12970 Intermitent SignatureDoesNotMatch Exception is fixed by 
> HADOOP-11687. 
> And from hadoop-aws >= 2.8, HADOOP-11687 is applied. but Spark 2.3.1 is still 
> using hadoop-aws 2.7.7. 
> I was stumpled upon Intermitent SignatureDoesNotMatch Exception with s3 with 
> spark 2.3.1. So I guess backporting might benefit some spark users. In my 
> situation, SignatureDoesNotMatch was reliably reproducible (4 out of 8 
> attemp). And after using patched version symptom was not seen (0 out of 10)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15775) [JDK9] Add missing javax.activation-api dependency

2018-09-21 Thread Akira Ajisaka (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15775?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira Ajisaka updated HADOOP-15775:
---
Summary: [JDK9] Add missing javax.activation-api dependency  (was: 
[JDK9] Add "--add-modules java.activation" to java option in unit tests)
Description: Many unit tests fail due to missing java.activation module. 
This failure can be fixed by adding javax.activation-api as third-party 
dependency.  (was: Many unit tests fail due to missing java.activation module. 
We need to configure maven surefire plugin to add "--add-modules 
java.activation" option to test JVM.)

> [JDK9] Add missing javax.activation-api dependency
> --
>
> Key: HADOOP-15775
> URL: https://issues.apache.org/jira/browse/HADOOP-15775
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: test
>Reporter: Akira Ajisaka
>Assignee: Akira Ajisaka
>Priority: Critical
> Attachments: HADOOP-15775.01.patch, HADOOP-15775.02.patch, 
> HADOOP-15775.03.patch
>
>
> Many unit tests fail due to missing java.activation module. This failure can 
> be fixed by adding javax.activation-api as third-party dependency.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15775) [JDK9] Add "--add-modules java.activation" to java option in unit tests

2018-09-21 Thread Akira Ajisaka (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15775?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623302#comment-16623302
 ] 

Akira Ajisaka commented on HADOOP-15775:


03 patch
* Shaded the third-party dependency from hadoop-client-minicluster

> [JDK9] Add "--add-modules java.activation" to java option in unit tests
> ---
>
> Key: HADOOP-15775
> URL: https://issues.apache.org/jira/browse/HADOOP-15775
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: test
>Reporter: Akira Ajisaka
>Assignee: Akira Ajisaka
>Priority: Critical
> Attachments: HADOOP-15775.01.patch, HADOOP-15775.02.patch, 
> HADOOP-15775.03.patch
>
>
> Many unit tests fail due to missing java.activation module. We need to 
> configure maven surefire plugin to add "--add-modules java.activation" option 
> to test JVM.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15775) [JDK9] Add "--add-modules java.activation" to java option in unit tests

2018-09-21 Thread Akira Ajisaka (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15775?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira Ajisaka updated HADOOP-15775:
---
Attachment: HADOOP-15775.03.patch

> [JDK9] Add "--add-modules java.activation" to java option in unit tests
> ---
>
> Key: HADOOP-15775
> URL: https://issues.apache.org/jira/browse/HADOOP-15775
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: test
>Reporter: Akira Ajisaka
>Assignee: Akira Ajisaka
>Priority: Critical
> Attachments: HADOOP-15775.01.patch, HADOOP-15775.02.patch, 
> HADOOP-15775.03.patch
>
>
> Many unit tests fail due to missing java.activation module. We need to 
> configure maven surefire plugin to add "--add-modules java.activation" option 
> to test JVM.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15711) Fix branch-2 builds

2018-09-21 Thread Allen Wittenauer (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15711?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623288#comment-16623288
 ] 

Allen Wittenauer commented on HADOOP-15711:
---

bq. Can we disable parallel tests for full builds then ?

I don't even want to contemplate the runtime.  With parallelization, it is 
already in the 14-15 hour mark, easily.  

[BTW, have a client that was using Yetus internally for 2.7.  As a result, one 
of the changes in Yetus 0.8.0 was to disable parallelization for 2.7 and lower 
since those aren't thread safe anyway.  That's why the 2.7 run timed out after 
20 hours. ]

bq. Will that help track down specific tests that go OOM for instance ?

Probably not, and definitely not without also setting the test order.  Also: no 
one has really done the homework to figure out why things are going catatonic.  
My guess is something in hadoop-common (and FileSystem in particular) that 
doesn't work well on JDK7.  That's why trunk doesn't have these issues.  My 
other hunch is that commit date of the problematic code is was committed 
between June 2016-June 2017.  Because we had other issues and no one was aware 
of how bad SUREFIRE-524 was, the issues had been hidden.

> Fix branch-2 builds
> ---
>
> Key: HADOOP-15711
> URL: https://issues.apache.org/jira/browse/HADOOP-15711
> Project: Hadoop Common
>  Issue Type: Task
>Reporter: Jonathan Hung
>Priority: Critical
> Attachments: HADOOP-15711.001.branch-2.patch
>
>
> Branch-2 builds have been disabled for a while: 
> https://builds.apache.org/view/H-L/view/Hadoop/job/hadoop-qbt-branch2-java7-linux-x86/
> A test run here causes hdfs tests to hang: 
> https://builds.apache.org/view/H-L/view/Hadoop/job/hadoop-qbt-branch2-java7-linux-x86-jhung/4/
> Running hadoop-hdfs tests locally reveal some errors such 
> as:{noformat}[ERROR] 
> testComplexAppend2(org.apache.hadoop.hdfs.TestFileAppend2)  Time elapsed: 
> 0.059 s  <<< ERROR!
> java.lang.OutOfMemoryError: unable to create new native thread
> at java.lang.Thread.start0(Native Method)
> at java.lang.Thread.start(Thread.java:714)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1164)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1128)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:174)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1172)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:403)
> at 
> org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:234)
> at 
> org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:1080)
> at 
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:883)
> at 
> org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514)
> at 
> org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473)
> at 
> org.apache.hadoop.hdfs.TestFileAppend2.testComplexAppend(TestFileAppend2.java:489)
> at 
> org.apache.hadoop.hdfs.TestFileAppend2.testComplexAppend2(TestFileAppend2.java:543)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){noformat}
> I was able to get more tests passing locally by increasing the max user 
> process count on my machine. But the error suggests that there's an issue in 
> the tests themselves. Not sure if the error seen locally is the same reason 
> as why jenkins builds are failing, I wasn't able to confirm based on the 
> jenkins builds' lack of output.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15774) Discovery of HA servers

2018-09-21 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15774?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623272#comment-16623272
 ] 

Steve Loughran commented on HADOOP-15774:
-

even though it says YARN, is explicitly not depending on any of the YARN 
libraries; it's always had a vision (from Eric14) to be broader than this. 
Essentially: discovery and binding to dynamically deployed applications, with 
Kerberos in the mix for security. For things YARN schedules, this has always 
been essential; but if you want dynamic deployment of core Hadoop services, it 
should work too. If not, something to fix

 

Pull [~sanjay.radia] [~billie.rinaldi] &[~elserj] for their insight here too.

> Discovery of HA servers
> ---
>
> Key: HADOOP-15774
> URL: https://issues.apache.org/jira/browse/HADOOP-15774
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Íñigo Goiri
>Priority: Major
> Attachments: Discovery Service.pdf
>
>
> Currently, Hadoop relies on configuration files to specify the servers.
> This requires maintaining these configuration files and propagating the 
> changes.
> Hadoop should have a framework to provide discovery.
> For example, in HDFS, we could define the Namenodes in a shared location and 
> the DNs would use the framework to find the Namenodes.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15775) [JDK9] Add "--add-modules java.activation" to java option in unit tests

2018-09-21 Thread Akira Ajisaka (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15775?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623259#comment-16623259
 ] 

Akira Ajisaka commented on HADOOP-15775:


02 patch:
* Add javax.activation-api dependency.
* TODO: Shade this dependency.

> [JDK9] Add "--add-modules java.activation" to java option in unit tests
> ---
>
> Key: HADOOP-15775
> URL: https://issues.apache.org/jira/browse/HADOOP-15775
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: test
>Reporter: Akira Ajisaka
>Assignee: Akira Ajisaka
>Priority: Critical
> Attachments: HADOOP-15775.01.patch, HADOOP-15775.02.patch
>
>
> Many unit tests fail due to missing java.activation module. We need to 
> configure maven surefire plugin to add "--add-modules java.activation" option 
> to test JVM.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15775) [JDK9] Add "--add-modules java.activation" to java option in unit tests

2018-09-21 Thread Akira Ajisaka (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15775?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira Ajisaka updated HADOOP-15775:
---
Attachment: HADOOP-15775.02.patch

> [JDK9] Add "--add-modules java.activation" to java option in unit tests
> ---
>
> Key: HADOOP-15775
> URL: https://issues.apache.org/jira/browse/HADOOP-15775
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: test
>Reporter: Akira Ajisaka
>Assignee: Akira Ajisaka
>Priority: Critical
> Attachments: HADOOP-15775.01.patch, HADOOP-15775.02.patch
>
>
> Many unit tests fail due to missing java.activation module. We need to 
> configure maven surefire plugin to add "--add-modules java.activation" option 
> to test JVM.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15775) [JDK9] Add "--add-modules java.activation" to java option in unit tests

2018-09-21 Thread Akira Ajisaka (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15775?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623248#comment-16623248
 ] 

Akira Ajisaka commented on HADOOP-15775:


bq. Is the third party dependency is available in Java 9 and 10?
It is available in Java 9 and 10. Great.

> [JDK9] Add "--add-modules java.activation" to java option in unit tests
> ---
>
> Key: HADOOP-15775
> URL: https://issues.apache.org/jira/browse/HADOOP-15775
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: test
>Reporter: Akira Ajisaka
>Assignee: Akira Ajisaka
>Priority: Critical
> Attachments: HADOOP-15775.01.patch
>
>
> Many unit tests fail due to missing java.activation module. We need to 
> configure maven surefire plugin to add "--add-modules java.activation" option 
> to test JVM.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15711) Fix branch-2 builds

2018-09-21 Thread Arun Suresh (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15711?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623202#comment-16623202
 ] 

Arun Suresh commented on HADOOP-15711:
--

Yeah... figured out with [~jhung] that isolating and ignoring specific tests 
wont work since as you mentioned, when executed in the parallel, a thread 
running a random batch of tests might go down.

Given that single YARN patch testing for instance  should run fine (Am assuming 
branch-2 Dockerfile is good currently since YARN-8658 ran fine), Can we disable 
parallel tests for full builds then ? Will that help track down specific tests 
that go OOM for instance ?

> Fix branch-2 builds
> ---
>
> Key: HADOOP-15711
> URL: https://issues.apache.org/jira/browse/HADOOP-15711
> Project: Hadoop Common
>  Issue Type: Task
>Reporter: Jonathan Hung
>Priority: Critical
> Attachments: HADOOP-15711.001.branch-2.patch
>
>
> Branch-2 builds have been disabled for a while: 
> https://builds.apache.org/view/H-L/view/Hadoop/job/hadoop-qbt-branch2-java7-linux-x86/
> A test run here causes hdfs tests to hang: 
> https://builds.apache.org/view/H-L/view/Hadoop/job/hadoop-qbt-branch2-java7-linux-x86-jhung/4/
> Running hadoop-hdfs tests locally reveal some errors such 
> as:{noformat}[ERROR] 
> testComplexAppend2(org.apache.hadoop.hdfs.TestFileAppend2)  Time elapsed: 
> 0.059 s  <<< ERROR!
> java.lang.OutOfMemoryError: unable to create new native thread
> at java.lang.Thread.start0(Native Method)
> at java.lang.Thread.start(Thread.java:714)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1164)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1128)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:174)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1172)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:403)
> at 
> org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:234)
> at 
> org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:1080)
> at 
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:883)
> at 
> org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514)
> at 
> org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473)
> at 
> org.apache.hadoop.hdfs.TestFileAppend2.testComplexAppend(TestFileAppend2.java:489)
> at 
> org.apache.hadoop.hdfs.TestFileAppend2.testComplexAppend2(TestFileAppend2.java:543)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){noformat}
> I was able to get more tests passing locally by increasing the max user 
> process count on my machine. But the error suggests that there's an issue in 
> the tests themselves. Not sure if the error seen locally is the same reason 
> as why jenkins builds are failing, I wasn't able to confirm based on the 
> jenkins builds' lack of output.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15775) [JDK9] Add "--add-modules java.activation" to java option in unit tests

2018-09-21 Thread Akira Ajisaka (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15775?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623165#comment-16623165
 ] 

Akira Ajisaka commented on HADOOP-15775:


Thank you for the information, [~apurtell]!
Is the third party dependency is available in Java 9 and 10? If so, adding the 
dependency in Java 9+ is better. Otherwise we need another profile to add the 
third party dependency only in Java 11+.

> [JDK9] Add "--add-modules java.activation" to java option in unit tests
> ---
>
> Key: HADOOP-15775
> URL: https://issues.apache.org/jira/browse/HADOOP-15775
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: test
>Reporter: Akira Ajisaka
>Assignee: Akira Ajisaka
>Priority: Critical
> Attachments: HADOOP-15775.01.patch
>
>
> Many unit tests fail due to missing java.activation module. We need to 
> configure maven surefire plugin to add "--add-modules java.activation" option 
> to test JVM.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14775) Change junit dependency in parent pom file to junit 5 while maintaining backward compatibility to junit4.

2018-09-21 Thread Akira Ajisaka (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-14775?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623162#comment-16623162
 ] 

Akira Ajisaka commented on HADOOP-14775:


Note: Apache Hadoop sets reuseForks=false and when reuseForks=false JUnit 5.3.0 
and Maven Surefire Plugin 2.22.0 fail to find tests. 
https://github.com/junit-team/junit5/issues/1579

Probably we need to wait for Maven Surefire Plugin 2.22.1 release.

> Change junit dependency in parent pom file to junit 5 while maintaining 
> backward compatibility to junit4. 
> --
>
> Key: HADOOP-14775
> URL: https://issues.apache.org/jira/browse/HADOOP-14775
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: build
>Affects Versions: 3.0.0-alpha4
>Reporter: Ajay Kumar
>Assignee: Akira Ajisaka
>Priority: Major
>  Labels: junit5
> Attachments: HADOOP-14775.01.patch, HADOOP-14775.02.patch, 
> HADOOP-14775.03.patch, HADOOP-14775.04.patch, HADOOP-14775.05.patch
>
>
> Change junit dependency in parent pom file to junit 5 while maintaining 
> backward compatibility to junit4. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14775) Change junit dependency in parent pom file to junit 5 while maintaining backward compatibility to junit4.

2018-09-21 Thread Akira Ajisaka (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-14775?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623105#comment-16623105
 ] 

Akira Ajisaka commented on HADOOP-14775:


Thanks [~ajayydv]!

Yes. If we want to run JUnit4 and JUnit5 tests in the same module, add vintage 
and jupitor engine and leave junit dependency as-is in pom.xml.
Sample: 
https://github.com/junit-team/junit5-samples/tree/master/junit5-migration-maven

> Change junit dependency in parent pom file to junit 5 while maintaining 
> backward compatibility to junit4. 
> --
>
> Key: HADOOP-14775
> URL: https://issues.apache.org/jira/browse/HADOOP-14775
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: build
>Affects Versions: 3.0.0-alpha4
>Reporter: Ajay Kumar
>Assignee: Akira Ajisaka
>Priority: Major
>  Labels: junit5
> Attachments: HADOOP-14775.01.patch, HADOOP-14775.02.patch, 
> HADOOP-14775.03.patch, HADOOP-14775.04.patch, HADOOP-14775.05.patch
>
>
> Change junit dependency in parent pom file to junit 5 while maintaining 
> backward compatibility to junit4. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org