[jira] [Reopened] (HDDS-1659) Define the process to add proposal/design docs to the Ozone subproject

2019-06-11 Thread Elek, Marton (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDDS-1659?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Elek, Marton reopened HDDS-1659:


> Define the process to add proposal/design docs to the Ozone subproject
> --
>
> Key: HDDS-1659
> URL: https://issues.apache.org/jira/browse/HDDS-1659
> Project: Hadoop Distributed Data Store
>  Issue Type: Task
>Reporter: Elek, Marton
>Assignee: Elek, Marton
>Priority: Major
>  Labels: pull-request-available
> Fix For: 0.4.1
>
>  Time Spent: 1h 50m
>  Remaining Estimate: 0h
>
> We think that it would be more effective to collect all the design docs in 
> one place and make it easier to review them by the community.
> We propose to follow an approach where the proposals are committed to the 
> hadoop-hdds/docs project and the review can be the same as a review of a PR



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1673) mapreduce smoketests are failig because an acl error

2019-06-11 Thread Elek, Marton (JIRA)
Elek, Marton created HDDS-1673:
--

 Summary: mapreduce smoketests are failig because an acl error
 Key: HDDS-1673
 URL: https://issues.apache.org/jira/browse/HDDS-1673
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Elek, Marton


After executing this command:

{code}
yarn jar ./share/hadoop/mapreduce/hadoop-mapreduce-examples-3.2.0.jar pi 3 3 2
{code}

The result is:

{code}
Number of Maps  = 3
Samples per Map = 3
2019-06-12 03:16:20 ERROR OzoneClientFactory:294 - Couldn't create protocol 
class org.apache.hadoop.ozone.client.rpc.RpcClient exception: 
java.lang.reflect.InvocationTargetException
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
at 
org.apache.hadoop.ozone.client.OzoneClientFactory.getClientProtocol(OzoneClientFactory.java:291)
at 
org.apache.hadoop.ozone.client.OzoneClientFactory.getRpcClient(OzoneClientFactory.java:169)
at 
org.apache.hadoop.fs.ozone.BasicOzoneClientAdapterImpl.(BasicOzoneClientAdapterImpl.java:134)
at 
org.apache.hadoop.fs.ozone.OzoneClientAdapterImpl.(OzoneClientAdapterImpl.java:50)
at 
org.apache.hadoop.fs.ozone.OzoneFileSystem.createAdapter(OzoneFileSystem.java:103)
at 
org.apache.hadoop.fs.ozone.BasicOzoneFileSystem.initialize(BasicOzoneFileSystem.java:143)
at 
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:3303)
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:124)
at 
org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:3352)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:3320)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:479)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:227)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:463)
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:361)
at 
org.apache.hadoop.mapreduce.lib.input.FileInputFormat.setInputPaths(FileInputFormat.java:522)
at 
org.apache.hadoop.examples.QuasiMonteCarlo.estimatePi(QuasiMonteCarlo.java:275)
at 
org.apache.hadoop.examples.QuasiMonteCarlo.run(QuasiMonteCarlo.java:360)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
at 
org.apache.hadoop.examples.QuasiMonteCarlo.main(QuasiMonteCarlo.java:368)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDriver.java:71)
at org.apache.hadoop.util.ProgramDriver.run(ProgramDriver.java:144)
at org.apache.hadoop.examples.ExampleDriver.main(ExampleDriver.java:74)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.hadoop.util.RunJar.run(RunJar.java:323)
at org.apache.hadoop.util.RunJar.main(RunJar.java:236)
Caused by: org.apache.hadoop.hdds.conf.ConfigurationException: Can't inject 
configuration to class 
org.apache.hadoop.ozone.security.acl.OzoneAclConfig.setUserDefaultRights
at 
org.apache.hadoop.hdds.conf.OzoneConfiguration.getObject(OzoneConfiguration.java:160)
at 
org.apache.hadoop.ozone.client.rpc.RpcClient.(RpcClient.java:148)
... 36 more
Caused by: java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.apache.hadoop.hdds.conf.OzoneConfiguration.getObject(OzoneConfiguration.java:137)
... 37 more
Caused by: java.lang.NullPointerException: Name is null
at java.lang.Enum.valueOf(Enum.java:236)
at 
org.apache.hadoop.ozone.security.acl.IAccessAuthorizer$ACLType.valueOf(IAccessAuthorizer.java:48)
at 
org.apache.hadoop.ozone.security.acl.OzoneAclConfig.setUserDefaultRights(OzoneAclConfig.java:43)
... 42 more
java.io.IOException: Could

[jira] [Created] (HDDS-1672) OzoneManager Lock change the volumeLock weight to 0

2019-06-11 Thread Bharat Viswanadham (JIRA)
Bharat Viswanadham created HDDS-1672:


 Summary: OzoneManager Lock change the volumeLock weight to 0
 Key: HDDS-1672
 URL: https://issues.apache.org/jira/browse/HDDS-1672
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Bharat Viswanadham
Assignee: Bharat Viswanadham


Currently after acquiring volume lock, we cannot acquire user lock. 

This is causing an issue in Volume request implementation, 
acquire/release/reacquire volume lock.

 

Case of Delete Volume Request: 
 # Acquire volume lock.
 # Get Volume Info from DB
 # Release Volume lock. (We are releasing the lock, because while acquiring 
volume lock, we cannot acquire user lock0
 # Get owner from volume Info read from DB
 # Acquire owner lock
 # Acquire volume lock
 # Do delete logic
 # release volume lock
 # release user lock

 

We can avoid this acquire/release/reacquire lock issue by making volume lock as 
low weight. 

 

In this way, the above deleteVolume request will change as below
 # Acquire volume lock
 # Get Volume Info from DB
 # Get owner from volume Info read from DB
 # Acquire owner lock
 # Do delete logic
 # release owner lock
 # release volume lock. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1671) Multiple unit test fails because of assertion while validating Acls

2019-06-11 Thread Mukul Kumar Singh (JIRA)
Mukul Kumar Singh created HDDS-1671:
---

 Summary: Multiple unit test fails because of assertion while 
validating Acls
 Key: HDDS-1671
 URL: https://issues.apache.org/jira/browse/HDDS-1671
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Security
Reporter: Mukul Kumar Singh


There are multiple unit test failures because of assertion in validateAcls
https://builds.apache.org/job/hadoop-multibranch/job/PR-846/7/testReport/




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1670) Add limit support to /api/containers and /api/containers/{id} endpoints

2019-06-11 Thread Vivek Ratnavel Subramanian (JIRA)
Vivek Ratnavel Subramanian created HDDS-1670:


 Summary: Add limit support to /api/containers and 
/api/containers/{id} endpoints
 Key: HDDS-1670
 URL: https://issues.apache.org/jira/browse/HDDS-1670
 Project: Hadoop Distributed Data Store
  Issue Type: Task
  Components: Ozone Recon
Affects Versions: 0.4.0
Reporter: Vivek Ratnavel Subramanian
Assignee: Vivek Ratnavel Subramanian


Add support for limit query param to limit the results of /api/containers and 
/api/containers/\{id} endpoints



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-14561) dfs.datanode.data.dir does not honor file:/// scheme

2019-06-11 Thread Artem Ervits (JIRA)
Artem Ervits created HDFS-14561:
---

 Summary: dfs.datanode.data.dir does not honor file:/// scheme
 Key: HDFS-14561
 URL: https://issues.apache.org/jira/browse/HDFS-14561
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: documentation, hdfs
Affects Versions: 2.8.5, 2.9.2
Reporter: Artem Ervits


_dfs.datanode.data.dir_ with {{file:///}} scheme is not being honored in Hadoop 
2.8.5 and 2.9.2. I didn't test 3.x branches but good to validate also. In my 
environment, only file:// seemed to work. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Re: [DISCUSS] A unified and open Hadoop community sync up schedule?

2019-06-11 Thread Anu Engineer
For Ozone, we have started using the Wiki itself as the agenda and after
the meeting is over, we convert it into the meeting notes.
Here is an example, the project owner can edit and maintain it, it is like
10 mins work - and allows anyone to add stuff into the agenda too.

https://cwiki.apache.org/confluence/display/HADOOP/2019-06-10+Meeting+notes

--Anu

On Tue, Jun 11, 2019 at 10:20 AM Yufei Gu  wrote:

> +1 for this idea. Thanks Wangda for bringing this up.
>
> Some comments to share:
>
>- Agenda needed to be posted ahead of meeting and welcome any interested
>party to contribute to topics.
>- We should encourage more people to attend. That's whole point of the
>meeting.
>- Hopefully, this can mitigate the situation that some patches are
>waiting for review for ever, which turns away new contributors.
>- 30m per session sounds a little bit short, we can try it out and see
>if extension is needed.
>
> Best,
>
> Yufei
>
> `This is not a contribution`
>
>
> On Fri, Jun 7, 2019 at 4:39 PM Wangda Tan  wrote:
>
> > Hi Hadoop-devs,
> >
> > Previous we have regular YARN community sync up (1 hr, biweekly, but not
> > open to public). Recently because of changes in our schedules, Less folks
> > showed up in the sync up for the last several months.
> >
> > I saw the K8s community did a pretty good job to run their sig meetings,
> > there's regular meetings for different topics, notes, agenda, etc. Such
> as
> >
> >
> https://docs.google.com/document/d/13mwye7nvrmV11q9_Eg77z-1w3X7Q1GTbslpml4J7F3A/edit
> >
> >
> > For Hadoop community, there are less such regular meetings open to the
> > public except for Ozone project and offline meetups or Bird-of-Features
> in
> > Hadoop/DataWorks Summit. Recently we have a few folks joined DataWorks
> > Summit at Washington DC and Barcelona, and lots (50+) of folks join the
> > Ozone/Hadoop/YARN BoF, ask (good) questions and roadmaps. I think it is
> > important to open such conversations to the public and let more
> > folk/companies join.
> >
> > Discussed a small group of community members and wrote a short proposal
> > about the form, time and topic of the community sync up, thanks for
> > everybody who have contributed to the proposal! Please feel free to add
> > your thoughts to the Proposal Google doc
> > <
> >
> https://docs.google.com/document/d/1GfNpYKhNUERAEH7m3yx6OfleoF3MqoQk3nJ7xqHD9nY/edit#
> > >
> > .
> >
> > Especially for the following parts:
> > - If you have interests to run any of the community sync-ups, please put
> > your name to the table inside the proposal. We need more volunteers to
> help
> > run the sync-ups in different timezones.
> > - Please add suggestions to the time, frequency and themes and feel free
> to
> > share your thoughts if we should do sync ups for other topics which are
> not
> > covered by the proposal.
> >
> > Link to the Proposal Google doc
> > <
> >
> https://docs.google.com/document/d/1GfNpYKhNUERAEH7m3yx6OfleoF3MqoQk3nJ7xqHD9nY/edit#
> > >
> >
> > Thanks,
> > Wangda Tan
> >
>


Re: [DISCUSS] A unified and open Hadoop community sync up schedule?

2019-06-11 Thread Yufei Gu
+1 for this idea. Thanks Wangda for bringing this up.

Some comments to share:

   - Agenda needed to be posted ahead of meeting and welcome any interested
   party to contribute to topics.
   - We should encourage more people to attend. That's whole point of the
   meeting.
   - Hopefully, this can mitigate the situation that some patches are
   waiting for review for ever, which turns away new contributors.
   - 30m per session sounds a little bit short, we can try it out and see
   if extension is needed.

Best,

Yufei

`This is not a contribution`


On Fri, Jun 7, 2019 at 4:39 PM Wangda Tan  wrote:

> Hi Hadoop-devs,
>
> Previous we have regular YARN community sync up (1 hr, biweekly, but not
> open to public). Recently because of changes in our schedules, Less folks
> showed up in the sync up for the last several months.
>
> I saw the K8s community did a pretty good job to run their sig meetings,
> there's regular meetings for different topics, notes, agenda, etc. Such as
>
> https://docs.google.com/document/d/13mwye7nvrmV11q9_Eg77z-1w3X7Q1GTbslpml4J7F3A/edit
>
>
> For Hadoop community, there are less such regular meetings open to the
> public except for Ozone project and offline meetups or Bird-of-Features in
> Hadoop/DataWorks Summit. Recently we have a few folks joined DataWorks
> Summit at Washington DC and Barcelona, and lots (50+) of folks join the
> Ozone/Hadoop/YARN BoF, ask (good) questions and roadmaps. I think it is
> important to open such conversations to the public and let more
> folk/companies join.
>
> Discussed a small group of community members and wrote a short proposal
> about the form, time and topic of the community sync up, thanks for
> everybody who have contributed to the proposal! Please feel free to add
> your thoughts to the Proposal Google doc
> <
> https://docs.google.com/document/d/1GfNpYKhNUERAEH7m3yx6OfleoF3MqoQk3nJ7xqHD9nY/edit#
> >
> .
>
> Especially for the following parts:
> - If you have interests to run any of the community sync-ups, please put
> your name to the table inside the proposal. We need more volunteers to help
> run the sync-ups in different timezones.
> - Please add suggestions to the time, frequency and themes and feel free to
> share your thoughts if we should do sync ups for other topics which are not
> covered by the proposal.
>
> Link to the Proposal Google doc
> <
> https://docs.google.com/document/d/1GfNpYKhNUERAEH7m3yx6OfleoF3MqoQk3nJ7xqHD9nY/edit#
> >
>
> Thanks,
> Wangda Tan
>


Apache Hadoop qbt Report: trunk+JDK8 on Linux/x86

2019-06-11 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1164/

[Jun 10, 2019 5:23:18 PM] (weichiu) HDFS-14494. Move Server logging of StatedId 
inside
[Jun 10, 2019 7:18:15 PM] (weichiu) YARN-9471. Cleanup in 
TestLogAggregationIndexFileController. Contributed
[Jun 10, 2019 8:45:37 PM] (weichiu) HDFS-10659. Namenode crashes after 
Journalnode re-installation in an HA
[Jun 10, 2019 9:13:53 PM] (weichiu) HDFS-10210. Remove the defunct startKdc 
profile from hdfs. Contributed
[Jun 10, 2019 9:33:24 PM] (sumasai) YARN-9569. Auto-created leaf queues do not 
honor cluster-wide min/max
[Jun 11, 2019 12:21:52 AM] (weichiu) HDFS-14553. Make queue size of 
BlockReportProcessingThread configurable.
[Jun 11, 2019 12:55:16 AM] (aengineer) HDFS-14234. Limit WebHDFS to specifc 
user, host, directory triples.




-1 overall


The following subsystems voted -1:
asflicense findbugs hadolint pathlen unit


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

FindBugs :

   
module:hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-documentstore
 
   Unread field:TimelineEventSubDoc.java:[line 56] 
   Unread field:TimelineMetricSubDoc.java:[line 44] 

FindBugs :

   
module:hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-mawo/hadoop-yarn-applications-mawo-core
 
   Class org.apache.hadoop.applications.mawo.server.common.TaskStatus 
implements Cloneable but does not define or use clone method At 
TaskStatus.java:does not define or use clone method At TaskStatus.java:[lines 
39-346] 
   Equals method for 
org.apache.hadoop.applications.mawo.server.worker.WorkerId assumes the argument 
is of type WorkerId At WorkerId.java:the argument is of type WorkerId At 
WorkerId.java:[line 114] 
   
org.apache.hadoop.applications.mawo.server.worker.WorkerId.equals(Object) does 
not check for null argument At WorkerId.java:null argument At 
WorkerId.java:[lines 114-115] 

Failed junit tests :

   hadoop.util.TestDiskChecker 
   hadoop.hdfs.web.TestWebHdfsTimeouts 
   hadoop.hdfs.server.balancer.TestBalancerWithMultipleNameNodes 
   hadoop.hdfs.server.datanode.TestDataNodeRollingUpgrade 
   hadoop.yarn.applications.distributedshell.TestDistributedShell 
   hadoop.mapreduce.v2.app.TestRuntimeEstimators 
   hadoop.mapreduce.jobhistory.TestJobHistoryEventHandler 
   hadoop.mapred.TestMRTimelineEventHandling 
   hadoop.yarn.sls.appmaster.TestAMSimulator 
   hadoop.ozone.container.common.impl.TestHddsDispatcher 
   hadoop.ozone.container.ozoneimpl.TestOzoneContainer 
   
hadoop.ozone.container.common.statemachine.commandhandler.TestBlockDeletion 
   hadoop.ozone.client.rpc.TestOzoneRpcClient 
   hadoop.ozone.client.rpc.TestOzoneRpcClientWithRatis 
   hadoop.ozone.client.rpc.TestSecureOzoneRpcClient 
   hadoop.ozone.client.rpc.TestOzoneAtRestEncryption 
   hadoop.fs.ozone.contract.ITestOzoneContractRootDir 
   hadoop.fs.ozone.contract.ITestOzoneContractDistCp 
   hadoop.fs.ozone.contract.ITestOzoneContractRename 
   hadoop.fs.ozone.contract.ITestOzoneContractGetFileStatus 
   hadoop.fs.ozone.contract.ITestOzoneContractOpen 
   hadoop.fs.ozone.contract.ITestOzoneContractCreate 
   hadoop.fs.ozone.contract.ITestOzoneContractSeek 
   hadoop.fs.ozone.contract.ITestOzoneContractMkdir 
   hadoop.fs.ozone.contract.ITestOzoneContractDelete 
   hadoop.ozone.freon.TestFreonWithPipelineDestroy 
   hadoop.ozone.freon.TestDataValidateWithUnsafeByteOperations 
   hadoop.ozone.freon.TestFreonWithDatanodeRestart 
   hadoop.ozone.freon.TestFreonWithDatanodeFastRestart 
   hadoop.ozone.fsck.TestContainerMapper 
   hadoop.ozone.freon.TestRandomKeyGenerator 
   hadoop.ozone.freon.TestDataValidateWithSafeByteOperations 
  

   cc:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1164/artifact/out/diff-compile-cc-root.txt
  [4.0K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1164/artifact/out/diff-compile-javac-root.txt
  [332K]

   checkstyle:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1164/artifact/out/diff-checkstyle-root.txt
  [17M]

   hadolint:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1164/artifact/out/diff-patch-hadolint.txt
  [8.0K]

   pathlen:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1164/artifact/out/pathlen.txt
  [12K]

   pylint:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1164/artifact/out/diff-patch-pylint.txt
  [120K]

   shellcheck:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux

[jira] [Created] (HDDS-1669) SCM startup is failing if network-topology-default.xml is part of a jar

2019-06-11 Thread Elek, Marton (JIRA)
Elek, Marton created HDDS-1669:
--

 Summary: SCM startup is failing if network-topology-default.xml is 
part of a jar
 Key: HDDS-1669
 URL: https://issues.apache.org/jira/browse/HDDS-1669
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Elek, Marton
Assignee: Elek, Marton


network-topology-default.xml can be loaded from file or classpath. But the 
NodeSchemaLoader assumes that the files on the classpath can be opened as a 
file. It's true if the file is in etc/hadoop (which is part of the classpath) 
but not true if the file is packaged to a jajr file:

{code}
scm_1 | 2019-06-11 13:18:03 INFO  NodeSchemaLoader:118 - Loading file 
from 
jar:file:/opt/hadoop/share/ozone/lib/hadoop-hdds-common-0.5.0-SNAPSHOT.jar!/network-topology-default.xml
scm_1 | 2019-06-11 13:18:03 ERROR NodeSchemaManager:74 - Failed to load 
schema file:network-topology-default.xml, error:
scm_1 | java.lang.IllegalArgumentException: URI is not hierarchical
scm_1 | at java.io.File.(File.java:418)
scm_1 | at 
org.apache.hadoop.hdds.scm.net.NodeSchemaLoader.loadSchemaFromFile(NodeSchemaLoader.java:119)
scm_1 | at 
org.apache.hadoop.hdds.scm.net.NodeSchemaManager.init(NodeSchemaManager.java:67)
scm_1 | at 
org.apache.hadoop.hdds.scm.net.NetworkTopologyImpl.(NetworkTopologyImpl.java:63)
scm_1 | at 
org.apache.hadoop.hdds.scm.server.StorageContainerManager.initializeSystemManagers(StorageContainerManager.java:382)
scm_1 | at 
org.apache.hadoop.hdds.scm.server.StorageContainerManager.(StorageContainerManager.java:275)
scm_1 | at 
org.apache.hadoop.hdds.scm.server.StorageContainerManager.(StorageContainerManager.java:208)
scm_1 | at 
org.apache.hadoop.hdds.scm.server.StorageContainerManager.createSCM(StorageContainerManager.java:586)
scm_1 | at 
org.apache.hadoop.hdds.scm.server.StorageContainerManagerStarter$SCMStarterHelper.start(StorageContainerManagerStarter.java:139)
scm_1 | at 
org.apache.hadoop.hdds.scm.server.StorageContainerManagerStarter.startScm(StorageContainerManagerStarter.java:115)
scm_1 | at 
org.apache.hadoop.hdds.scm.server.StorageContainerManagerStarter.call(StorageContainerManagerStarter.java:67)
scm_1 | at 
org.apache.hadoop.hdds.scm.server.StorageContainerManagerStarter.call(StorageContainerManagerStarter.java:42)
scm_1 | at picocli.CommandLine.execute(CommandLine.java:1173)
scm_1 | at picocli.CommandLine.access$800(CommandLine.java:141)
scm_1 | at 
picocli.CommandLine$RunLast.handle(CommandLine.java:1367)
scm_1 | at 
picocli.CommandLine$RunLast.handle(CommandLine.java:1335)
scm_1 | at 
picocli.CommandLine$AbstractParseResultHandler.handleParseResult(CommandLine.java:1243)
scm_1 | at 
picocli.CommandLine.parseWithHandlers(CommandLine.java:1526)
scm_1 | at 
picocli.CommandLine.parseWithHandler(CommandLine.java:1465)
scm_1 | at 
org.apache.hadoop.hdds.cli.GenericCli.execute(GenericCli.java:65)
scm_1 | at 
org.apache.hadoop.hdds.cli.GenericCli.run(GenericCli.java:56)
scm_1 | at 
org.apache.hadoop.hdds.scm.server.StorageContainerManagerStarter.main(StorageContainerManagerStarter.java:56)
scm_1 | Failed to load schema file:network-topology-default.xml, error:
{code}

The quick fix is to keep the current behaviour but read the file from 
classloader.getResourceAsStream() instead of classloader.getResource().toURI()



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Apache Hadoop qbt Report: branch2+JDK7 on Linux/x86

2019-06-11 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/

No changes




-1 overall


The following subsystems voted -1:
asflicense findbugs hadolint pathlen unit xml


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

XML :

   Parsing Error(s): 
   
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/empty-configuration.xml
 
   hadoop-tools/hadoop-azure/src/config/checkstyle-suppressions.xml 
   hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/public/crossdomain.xml 
   
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/public/crossdomain.xml
 

FindBugs :

   module:hadoop-common-project/hadoop-common 
   Class org.apache.hadoop.fs.GlobalStorageStatistics defines non-transient 
non-serializable instance field map In GlobalStorageStatistics.java:instance 
field map In GlobalStorageStatistics.java 

FindBugs :

   
module:hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-client
 
   Boxed value is unboxed and then immediately reboxed in 
org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result,
 byte[], byte[], KeyConverter, ValueConverter, boolean) At 
ColumnRWHelper.java:then immediately reboxed in 
org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result,
 byte[], byte[], KeyConverter, ValueConverter, boolean) At 
ColumnRWHelper.java:[line 335] 

Failed junit tests :

   hadoop.hdfs.qjournal.server.TestJournalNodeRespectsBindHostKeys 
   hadoop.hdfs.web.TestWebHdfsTimeouts 
   hadoop.registry.secure.TestSecureLogins 
   hadoop.yarn.server.timelineservice.security.TestTimelineAuthFilterForV2 
  

   cc:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/diff-compile-cc-root-jdk1.7.0_95.txt
  [4.0K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/diff-compile-javac-root-jdk1.7.0_95.txt
  [328K]

   cc:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/diff-compile-cc-root-jdk1.8.0_212.txt
  [4.0K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/diff-compile-javac-root-jdk1.8.0_212.txt
  [308K]

   checkstyle:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/diff-checkstyle-root.txt
  [16M]

   hadolint:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/diff-patch-hadolint.txt
  [4.0K]

   pathlen:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/pathlen.txt
  [12K]

   pylint:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/diff-patch-pylint.txt
  [24K]

   shellcheck:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/diff-patch-shellcheck.txt
  [72K]

   shelldocs:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/diff-patch-shelldocs.txt
  [8.0K]

   whitespace:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/whitespace-eol.txt
  [12M]
   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/whitespace-tabs.txt
  [1.2M]

   xml:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/xml.txt
  [12K]

   findbugs:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/branch-findbugs-hadoop-common-project_hadoop-common-warnings.html
  [8.0K]
   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/branch-findbugs-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-timelineservice-hbase_hadoop-yarn-server-timelineservice-hbase-client-warnings.html
  [8.0K]

   javadoc:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/diff-javadoc-javadoc-root-jdk1.7.0_95.txt
  [16K]
   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/diff-javadoc-javadoc-root-jdk1.8.0_212.txt
  [1.1M]

   unit:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt
  [284K]
   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-registry.txt
  [12K]
   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/349/art

[jira] [Created] (HDFS-14560) Allow block replication parameters to be refreshable

2019-06-11 Thread Stephen O'Donnell (JIRA)
Stephen O'Donnell created HDFS-14560:


 Summary: Allow block replication parameters to be refreshable
 Key: HDFS-14560
 URL: https://issues.apache.org/jira/browse/HDFS-14560
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Affects Versions: 3.3.0
Reporter: Stephen O'Donnell
Assignee: Stephen O'Donnell


There are 3 key parameters that control the speed of block replication across 
the cluster:

{code}
dfs.namenode.replication.max-streams
dfs.namenode.replication.max-streams-hard-limit
dfs.namenode.replication.work.multiplier.per.iteration
{code}

These are used when decomissioning nodes and when under replicated blocks are 
being recovered across the cluster. There are times when it may be desirable to 
increase the speed of replication and then reduce it again (eg during off peak 
hours) without restarting the namenode.

Therefore it would be good to be able to reconfigure / refresh these parameters 
dynamically without a namenode restart.

This Jira is to allow these parameters to be refreshed at runtime without a 
restart.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1668) Add liveness probe to the example k8s resources files

2019-06-11 Thread Elek, Marton (JIRA)
Elek, Marton created HDDS-1668:
--

 Summary: Add liveness probe to the example k8s resources files
 Key: HDDS-1668
 URL: https://issues.apache.org/jira/browse/HDDS-1668
 Project: Hadoop Distributed Data Store
  Issue Type: Sub-task
Reporter: Elek, Marton
Assignee: Elek, Marton


In kubernetes resources we can define livebess probes which can help to detect 
any failure. If the define port is not available the pod will be rescheduled.

We need to add the liveness probes to our k8s resource files.

Note: We shouldn't add readiness probes. Readiness probe is about the service 
availability. The service/dns can be available only after the service is 
restarted. This is not good for us as:

 * We need DNS resolution during the startup (See OzoneManager.loadOMHAConfigs)
 * We already implemented retry in case of missing DNS entries



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-14559) Optimizing safemode leave mechanism

2019-06-11 Thread He Xiaoqiao (JIRA)
He Xiaoqiao created HDFS-14559:
--

 Summary: Optimizing safemode leave mechanism
 Key: HDFS-14559
 URL: https://issues.apache.org/jira/browse/HDFS-14559
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Reporter: He Xiaoqiao
Assignee: He Xiaoqiao


As HDFS-14186 mentioned, The last stage of namenode startup, it will leave 
safemode based on the condition that if blocks num reach to threshold. However 
the current condition is complete based on total blocks rather than total 
replications. So for a large cluster, after total blocks has reported from 
datanode, there are still large block replication pending report and load of 
namenode is continue high for long times. In some extreme case, between leave 
safemode time and process block report completely, namenode will not provide 
normal service and some datanodes could dead then register/blockreport again 
and again.
In one word, we need to upgrade safemode leave mechanism to support large 
cluster restart smooth.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-14424) NN failover failed beacuse of lossing paxos directory

2019-06-11 Thread star (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDFS-14424?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

star resolved HDFS-14424.
-
Resolution: Duplicate

> NN failover failed beacuse of lossing paxos directory
> -
>
> Key: HDFS-14424
> URL: https://issues.apache.org/jira/browse/HDFS-14424
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: star
>Assignee: star
>Priority: Major
>
> Recently, our hadoop namenode shutdown when switching active namenode, just 
> because of missing paxos directory. It is created in the default /tmp path 
> and deleted by os for no operation in 7 days. We can avoid this by moving 
> journal directory to a none tmp dir, but it‘s better to make sure namenode 
> works well by a default config.
> The issue throws exception similar to HDFS-10659, also caused by missing 
> paxos directory.
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Re: [VOTE] Merge HDFS-13891(RBF) to trunk

2019-06-11 Thread Takanobu Asanuma
+1(non-binding).

Regards,
- Takanobu


From: Xiaoqiao He 
Sent: Monday, June 10, 2019 15:47
To: Ranith Sardar
Cc: Brahma Reddy Battula; Hadoop Common; Hdfs-dev
Subject: Re: [VOTE] Merge HDFS-13891(RBF) to trunk

+1 (non-binding)

- Try to merge branch HDFS-13891(RBF) to trunk at local and no conflict or
failure.
- Built from merged sources.
- Ran 85 RBF test class at local and result shows: (Tests run: 639,
Failures: 2, Errors: 2, Skipped: 2) - failed tests include
#TestRouterWithSecureStartup and #TestRouterHttpDelegationToken. I don't
think it is blocked issue.

Thanks Brahma for organizing the merge.


On Mon, Jun 10, 2019 at 12:27 PM Ranith Sardar 
wrote:

> +1 (Non-binding)
>
> -Original Message-
> From: Brahma Reddy Battula [mailto:bra...@apache.org]
> Sent: 09 June 2019 19:31
> To: Hadoop Common ; Hdfs-dev <
> hdfs-dev@hadoop.apache.org>
> Subject: [VOTE] Merge HDFS-13891(RBF) to trunk
>
> Updated mail...
>
> -- Forwarded message -
> From: Brahma Reddy Battula 
> Date: Sun, Jun 9, 2019 at 7:26 PM
> Subject: Re: [VOTE] Merge HDFS-13891(RBF) to trunk
> To: Xiaoqiao He 
> Cc: Akira Ajisaka , Chittaranjan Hota <
> chitts.h...@gmail.com>, Giovanni Matteo Fumarola <
> giovanni.fumar...@gmail.com>, Hadoop Common ,
> Hdfs-dev , Iñigo Goiri 
>
>
> Hi All,
>
> Given the positive response to the discussion thread [1], here is the
> formal vote thread to merge HDFS-13891 in to trunk.
>
> Summary of code changes:
> 1. Code changes for this branch are done in the hadoop-hdfs-rbf
> subproject, there is no impact to hadoop-hdfs and hadoop-common.
> 2. Added Security support for RBF
> 3. Added Missing Client Protocol API's
> 4. Bug Fixes/ Improvments
>
>
>  The vote will run for 7 days, ending Sat June 15th. I will start this
> vote with my +1.
>
> Regards,
> Brahma Reddy Battula
>
> 1).
>
> https://lists.apache.org/thread.html/cdc2e084874b30bf6af2dd827bcbdba4ab8d3d983a8b9796e61608e6@%3Ccommon-dev.hadoop.apache.org%3E
>
>
> On Wed, Jun 5, 2019 at 3:44 PM Xiaoqiao He  wrote:
>
> > Thanks Brahma for starting the thread.
> > +1 for merging.
> >
> > He Xiaoqiao
> >
> > On Tue, Jun 4, 2019 at 1:53 PM Chittaranjan Hota
> > 
> > wrote:
> >
> >> Thanks Brahma initiating this.
> >> +1(non-binding) for merge.
> >>
> >> @Uber we have almost all changes specially rbf security in production
> >> for a while now without issues.
> >>
> >> On Mon, Jun 3, 2019 at 12:56 PM Giovanni Matteo Fumarola <
> >> giovanni.fumar...@gmail.com> wrote:
> >>
> >> > +1 on merging.
> >> >
> >> > Thanks Brahma for starting the thread.
> >> >
> >> > On Mon, Jun 3, 2019 at 10:00 AM Iñigo Goiri 
> wrote:
> >> >
> >> > > Thank you Brahma for pushing this.
> >> > >
> >> > > As you mentioned, we have already taken most of the changes into
> >> > > production.
> >> > > I want to highlight that the main contribution is the addition of
> >> > security.
> >> > > We have been able to test this at a smaller scale (~500 servers
> >> > > and 4
> >> > > subclusters) and the performance is great with our current
> >> > > ZooKeeper deployment.
> >> > > I would also like to highlight that all the changes are
> >> > > constrained to hadoop-hdfs-rbf and there is no differences in
> commons or HDFS.
> >> > >
> >> > > +1 on merging
> >> > >
> >> > > Inigo
> >> > >
> >> > > On Sun, Jun 2, 2019 at 10:19 PM Akira Ajisaka
> >> > > 
> >> > wrote:
> >> > >
> >> > > > Thanks Brahma for starting the discussion.
> >> > > > I'm +1 for merging this.
> >> > > >
> >> > > > FYI: In Yahoo! JAPAN, deployed all these changes in 20 nodes
> >> > > > cluster with 2 routers (not in production) and running several
> tests.
> >> > > >
> >> > > > Regards,
> >> > > > Akira
> >> > > >
> >> > > > On Sun, Jun 2, 2019 at 12:40 PM Brahma Reddy Battula <
> >> > bra...@apache.org>
> >> > > > wrote:
> >> > > > >
> >> > > > > Dear Hadoop Developers
> >> > > > >
> >> > > > > I would like to propose RBF Branch (HDFS-13891) merge into
> trunk.
> >> We
> >> > > have
> >> > > > > been working on this feature from last several months.
> >> > > > > This feature work received the contributions from different
> >> > companies.
> >> > > > All
> >> > > > > of the feature development happened smoothly and
> >> > > > > collaboratively
> >> in
> >> > > > JIRAs.
> >> > > > >
> >> > > > > Kindly do take a look at the branch and raise issues/concerns
> >> > > > > that
> >> > need
> >> > > > to
> >> > > > > be addressed before the merge.
> >> > > > >
> >> > > > > *Highlights of HDFS-13891 Branch:*
> >> > > > > =
> >> > > > >
> >> > > > > Adding Security to RBF(1)
> >> > > > > Adding Missing Client API's(2) Improvements/Bug Fixing
> >> > > > >   Critical - HDFS-13637, HDFS-13834
> >> > > > >
> >> > > > > *Commits:*
> >> > > > > 
> >> > > > >
> >> > > > > No of JIRAs Resolved: 72
> >> > > > >
> >> > > > > All this commits are in RBF Module. No changes in hdfs/common.
> >> > > > >
> >> > > > > *Test