Could you check which xerces version you have on your classpath?
Apparently, it cannot read core-default.xml as Ted pointed out. This might
be the root cause for the failure.

Cheers,
Till

On Mon, Jun 18, 2018 at 1:31 PM Garvit Sharma <garvit...@gmail.com> wrote:

> Hi,
>
> Sorry for the confusion, but the yarn is running on Hadoop version 2.7
> only and hence I am using Flink 1.5 Hadoop 2.7 binary.
>
> Below are the details provided by Yarn version command :
>
> Hadoop 2.7.3.2.6.3.0-235
> Subversion g...@github.com:hortonworks/hadoop.git -r
> 45bfd33bba8acadfa0e6024c80981c023b28d454
> Compiled by jenkins on 2017-10-30T02:31Z
> Compiled with protoc 2.5.0
> From source with checksum cd1a4a466ef450f547c279989f3aa3
> This command was run using
> /usr/hdp/2.6.3.0-235/hadoop/hadoop-common-2.7.3.2.6.3.0-235.jar
>
> Please let me know if you have found the resolution to my issue :)
>
> Thanks,
>
>
> On Mon, Jun 18, 2018 at 4:50 PM Till Rohrmann <trohrm...@apache.org>
> wrote:
>
>> Which Hadoop version have you installed? It looks as if Flink has been
>> build with Hadoop 2.7 but I see /usr/hdp/2.6.3.0-235 in the class path. If
>> you want to run Flink on Hadoop 2.6, then try to use the Hadoop free Flink
>> binaries or the one built for Hadoop 2.6.
>>
>> Cheers,
>> Till
>>
>> On Mon, Jun 18, 2018 at 10:48 AM Garvit Sharma <garvit...@gmail.com>
>> wrote:
>>
>>> Ok, I have attached the log file.
>>>
>>> Please check and let me know.
>>>
>>> Thanks,
>>>
>>> On Mon, Jun 18, 2018 at 2:07 PM Amit Jain <aj201...@gmail.com> wrote:
>>>
>>>> Hi Gravit,
>>>>
>>>> I think Till is interested to know about classpath details present at
>>>> the start of JM and TM logs e.g. following logs provide classpath details
>>>> used by TM in our case.
>>>>
>>>> 2018-06-17 19:01:30,656 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>> --------------------------------------------------------------------------------
>>>> 2018-06-17 19:01:30,658 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -  Starting
>>>> YARN TaskExecutor runner (Version: 1.5.0, Rev:c61b108, Date:24.05.2018 @
>>>> 14:54:44 UTC)
>>>> 2018-06-17 19:01:30,659 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -  OS
>>>> current user: yarn
>>>> 2018-06-17 19:01:31,662 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -  Current
>>>> Hadoop/Kerberos user: hadoop
>>>> 2018-06-17 19:01:31,663 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -  JVM:
>>>> OpenJDK 64-Bit Server VM - Oracle Corporation - 1.8/25.171-b10
>>>> 2018-06-17 19:01:31,663 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -  Maximum
>>>> heap size: 6647 MiBytes
>>>> 2018-06-17 19:01:31,663 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>>  JAVA_HOME: /usr/lib/jvm/java-openjdk
>>>> 2018-06-17 19:01:31,664 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -  Hadoop
>>>> version: 2.8.3
>>>> 2018-06-17 19:01:31,664 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -  JVM
>>>> Options:
>>>> 2018-06-17 19:01:31,665 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>> -Xms6936m
>>>> 2018-06-17 19:01:31,665 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>> -Xmx6936m
>>>> 2018-06-17 19:01:31,665 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>> -XX:MaxDirectMemorySize=4072m
>>>> 2018-06-17 19:01:31,665 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>> -Dlog.file=/var/log/hadoop-yarn/containers/application_1528342246614_0002/container_1528342246614_0002_01_282649/taskmanager.log
>>>> 2018-06-17 19:01:31,665 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>> -Dlogback.configurationFile=file:./logback.xml
>>>> 2018-06-17 19:01:31,665 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>> -Dlog4j.configuration=file:./log4j.properties
>>>> 2018-06-17 19:01:31,665 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -  Program
>>>> Arguments:
>>>> 2018-06-17 19:01:31,665 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>> --configDir
>>>> 2018-06-17 19:01:31,665 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -     .
>>>> *2018-06-17 19:01:31,666 INFO
>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>>  Classpath:
>>>> lib/flink-dist_2.11-1.5.0.jar:lib/flink-python_2.11-1.5.0.jar:lib/flink-shaded-hadoop2-uber-1.5.0.jar:lib/flink-shaded-include-yarn-0.9.1.jar:lib/guava-18.0.jar:lib/log4j-1.2.17.jar:lib/slf4j-log4j12-1.7.7.jar:log4j.properties:logback.xml:flink.jar:flink-conf.yaml::/etc/hadoop/conf:/usr/lib/hadoop/hadoop-common-2.8.3-amzn-0.jar:/usr/lib/hadoop/hadoop-archive-logs.jar:/usr/lib/hadoop/hadoop-auth.jar:/usr/lib/hadoop/hadoop-archives-2.8.3-amzn-0.jar:/usr/lib/hadoop/hadoop-archive-logs-2.8.3-amzn-0.jar:/usr/lib/hadoop/hadoop-azure-datalake-2.8.3-amzn-0.jar.........*
>>>>
>>>> --
>>>> Thanks,
>>>> Amit
>>>>
>>>> On Mon, Jun 18, 2018 at 2:00 PM, Garvit Sharma <garvit...@gmail.com>
>>>> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> Please refer to my previous mail for complete logs.
>>>>>
>>>>> Thanks,
>>>>>
>>>>> On Mon, Jun 18, 2018 at 1:17 PM Till Rohrmann <trohrm...@apache.org>
>>>>> wrote:
>>>>>
>>>>>> Could you also please share the complete log file with us.
>>>>>>
>>>>>> Cheers,
>>>>>> Till
>>>>>>
>>>>>> On Sat, Jun 16, 2018 at 5:22 PM Ted Yu <yuzhih...@gmail.com> wrote:
>>>>>>
>>>>>>> The error for core-default.xml is interesting.
>>>>>>>
>>>>>>> Flink doesn't have this file. Probably it came with Yarn. Please
>>>>>>> check the hadoop version Flink was built with versus the hadoop version 
>>>>>>> in
>>>>>>> your cluster.
>>>>>>>
>>>>>>> Thanks
>>>>>>>
>>>>>>> -------- Original message --------
>>>>>>> From: Garvit Sharma <garvit...@gmail.com>
>>>>>>> Date: 6/16/18 7:23 AM (GMT-08:00)
>>>>>>> To: trohrm...@apache.org
>>>>>>> Cc: Chesnay Schepler <ches...@apache.org>, user@flink.apache.org
>>>>>>> Subject: Re: Exception while submitting jobs through Yarn
>>>>>>>
>>>>>>> I am not able to figure out, got stuck badly in this since last 1
>>>>>>> week. Any little help would be appreciated.
>>>>>>>
>>>>>>>
>>>>>>> 2018-06-16 19:25:10,523 DEBUG
>>>>>>> org.apache.flink.streaming.api.graph.StreamingJobGraphGenerator  -
>>>>>>> Parallelism set: 1 for 8
>>>>>>>
>>>>>>> 2018-06-16 19:25:10,578 DEBUG
>>>>>>> org.apache.flink.streaming.api.graph.StreamingJobGraphGenerator  -
>>>>>>> Parallelism set: 1 for 1
>>>>>>>
>>>>>>> 2018-06-16 19:25:10,588 DEBUG
>>>>>>> org.apache.flink.streaming.api.graph.StreamingJobGraphGenerator  -
>>>>>>> CONNECTED: KeyGroupStreamPartitioner - 1 -> 8
>>>>>>>
>>>>>>> 2018-06-16 19:25:10,591 DEBUG
>>>>>>> org.apache.flink.streaming.api.graph.StreamingJobGraphGenerator  -
>>>>>>> Parallelism set: 1 for 5
>>>>>>>
>>>>>>> 2018-06-16 19:25:10,597 DEBUG
>>>>>>> org.apache.flink.streaming.api.graph.StreamingJobGraphGenerator  -
>>>>>>> CONNECTED: KeyGroupStreamPartitioner - 5 -> 8
>>>>>>>
>>>>>>> 2018-06-16 19:25:10,618 FATAL org.apache.hadoop.conf.Configuration
>>>>>>>                         - error parsing conf core-default.xml
>>>>>>>
>>>>>>> javax.xml.parsers.ParserConfigurationException: Feature '
>>>>>>> http://apache.org/xml/features/xinclude' is not recognized.
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.xerces.jaxp.DocumentBuilderFactoryImpl.newDocumentBuilder(Unknown
>>>>>>> Source)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2482)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2444)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2361)
>>>>>>>
>>>>>>> at org.apache.hadoop.conf.Configuration.get(Configuration.java:1188)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.hadoop.yarn.factory.providers.RecordFactoryProvider.getRecordFactory(RecordFactoryProvider.java:49)
>>>>>>>
>>>>>>> at org.apache.hadoop.yarn.util.Records.<clinit>(Records.java:32)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getQueueInfoRequest(YarnClientImpl.java:495)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getAllQueues(YarnClientImpl.java:525)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.yarn.AbstractYarnClusterDescriptor.checkYarnQueues(AbstractYarnClusterDescriptor.java:658)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.yarn.AbstractYarnClusterDescriptor.deployInternal(AbstractYarnClusterDescriptor.java:486)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.yarn.YarnClusterDescriptor.deployJobCluster(YarnClusterDescriptor.java:75)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.client.cli.CliFrontend.runProgram(CliFrontend.java:235)
>>>>>>>
>>>>>>> at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:210)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:1020)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.client.cli.CliFrontend.lambda$main$9(CliFrontend.java:1096)
>>>>>>>
>>>>>>> at java.security.AccessController.doPrivileged(Native Method)
>>>>>>>
>>>>>>> at javax.security.auth.Subject.doAs(Subject.java:422)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1692)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1096)
>>>>>>>
>>>>>>> 2018-06-16 19:25:10,620 WARN  
>>>>>>> org.apache.flink.yarn.AbstractYarnClusterDescriptor
>>>>>>>           - Error while getting queue information from YARN: null
>>>>>>>
>>>>>>> 2018-06-16 19:25:10,621 DEBUG
>>>>>>> org.apache.flink.yarn.AbstractYarnClusterDescriptor           -
>>>>>>> Error details
>>>>>>>
>>>>>>> java.lang.ExceptionInInitializerError
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getQueueInfoRequest(YarnClientImpl.java:495)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getAllQueues(YarnClientImpl.java:525)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.yarn.AbstractYarnClusterDescriptor.checkYarnQueues(AbstractYarnClusterDescriptor.java:658)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.yarn.AbstractYarnClusterDescriptor.deployInternal(AbstractYarnClusterDescriptor.java:486)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.yarn.YarnClusterDescriptor.deployJobCluster(YarnClusterDescriptor.java:75)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.client.cli.CliFrontend.runProgram(CliFrontend.java:235)
>>>>>>>
>>>>>>> at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:210)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:1020)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.client.cli.CliFrontend.lambda$main$9(CliFrontend.java:1096)
>>>>>>>
>>>>>>
>>>>>
>>>>> --
>>>>>
>>>>> Garvit Sharma
>>>>> github.com/garvitlnmiit/
>>>>>
>>>>> No Body is a Scholar by birth, its only hard work and strong
>>>>> determination that makes him master.
>>>>>
>>>>
>>>>
>>>
>>> --
>>>
>>> Garvit Sharma
>>> github.com/garvitlnmiit/
>>>
>>> No Body is a Scholar by birth, its only hard work and strong
>>> determination that makes him master.
>>>
>>
>
> --
>
> Garvit Sharma
> github.com/garvitlnmiit/
>
> No Body is a Scholar by birth, its only hard work and strong determination
> that makes him master.
>

Reply via email to