Hi Mukesh,

Those line numbers in ConverterUtils in the stack trace don't appear to
line up with CDH 5.3:
https://github.com/cloudera/hadoop-common/blob/cdh5-2.5.0_5.3.0/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/util/ConverterUtils.java

Is it possible you're still including the old jars on the classpath in some
way?

-Sandy

On Thu, Jan 8, 2015 at 3:38 AM, Mukesh Jha <me.mukesh....@gmail.com> wrote:

> Hi Experts,
>
> I am running spark inside YARN job.
>
> The spark-streaming job is running fine in CDH-5.0.0 but after the upgrade
> to 5.3.0 it cannot fetch containers with the below errors. Looks like the
> container id is incorrect and a string is present in a pace where it's
> expecting a number.
>
>
>
> java.lang.IllegalArgumentException: Invalid ContainerId:
> container_e01_1420481081140_0006_01_000001
>
> Caused by: java.lang.NumberFormatException: For input string: "e01"
>
>
>
> Is this a bug?? Did you face something similar and any ideas how to fix
> this?
>
>
>
> 15/01/08 09:50:28 INFO yarn.ApplicationMaster: Registered signal handlers
> for [TERM, HUP, INT]
>
> 15/01/08 09:50:29 ERROR yarn.ApplicationMaster: Uncaught exception:
>
> java.lang.IllegalArgumentException: Invalid ContainerId:
> container_e01_1420481081140_0006_01_000001
>
> at
> org.apache.hadoop.yarn.util.ConverterUtils.toContainerId(ConverterUtils.java:182)
>
> at
> org.apache.spark.deploy.yarn.YarnRMClientImpl.getAttemptId(YarnRMClientImpl.scala:79)
>
> at
> org.apache.spark.deploy.yarn.ApplicationMaster.run(ApplicationMaster.scala:79)
>
> at
> org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$main$1.apply$mcV$sp(ApplicationMaster.scala:515)
>
> at
> org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:60)
>
> at
> org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:59)
>
> at java.security.AccessController.doPrivileged(Native Method)
>
> at javax.security.auth.Subject.doAs(Subject.java:415)
>
> at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>
> at
> org.apache.spark.deploy.SparkHadoopUtil.runAsSparkUser(SparkHadoopUtil.scala:59)
>
> at
> org.apache.spark.deploy.yarn.ApplicationMaster$.main(ApplicationMaster.scala:513)
>
> at
> org.apache.spark.deploy.yarn.ApplicationMaster.main(ApplicationMaster.scala)
>
> Caused by: java.lang.NumberFormatException: For input string: "e01"
>
> at
> java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)
>
> at java.lang.Long.parseLong(Long.java:441)
>
> at java.lang.Long.parseLong(Long.java:483)
>
> at
> org.apache.hadoop.yarn.util.ConverterUtils.toApplicationAttemptId(ConverterUtils.java:137)
>
> at
> org.apache.hadoop.yarn.util.ConverterUtils.toContainerId(ConverterUtils.java:177)
>
> ... 11 more
>
> 15/01/08 09:50:29 INFO yarn.ApplicationMaster: Final app status: FAILED,
> exitCode: 10, (reason: Uncaught exception: Invalid ContainerId:
> container_e01_1420481081140_0006_01_000001)
>
> --
> Thanks & Regards,
>
> *Mukesh Jha <me.mukesh....@gmail.com>*
>

Reply via email to