Why you have 2 hadoop version in the same pom file? In this case, you are not going to know which hadoop class you are actually using.
<dependency> <groupId>org.apache.hadoop</groupId> <artifactId>hadoop-client</artifactId> <version>2.3.0</version> </dependency> <dependency> <groupId>org.apache.hadoop</groupId> <artifactId>hadoop-core</artifactId> <version>1.2.1</version> </dependency> Regards, *Stanley Shi,* On Tue, Mar 4, 2014 at 1:15 AM, Margusja <mar...@roo.ee> wrote: > Hi > > 2.2.0 and 2.3.0 gave me the same container log. > > A little bit more details. > I'll try to use external java client who submits job. > some lines from maven pom.xml file: > <dependency> > <groupId>org.apache.hadoop</groupId> > <artifactId>hadoop-client</artifactId> > <version>2.3.0</version> > </dependency> > <dependency> > <groupId>org.apache.hadoop</groupId> > <artifactId>hadoop-core</artifactId> > <version>1.2.1</version> > </dependency> > > lines from external client: > ... > 2014-03-03 17:36:01 INFO FileInputFormat:287 - Total input paths to > process : 1 > 2014-03-03 17:36:02 INFO JobSubmitter:396 - number of splits:1 > 2014-03-03 17:36:03 INFO JobSubmitter:479 - Submitting tokens for job: > job_1393848686226_0018 > 2014-03-03 17:36:04 INFO YarnClientImpl:166 - Submitted application > application_1393848686226_0018 > 2014-03-03 17:36:04 INFO Job:1289 - The url to track the job: > http://vm38.dbweb.ee:8088/proxy/application_1393848686226_0018/ > 2014-03-03 17:36:04 INFO Job:1334 - Running job: job_1393848686226_0018 > 2014-03-03 17:36:10 INFO Job:1355 - Job job_1393848686226_0018 running in > uber mode : false > 2014-03-03 17:36:10 INFO Job:1362 - map 0% reduce 0% > 2014-03-03 17:36:10 INFO Job:1375 - Job job_1393848686226_0018 failed > with state FAILED due to: Application application_1393848686226_0018 failed > 2 times due to AM Container for appattempt_1393848686226_0018_000002 > exited with exitCode: 1 due to: Exception from container-launch: > org.apache.hadoop.util.Shell$ExitCodeException: > at org.apache.hadoop.util.Shell.runCommand(Shell.java:464) > at org.apache.hadoop.util.Shell.run(Shell.java:379) > at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute( > Shell.java:589) > at org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor. > launchContainer(DefaultContainerExecutor.java:195) > at org.apache.hadoop.yarn.server.nodemanager.containermanager. > launcher.ContainerLaunch.call(ContainerLaunch.java:283) > at org.apache.hadoop.yarn.server.nodemanager.containermanager. > launcher.ContainerLaunch.call(ContainerLaunch.java:79) > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > at java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPoolExecutor.java:1145) > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:744) > ... > > Lines from namenode: > ... > 14/03/03 19:12:42 INFO namenode.FSEditLog: Number of transactions: 900 > Total time for transactions(ms): 69 Number of transactions batched in > Syncs: 0 Number of syncs: 542 SyncTimes(ms): 9783 > 14/03/03 19:12:42 INFO BlockStateChange: BLOCK* addToInvalidates: blk_ > 1073742050_1226 90.190.106.33:50010 > 14/03/03 19:12:42 INFO hdfs.StateChange: BLOCK* allocateBlock: > /user/hduser/input/data666.noheader.data. > BP-802201089-90.190.106.33-1393506052071 > blk_1073742056_1232{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, > replicas=[ReplicaUnderConstruction[90.190.106.33:50010|RBW]]} > 14/03/03 19:12:44 INFO hdfs.StateChange: BLOCK* InvalidateBlocks: ask > 90.190.106.33:50010 to delete [blk_1073742050_1226] > 14/03/03 19:12:53 INFO BlockStateChange: BLOCK* addStoredBlock: blockMap > updated: 90.190.106.33:50010 is added to blk_1073742056 > _1232{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ > ReplicaUnderConstruction[90.190.106.33:50010|RBW]]} size 0 > 14/03/03 19:12:53 INFO hdfs.StateChange: DIR* completeFile: > /user/hduser/input/data666.noheader.data is closed by > DFSClient_NONMAPREDUCE_-915999412_15 > 14/03/03 19:12:54 INFO BlockStateChange: BLOCK* addToInvalidates: blk_ > 1073742051_1227 90.190.106.33:50010 > 14/03/03 19:12:54 INFO hdfs.StateChange: BLOCK* allocateBlock: > /user/hduser/input/data666.noheader.data.info. > BP-802201089-90.190.106.33-1393506052071 > blk_1073742057_1233{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, > replicas=[ReplicaUnderConstruction[90.190.106.33:50010|RBW]]} > 14/03/03 19:12:54 INFO BlockStateChange: BLOCK* addStoredBlock: blockMap > updated: 90.190.106.33:50010 is added to blk_1073742057 > _1233{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ > ReplicaUnderConstruction[90.190.106.33:50010|RBW]]} size 0 > 14/03/03 19:12:54 INFO hdfs.StateChange: DIR* completeFile: > /user/hduser/input/data666.noheader.data.info is closed by > DFSClient_NONMAPREDUCE_-915999412_15 > 14/03/03 19:12:55 INFO hdfs.StateChange: BLOCK* allocateBlock: > /user/hduser/.staging/job_1393848686226_0019/job.jar. > BP-802201089-90.190.106.33-1393506052071 blk_1073742058_1234{ > blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ > ReplicaUnderConstruction[90.190.106.33:50010|RBW]]} > 14/03/03 19:12:56 INFO hdfs.StateChange: BLOCK* InvalidateBlocks: ask > 90.190.106.33:50010 to delete [blk_1073742051_1227] > 14/03/03 19:13:12 INFO BlockStateChange: BLOCK* addStoredBlock: blockMap > updated: 90.190.106.33:50010 is added to blk_1073742058 > _1234{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ > ReplicaUnderConstruction[90.190.106.33:50010|RBW]]} size 0 > 14/03/03 19:13:12 INFO hdfs.StateChange: DIR* completeFile: > /user/hduser/.staging/job_1393848686226_0019/job.jar is closed by > DFSClient_NONMAPREDUCE_-915999412_15 > 14/03/03 19:13:12 INFO blockmanagement.BlockManager: Increasing > replication from 3 to 10 for /user/hduser/.staging/job_ > 1393848686226_0019/job.jar > 14/03/03 19:13:12 INFO blockmanagement.BlockManager: Increasing > replication from 3 to 10 for /user/hduser/.staging/job_ > 1393848686226_0019/job.split > 14/03/03 19:13:12 INFO hdfs.StateChange: BLOCK* allocateBlock: > /user/hduser/.staging/job_1393848686226_0019/job.split. > BP-802201089-90.190.106.33-1393506052071 blk_1073742059_1235{ > blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ > ReplicaUnderConstruction[90.190.106.33:50010|RBW]]} > 14/03/03 19:13:12 INFO BlockStateChange: BLOCK* addStoredBlock: blockMap > updated: 90.190.106.33:50010 is added to blk_1073742059 > _1235{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ > ReplicaUnderConstruction[90.190.106.33:50010|RBW]]} size 0 > 14/03/03 19:13:12 INFO hdfs.StateChange: DIR* completeFile: > /user/hduser/.staging/job_1393848686226_0019/job.split is closed by > DFSClient_NONMAPREDUCE_-915999412_15 > 14/03/03 19:13:12 INFO hdfs.StateChange: BLOCK* allocateBlock: > /user/hduser/.staging/job_1393848686226_0019/job.splitmetainfo. > BP-802201089-90.190.106.33-1393506052071 blk_1073742060_1236{ > blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ > ReplicaUnderConstruction[90.190.106.33:50010|RBW]]} > 14/03/03 19:13:12 INFO BlockStateChange: BLOCK* addStoredBlock: blockMap > updated: 90.190.106.33:50010 is added to blk_1073742060 > _1236{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ > ReplicaUnderConstruction[90.190.106.33:50010|RBW]]} size 0 > 14/03/03 19:13:12 INFO hdfs.StateChange: DIR* completeFile: > /user/hduser/.staging/job_1393848686226_0019/job.splitmetainfo is closed > by DFSClient_NONMAPREDUCE_-915999412_15 > 14/03/03 19:13:12 INFO hdfs.StateChange: BLOCK* allocateBlock: > /user/hduser/.staging/job_1393848686226_0019/job.xml. > BP-802201089-90.190.106.33-1393506052071 blk_1073742061_1237{ > blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ > ReplicaUnderConstruction[90.190.106.33:50010|RBW]]} > 14/03/03 19:13:13 INFO BlockStateChange: BLOCK* addStoredBlock: blockMap > updated: 90.190.106.33:50010 is added to blk_1073742061 > _1237{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ > ReplicaUnderConstruction[90.190.106.33:50010|RBW]]} size 0 > 14/03/03 19:13:13 INFO hdfs.StateChange: DIR* completeFile: > /user/hduser/.staging/job_1393848686226_0019/job.xml is closed by > DFSClient_NONMAPREDUCE_-915999412_15 > ... > > Lines from namemanager log: > ... > 2014-03-03 19:13:19,473 WARN > org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor: > Exit code from container container_1393848686226_0019_02_000001 is : 1 > 2014-03-03 19:13:19,474 WARN > org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor: > Exception from container-launch with container ID: > container_1393848686226_0019_02_000001 and exit code: 1 > org.apache.hadoop.util.Shell$ExitCodeException: > at org.apache.hadoop.util.Shell.runCommand(Shell.java:464) > at org.apache.hadoop.util.Shell.run(Shell.java:379) > at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute( > Shell.java:589) > at org.apache.hadoop.yarn.server.nodemanager. > DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java: > 195) > at org.apache.hadoop.yarn.server.nodemanager.containermanager. > launcher.ContainerLaunch.call(ContainerLaunch.java:283) > at org.apache.hadoop.yarn.server.nodemanager.containermanager. > launcher.ContainerLaunch.call(ContainerLaunch.java:79) > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > at java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPoolExecutor.java:1145) > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:744) > 2014-03-03 19:13:19,474 INFO org.apache.hadoop.yarn.server. > nodemanager.ContainerExecutor: > 2014-03-03 19:13:19,474 WARN org.apache.hadoop.yarn.server. > nodemanager.containermanager.launcher.ContainerLaunch: Container exited > with a non-zero exit code 1 > 2014-03-03 19:13:19,475 INFO org.apache.hadoop.yarn.server. > nodemanager.containermanager.container.Container: Container > container_1393848686226_0019_02_000001 transitioned from RUNNING to > EXITED_WITH_FAILURE > 2014-03-03 19:13:19,475 INFO org.apache.hadoop.yarn.server. > nodemanager.containermanager.launcher.ContainerLaunch: Cleaning up > container container_1393848686226_0019_02_000001 > 2014-03-03 19:13:19,496 INFO > org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor: > Deleting absolute path : /tmp/hadoop-hdfs/nm-local-dir/ > usercache/hduser/appcache/application_1393848686226_ > 0019/container_1393848686226_0019_02_000001 > 2014-03-03 19:13:19,498 WARN > org.apache.hadoop.yarn.server.nodemanager.NMAuditLogger: > USER=hduser OPERATION=Container Finished - Failed > TARGET=ContainerImpl RESULT=FAILURE DESCRIPTION=Container failed > with state: EXITED_WITH_FAILURE APPID=application_1393848686226_0019 > CONTAINERID=container_1393848686226_0019_02_000001 > 2014-03-03 19:13:19,498 INFO org.apache.hadoop.yarn.server. > nodemanager.containermanager.container.Container: Container > container_1393848686226_0019_02_000001 transitioned from > EXITED_WITH_FAILURE to DONE > 2014-03-03 19:13:19,498 INFO org.apache.hadoop.yarn.server. > nodemanager.containermanager.application.Application: Removing > container_1393848686226_0019_02_000001 from application > application_1393848686226_0019 > 2014-03-03 19:13:19,499 INFO org.apache.hadoop.yarn.server. > nodemanager.containermanager.AuxServices: Got event CONTAINER_STOP for > appId application_1393848686226_0019 > 2014-03-03 19:13:20,160 INFO > org.apache.hadoop.yarn.server.nodemanager.NodeStatusUpdaterImpl: > Sending out status for container: container_id { app_attempt_id { > application_id { id: 19 cluster_timestamp: 1393848686226 } attemptId: 2 } > id: 1 } state: C_COMPLETE diagnostics: "Exception from container-launch: > \norg.apache.hadoop.util.Shell$ExitCodeException: \n\tat > org.apache.hadoop.util.Shell.runCommand(Shell.java:464)\n\tat > org.apache.hadoop.util.Shell.run(Shell.java:379)\n\tat > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:589)\n\tat > org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor. > launchContainer(DefaultContainerExecutor.java:195)\n\tat > org.apache.hadoop.yarn.server.nodemanager.containermanager. > launcher.ContainerLaunch.call(ContainerLaunch.java:283)\n\tat > org.apache.hadoop.yarn.server.nodemanager.containermanager. > launcher.ContainerLaunch.call(ContainerLaunch.java:79)\n\tat > java.util.concurrent.FutureTask.run(FutureTask.java:262)\n\tat > java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPoolExecutor.java:1145)\n\tat java.util.concurrent. > ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)\n\tat > java.lang.Thread.run(Thread.java:744)\n\n\n" exit_status: 1 > 2014-03-03 19:13:20,161 INFO > org.apache.hadoop.yarn.server.nodemanager.NodeStatusUpdaterImpl: > Removed completed container container_1393848686226_0019_02_000001 > 2014-03-03 19:13:20,542 INFO org.apache.hadoop.yarn.server. > nodemanager.containermanager.monitor.ContainersMonitorImpl: Starting > resource-monitoring for container_1393848686226_0019_02_000001 > 2014-03-03 19:13:20,543 INFO org.apache.hadoop.yarn.server. > nodemanager.containermanager.monitor.ContainersMonitorImpl: Stopping > resource-monitoring for container_1393848686226_0019_02_000001 > 2014-03-03 19:13:21,164 INFO org.apache.hadoop.yarn.server. > nodemanager.containermanager.application.Application: Application > application_1393848686226_0019 transitioned from RUNNING to > APPLICATION_RESOURCES_CLEANINGUP > 2014-03-03 19:13:21,164 INFO > org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor: > Deleting absolute path : /tmp/hadoop-hdfs/nm-local-dir/ > usercache/hduser/appcache/application_1393848686226_0019 > 2014-03-03 19:13:21,165 INFO org.apache.hadoop.yarn.server. > nodemanager.containermanager.AuxServices: Got event APPLICATION_STOP for > appId application_1393848686226_0019 > 2014-03-03 19:13:21,165 INFO org.apache.hadoop.yarn.server. > nodemanager.containermanager.application.Application: Application > application_1393848686226_0019 transitioned from > APPLICATION_RESOURCES_CLEANINGUP > to FINISHED > 2014-03-03 19:13:21,165 INFO org.apache.hadoop.yarn.server. > nodemanager.containermanager.loghandler.NonAggregatingLogHandler: > Scheduling Log Deletion for application: application_1393848686226_0019, > with delay of 10800 seconds > ... > > > > Tervitades, Margus (Margusja) Roo > +372 51 48 780 > http://margus.roo.ee > http://ee.linkedin.com/in/margusroo > skype: margusja > ldapsearch -x -h ldap.sk.ee -b c=EE "(serialNumber=37303140314)" > -----BEGIN PUBLIC KEY----- > MIGfMA0GCSqGSIb3DQEBAQUAA4GNADCBiQKBgQCvbeg7LwEC2SCpAEewwpC3ajxE > 5ZsRMCB77L8bae9G7TslgLkoIzo9yOjPdx2NN6DllKbV65UjTay43uUDyql9g3tl > RhiJIcoAExkSTykWqAIPR88LfilLy1JlQ+0RD8OXiWOVVQfhOHpQ0R/jcAkM2lZa > BjM8j36yJvoBVsfOHQIDAQAB > -----END PUBLIC KEY----- > > On 03/03/14 19:05, Ted Yu wrote: > >> Can you tell us the hadoop release you're using ? >> >> Seems there is inconsistency in protobuf library. >> >> >> On Mon, Mar 3, 2014 at 8:01 AM, Margusja <mar...@roo.ee <mailto: >> mar...@roo.ee>> wrote: >> >> Hi >> >> I even don't know what information to provide but my container log is: >> >> 2014-03-03 17:36:05,311 FATAL [main] >> org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Error starting >> MRAppMaster >> java.lang.VerifyError: class >> org.apache.hadoop.yarn.proto.YarnProtos$ApplicationIdProto >> overrides final method >> getUnknownFields.()Lcom/google/protobuf/UnknownFieldSet; >> at java.lang.ClassLoader.defineClass1(Native Method) >> at java.lang.ClassLoader.defineClass(ClassLoader.java:800) >> at >> java.security.SecureClassLoader.defineClass( >> SecureClassLoader.java:142) >> at >> java.net.URLClassLoader.defineClass(URLClassLoader.java:449) >> at java.net.URLClassLoader.access$100(URLClassLoader.java:71) >> at java.net.URLClassLoader$1.run(URLClassLoader.java:361) >> at java.net.URLClassLoader$1.run(URLClassLoader.java:355) >> at java.security.AccessController.doPrivileged(Native Method) >> at java.net.URLClassLoader.findClass(URLClassLoader.java:354) >> at java.lang.ClassLoader.loadClass(ClassLoader.java:425) >> at >> sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) >> at java.lang.ClassLoader.loadClass(ClassLoader.java:358) >> at java.lang.Class.getDeclaredConstructors0(Native Method) >> at >> java.lang.Class.privateGetDeclaredConstructors(Class.java:2493) >> at java.lang.Class.getConstructor0(Class.java:2803) >> at java.lang.Class.getConstructor(Class.java:1718) >> at >> org.apache.hadoop.yarn.factories.impl.pb.RecordFactoryPBImpl. >> newRecordInstance(RecordFactoryPBImpl.java:62) >> at >> org.apache.hadoop.yarn.util.Records.newRecord(Records.java:36) >> at >> org.apache.hadoop.yarn.api.records.ApplicationId. >> newInstance(ApplicationId.java:49) >> at >> org.apache.hadoop.yarn.util.ConverterUtils.toApplicationAttemptId( >> ConverterUtils.java:137) >> at >> org.apache.hadoop.yarn.util.ConverterUtils.toContainerId( >> ConverterUtils.java:177) >> at >> org.apache.hadoop.mapreduce.v2.app.MRAppMaster.main( >> MRAppMaster.java:1343) >> >> >> Where to start digging? >> >> -- Tervitades, Margus (Margusja) Roo >> +372 51 48 780 <tel:%2B372%2051%2048%20780> >> >> http://margus.roo.ee >> http://ee.linkedin.com/in/margusroo >> skype: margusja >> ldapsearch -x -h ldap.sk.ee <http://ldap.sk.ee> -b c=EE >> >> "(serialNumber=37303140314)" >> -----BEGIN PUBLIC KEY----- >> MIGfMA0GCSqGSIb3DQEBAQUAA4GNADCBiQKBgQCvbeg7LwEC2SCpAEewwpC3ajxE >> 5ZsRMCB77L8bae9G7TslgLkoIzo9yOjPdx2NN6DllKbV65UjTay43uUDyql9g3tl >> RhiJIcoAExkSTykWqAIPR88LfilLy1JlQ+0RD8OXiWOVVQfhOHpQ0R/jcAkM2lZa >> BjM8j36yJvoBVsfOHQIDAQAB >> -----END PUBLIC KEY----- >> >> >> >