[ https://issues.apache.org/jira/browse/GIRAPH-583?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13612180#comment-13612180 ]
Gianmarco De Francisci Morales commented on GIRAPH-583: ------------------------------------------------------- I tried restarting the process from scratch (recompiling Giraph, re-authenticating, etc...) and now I manage to start the job. However I get a different error message, which might be unrelated: {code} 2013-03-24 18:02:13,528 FATAL [IPC Server handler 3 on 48891] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Task: attempt_1362131148734_326910_m_000000_0 - exited : java.lang.IllegalStateException: run: Caught an unrecoverable exception java.io.FileNotFoundException: File _bsp/_defaultZkManagerDir/job_1362131148734_326910/_zkServer does not exist. at org.apache.giraph.graph.GraphMapper.run(GraphMapper.java:102) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:726) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:333) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:157) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1221) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:152) Caused by: java.lang.RuntimeException: java.io.FileNotFoundException: File _bsp/_defaultZkManagerDir/job_1362131148734_326910/_zkServer does not exist. at org.apache.giraph.zk.ZooKeeperManager.onlineZooKeeperServers(ZooKeeperManager.java:795) at org.apache.giraph.graph.GraphTaskManager.startZooKeeperManager(GraphTaskManager.java:333) at org.apache.giraph.graph.GraphTaskManager.setup(GraphTaskManager.java:182) at org.apache.giraph.graph.GraphMapper.setup(GraphMapper.java:60) at org.apache.giraph.graph.GraphMapper.run(GraphMapper.java:90) ... 7 more Caused by: java.io.FileNotFoundException: File _bsp/_defaultZkManagerDir/job_1362131148734_326910/_zkServer does not exist. at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:362) at org.apache.giraph.zk.ZooKeeperManager.onlineZooKeeperServers(ZooKeeperManager.java:754) ... 11 more 2013-03-24 18:02:13,528 INFO [IPC Server handler 3 on 48891] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Diagnostics report from attempt_1362131148734_326910_m_000000_0: Error: java.lang.IllegalStateException: run: Caught an unrecoverable exception java.io.FileNotFoundException: File _bsp/_defaultZkManagerDir/job_1362131148734_326910/_zkServer does not exist. at org.apache.giraph.graph.GraphMapper.run(GraphMapper.java:102) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:726) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:333) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:157) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1221) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:152) Caused by: java.lang.RuntimeException: java.io.FileNotFoundException: File _bsp/_defaultZkManagerDir/job_1362131148734_326910/_zkServer does not exist. at org.apache.giraph.zk.ZooKeeperManager.onlineZooKeeperServers(ZooKeeperManager.java:795) at org.apache.giraph.graph.GraphTaskManager.startZooKeeperManager(GraphTaskManager.java:333) at org.apache.giraph.graph.GraphTaskManager.setup(GraphTaskManager.java:182) at org.apache.giraph.graph.GraphMapper.setup(GraphMapper.java:60) at org.apache.giraph.graph.GraphMapper.run(GraphMapper.java:90) ... 7 more Caused by: java.io.FileNotFoundException: File _bsp/_defaultZkManagerDir/job_1362131148734_326910/_zkServer does not exist. at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:362) at org.apache.giraph.zk.ZooKeeperManager.onlineZooKeeperServers(ZooKeeperManager.java:754) ... 11 more {code} I guess we can close this Jira, as it seems a temporary problem of my environment. Though I would appreciate any help in solving this new issue. > Problem with authentication on Hadoop 0.23 > ------------------------------------------ > > Key: GIRAPH-583 > URL: https://issues.apache.org/jira/browse/GIRAPH-583 > Project: Giraph > Issue Type: Bug > Reporter: Gianmarco De Francisci Morales > > Hi, > I am trying to run the RWR code on trunk and Hadoop 0.23 with Kerberos > authentication, but I get this exception: > {code} > 13/03/23 17:32:36 ERROR security.UserGroupInformation: > PriviledgedActionException as:gdfm (auth:KERBEROS) > cause:javax.security.sasl.SaslException: GSS initiate failed [Caused by > GSSException: No valid credentials provided (Mechanism level: Failed to find > any Kerberos tgt)] > 13/03/23 17:32:36 WARN ipc.Client: Exception encountered while connecting to > the server : javax.security.sasl.SaslException: GSS initiate failed [Caused > by GSSException: No valid credentials provided (Mechanism level: Failed to > find any Kerberos tgt)] > 13/03/23 17:32:36 ERROR security.UserGroupInformation: > PriviledgedActionException as:gdfm (auth:KERBEROS) cause:java.io.IOException: > javax.security.sasl.SaslException: GSS initiate failed [Caused by > GSSException: No valid credentials provided (Mechanism level: Failed to find > any Kerberos tgt)] > 13/03/23 17:32:36 ERROR security.UserGroupInformation: > PriviledgedActionException as:gdfm (auth:KERBEROS) cause:java.io.IOException: > Failed on local exception: java.io.IOException: > javax.security.sasl.SaslException: GSS initiate failed [Caused by > GSSException: No valid credentials provided (Mechanism level: Failed to find > any Kerberos tgt)]; Host Details : local host is: > "gwta3005.tan.ygrid.yahoo.com/98.138.127.244"; destination host is: > ""tiberiumtan-nn1.tan.ygrid.yahoo.com":8020; > Exception in thread "main" java.io.IOException: Failed on local exception: > java.io.IOException: javax.security.sasl.SaslException: GSS initiate failed > [Caused by GSSException: No valid credentials provided (Mechanism level: > Failed to find any Kerberos tgt)]; Host Details : local host is: > "gwta3005.tan.ygrid.yahoo.com/98.138.127.244"; destination host is: > ""tiberiumtan-nn1.tan.ygrid.yahoo.com":8020; > at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:738) > at org.apache.hadoop.ipc.Client.call(Client.java:1092) > at > org.apache.hadoop.ipc.WritableRpcEngine$Invoker.invoke(WritableRpcEngine.java:195) > at $Proxy6.getDelegationToken(Unknown Source) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:102) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:67) > at $Proxy6.getDelegationToken(Unknown Source) > at > org.apache.hadoop.hdfs.DFSClient.getDelegationToken(DFSClient.java:603) > at > org.apache.hadoop.hdfs.DistributedFileSystem.getDelegationToken(DistributedFileSystem.java:786) > at > org.apache.hadoop.fs.FileSystem.collectDelegationTokens(FileSystem.java:466) > at > org.apache.hadoop.fs.FileSystem.addDelegationTokens(FileSystem.java:444) > at > org.apache.hadoop.mapreduce.security.TokenCache.obtainTokensForNamenodesInternal(TokenCache.java:122) > at > org.apache.hadoop.mapreduce.security.TokenCache.obtainTokensForNamenodesInternal(TokenCache.java:101) > at > org.apache.hadoop.mapreduce.security.TokenCache.obtainTokensForNamenodes(TokenCache.java:81) > at > org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.checkOutputSpecs(FileOutputFormat.java:137) > at > org.apache.giraph.io.formats.TextVertexOutputFormat.checkOutputSpecs(TextVertexOutputFormat.java:55) > at > org.apache.giraph.bsp.BspOutputFormat.checkOutputSpecs(BspOutputFormat.java:50) > at > org.apache.hadoop.mapreduce.JobSubmitter.checkSpecs(JobSubmitter.java:417) > at > org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:332) > at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1218) > at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1215) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1221) > at org.apache.hadoop.mapreduce.Job.submit(Job.java:1215) > at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1236) > at org.apache.giraph.job.GiraphJob.run(GiraphJob.java:237) > at org.apache.giraph.GiraphRunner.run(GiraphRunner.java:81) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:69) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:83) > at org.apache.giraph.GiraphRunner.main(GiraphRunner.java:111) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at org.apache.hadoop.util.RunJar.main(RunJar.java:208) > Caused by: java.io.IOException: javax.security.sasl.SaslException: GSS > initiate failed [Caused by GSSException: No valid credentials provided > (Mechanism level: Failed to find any Kerberos tgt)] > at org.apache.hadoop.ipc.Client$Connection$1.run(Client.java:534) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1221) > at > org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:498) > at > org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:582) > at org.apache.hadoop.ipc.Client$Connection.access$2000(Client.java:205) > at org.apache.hadoop.ipc.Client.getConnection(Client.java:1198) > at org.apache.hadoop.ipc.Client.call(Client.java:1068) > ... 38 more > Caused by: javax.security.sasl.SaslException: GSS initiate failed [Caused by > GSSException: No valid credentials provided (Mechanism level: Failed to find > any Kerberos tgt)] > at > com.sun.security.sasl.gsskerb.GssKrb5Client.evaluateChallenge(GssKrb5Client.java:212) > at > org.apache.hadoop.security.SaslRpcClient.saslConnect(SaslRpcClient.java:137) > at > org.apache.hadoop.ipc.Client$Connection.setupSaslConnection(Client.java:406) > at org.apache.hadoop.ipc.Client$Connection.access$1200(Client.java:205) > at org.apache.hadoop.ipc.Client$Connection$2.run(Client.java:575) > at org.apache.hadoop.ipc.Client$Connection$2.run(Client.java:572) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1221) > at > org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:571) > ... 41 more > Caused by: GSSException: No valid credentials provided (Mechanism level: > Failed to find any Kerberos tgt) > at > sun.security.jgss.krb5.Krb5InitCredential.getInstance(Krb5InitCredential.java:147) > at > sun.security.jgss.krb5.Krb5MechFactory.getCredentialElement(Krb5MechFactory.java:121) > at > sun.security.jgss.krb5.Krb5MechFactory.getMechanismContext(Krb5MechFactory.java:187) > at > sun.security.jgss.GSSManagerImpl.getMechanismContext(GSSManagerImpl.java:223) > at > sun.security.jgss.GSSContextImpl.initSecContext(GSSContextImpl.java:212) > at > sun.security.jgss.GSSContextImpl.initSecContext(GSSContextImpl.java:179) > at > com.sun.security.sasl.gsskerb.GssKrb5Client.evaluateChallenge(GssKrb5Client.java:193) > ... 50 more > {code} > For reference, I am using -Dgiraph.useNetty="true" > -Dgiraph.authenticate="true" -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira