I have enabled Kerberos debugging in Hadoop command line, so when i run the
"hadoop fs -ls /" command, i get following output, I can't interpret this.
Can you please tell me is something wrong with Kerberos configuration or
everything is fine ?


16/08/02 18:34:10 DEBUG util.Shell: setsid exited with exit code 0
16/08/02 18:34:10 DEBUG conf.Configuration: parsing URL
jar:file:/usr/local/hadoop/share/hadoop/common/hadoop-common-2.7.2.jar!/core-default.xml
16/08/02 18:34:10 DEBUG conf.Configuration: parsing input stream
sun.net.www.protocol.jar.JarURLConnection$JarURLInputStream@4fbc7b65
16/08/02 18:34:10 DEBUG conf.Configuration: parsing URL
file:/usr/local/hadoop/etc/hadoop/core-site.xml
16/08/02 18:34:10 DEBUG conf.Configuration: parsing input stream
java.io.BufferedInputStream@69c1adfa
16/08/02 18:34:11 DEBUG lib.MutableMetricsFactory: field
org.apache.hadoop.metrics2.lib.MutableRate
org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginSuccess
with annotation
@org.apache.hadoop.metrics2.annotation.Metric(valueName=Time, value=[Rate
of successful kerberos logins and latency (milliseconds)], about=,
always=false, type=DEFAULT, sampleName=Ops)
16/08/02 18:34:11 DEBUG lib.MutableMetricsFactory: field
org.apache.hadoop.metrics2.lib.MutableRate
org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginFailure
with annotation
@org.apache.hadoop.metrics2.annotation.Metric(valueName=Time, value=[Rate
of failed kerberos logins and latency (milliseconds)], about=,
always=false, type=DEFAULT, sampleName=Ops)
16/08/02 18:34:11 DEBUG lib.MutableMetricsFactory: field
org.apache.hadoop.metrics2.lib.MutableRate
org.apache.hadoop.security.UserGroupInformation$UgiMetrics.getGroups with
annotation @org.apache.hadoop.metrics2.annotation.Metric(valueName=Time,
value=[GetGroups], about=, always=false, type=DEFAULT, sampleName=Ops)
16/08/02 18:34:11 DEBUG impl.MetricsSystemImpl: UgiMetrics, User and group
related metrics
Java config name: null
Native config name: /etc/krb5.conf
Loaded from native config
16/08/02 18:34:11 DEBUG security.Groups:  Creating new Groups object
16/08/02 18:34:11 DEBUG security.Groups: Group mapping
impl=org.apache.hadoop.security.LdapGroupsMapping; cacheTimeout=300000;
warningDeltaMs=5000
>>>KinitOptions cache name is /tmp/krb5cc_0
>>>DEBUG <CCacheInputStream>  client principal is
nn/hadoop-master@platalyticsrealm
>>>DEBUG <CCacheInputStream> server principal is
krbtgt/platalyticsrealm@platalyticsrealm
>>>DEBUG <CCacheInputStream> key type: 16
>>>DEBUG <CCacheInputStream> auth time: Tue Aug 02 18:23:59 PKT 2016
>>>DEBUG <CCacheInputStream> start time: Tue Aug 02 18:23:59 PKT 2016
>>>DEBUG <CCacheInputStream> end time: Wed Aug 03 06:23:59 PKT 2016
>>>DEBUG <CCacheInputStream> renew_till time: Tue Aug 09 18:23:59 PKT 2016
>>> CCacheInputStream: readFlags()  FORWARDABLE; RENEWABLE; INITIAL;
>>>DEBUG <CCacheInputStream>  client principal is
nn/hadoop-master@platalyticsrealm
>>>DEBUG <CCacheInputStream> server principal is
X-CACHECONF:/krb5_ccache_conf_data/fast_avail/krbtgt/platalyticsrealm@platalyticsrealm
>>>DEBUG <CCacheInputStream> key type: 0
>>>DEBUG <CCacheInputStream> auth time: Thu Jan 01 05:00:00 PKT 1970
>>>DEBUG <CCacheInputStream> start time: null
>>>DEBUG <CCacheInputStream> end time: Thu Jan 01 05:00:00 PKT 1970
>>>DEBUG <CCacheInputStream> renew_till time: null
>>> CCacheInputStream: readFlags()
16/08/02 18:34:11 DEBUG security.UserGroupInformation: hadoop login
16/08/02 18:34:11 DEBUG security.UserGroupInformation: hadoop login commit
16/08/02 18:34:11 DEBUG security.UserGroupInformation: using kerberos
user:nn/hadoop-master@platalyticsrealm
16/08/02 18:34:11 DEBUG security.UserGroupInformation: Using user:
"nn/hadoop-master@platalyticsrealm" with name
nn/hadoop-master@platalyticsrealm
16/08/02 18:34:11 DEBUG security.UserGroupInformation: User entry:
"nn/hadoop-master@platalyticsrealm"
16/08/02 18:34:11 DEBUG security.UserGroupInformation: UGI
loginUser:nn/hadoop-master@platalyticsrealm (auth:KERBEROS)
16/08/02 18:34:12 DEBUG security.UserGroupInformation: Found tgt Ticket
(hex) =
0000: 61 82 01 72 30 82 01 6E   A0 03 02 01 05 A1 12 1B  a..r0..n........
0010: 10 70 6C 61 74 61 6C 79   74 69 63 73 72 65 61 6C  .platalyticsreal
0020: 6D A2 25 30 23 A0 03 02   01 02 A1 1C 30 1A 1B 06  m.%0#.......0...
0030: 6B 72 62 74 67 74 1B 10   70 6C 61 74 61 6C 79 74  krbtgt..platalyt
0040: 69 63 73 72 65 61 6C 6D   A3 82 01 2A 30 82 01 26  icsrealm...*0..&
0050: A0 03 02 01 10 A1 03 02   01 01 A2 82 01 18 04 82  ................
0060: 01 14 A5 A9 41 A6 B7 0E   8F 70 F4 03 41 64 8D DC  ....A....p..Ad..
0070: 78 2F FB 08 58 C9 39 44   CF D0 8D B0 85 09 62 8C  x/..X.9D......b.
0080: 40 CF 45 13 D3 B9 CD 38   84 92 33 24 B2 0D C1 65  @.E....8..3$...e
0090: C7 1B 0D 3E F2 92 A2 8B   58 34 77 5F F6 E3 AA B6  ...>....X4w_....
00A0: EB 8E 58 46 AC 54 DB 9B   79 3E ED A1 83 0C D3 D3  ..XF.T..y>......
00B0: 02 8B 42 52 6D 92 F1 39   BA E7 56 D4 BA A6 03 B6  ..BRm..9..V.....
00C0: 16 5A DC 1A 69 F4 DF A5   CD F6 48 AC 08 32 D3 AD  .Z..i.....H..2..
00D0: 22 8E E9 52 00 93 78 41   1C 26 4F 0B 42 2C EF E9  "..R..xA.&O.B,..
00E0: B8 0E 84 39 E4 AF 3A 60   7D 04 EE 70 18 C0 E7 21  ...9..:`...p...!
00F0: 0B 70 18 42 33 5E D9 CA   94 C0 6F 6A C0 39 72 7B  .p.B3^....oj.9r.
0100: FD 6E F1 09 CE 2D 02 EA   DA 52 5C 1B B2 18 36 0E  .n...-...R\...6.
0110: 54 94 DD 7A 47 A8 F2 36   53 18 3D D7 5C 68 58 71  T..zG..6S.=.\hXq
0120: 63 DB 36 88 B9 87 62 DC   BA 86 C3 F0 55 05 D8 15  c.6...b.....U...
0130: 6E 70 FD 8E 64 63 3D 51   36 EC 9E 63 30 77 BE 98  np..dc=Q6..c0w..
0140: 1D A0 DC 97 04 6F 03 AB   12 52 F8 68 7C 6C D0 88  .....o...R.h.l..
0150: 16 FC 17 69 3E 02 4B 59   E8 22 B3 1B 13 70 B2 6A  ...i>.KY."...p.j
0160: 3F 05 3B 1C 91 3D 03 A8   30 64 1C B1 59 42 17 FB  ?.;..=..0d..YB..
0170: 1B B2 76 E0 BC 49                                  ..v..I

Client Principal = nn/hadoop-master@platalyticsrealm
Server Principal = krbtgt/platalyticsrealm@platalyticsrealm
Session Key = EncryptionKey: keyType=16 keyBytes (hex dump)=
0000: B5 4A 9B 0E 1C 6D 1C 34   D5 DF DA F2 9D 4C C2 FE  .J...m.4.....L..
0010: D9 0D 67 A2 79 6D 8C 0D                            ..g.ym..


Forwardable Ticket true
Forwarded Ticket false
Proxiable Ticket false
Proxy Ticket false
Postdated Ticket false
Renewable Ticket true
Initial Ticket true
Auth Time = Tue Aug 02 18:23:59 PKT 2016
Start Time = Tue Aug 02 18:23:59 PKT 2016
End Time = Wed Aug 03 06:23:59 PKT 2016
Renew Till = Tue Aug 09 18:23:59 PKT 2016
Client Addresses  Null
16/08/02 18:34:12 DEBUG security.UserGroupInformation: Current time is
1470144852023
16/08/02 18:34:12 DEBUG security.UserGroupInformation: Next refresh is
1470178799000
16/08/02 18:34:12 TRACE tracing.SpanReceiverHost: No span receiver names
found in dfs.client.htrace.spanreceiver.classes.
16/08/02 18:34:12 DEBUG hdfs.BlockReaderLocal:
dfs.client.use.legacy.blockreader.local = false
16/08/02 18:34:12 DEBUG hdfs.BlockReaderLocal: dfs.client.read.shortcircuit
= false
16/08/02 18:34:12 DEBUG hdfs.BlockReaderLocal:
dfs.client.domain.socket.data.traffic = false
16/08/02 18:34:12 DEBUG hdfs.BlockReaderLocal: dfs.domain.socket.path =
16/08/02 18:34:12 DEBUG retry.RetryUtils: multipleLinearRandomRetry = null
16/08/02 18:34:12 DEBUG ipc.Server: rpcKind=RPC_PROTOCOL_BUFFER,
rpcRequestWrapperClass=class
org.apache.hadoop.ipc.ProtobufRpcEngine$RpcRequestWrapper,
rpcInvoker=org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker@4219a40f
16/08/02 18:34:12 DEBUG ipc.Client: getting client out of cache:
org.apache.hadoop.ipc.Client@5e0df7af
16/08/02 18:34:13 DEBUG util.NativeCodeLoader: Trying to load the
custom-built native-hadoop library...
16/08/02 18:34:13 DEBUG util.NativeCodeLoader: Loaded the native-hadoop
library
16/08/02 18:34:13 DEBUG unix.DomainSocketWatcher:
org.apache.hadoop.net.unix.DomainSocketWatcher$2@1a1ff7d1: starting with
interruptCheckPeriodMs = 60000
16/08/02 18:34:13 TRACE unix.DomainSocketWatcher:
DomainSocketWatcher(1934811148): adding notificationSocket 191, connected
to 190
16/08/02 18:34:13 DEBUG util.PerformanceAdvisory: Both short-circuit local
reads and UNIX domain socket are disabled.
16/08/02 18:34:13 DEBUG sasl.DataTransferSaslUtil: DataTransferProtocol not
using SaslPropertiesResolver, no QOP found in configuration for
dfs.data.transfer.protection
16/08/02 18:34:13 TRACE ipc.ProtobufRpcEngine: 1: Call -> /
192.168.23.206:8020: getFileInfo {src: "/"}
16/08/02 18:34:13 DEBUG ipc.Client: The ping interval is 60000 ms.
16/08/02 18:34:13 DEBUG ipc.Client: Connecting to /192.168.23.206:8020
16/08/02 18:34:13 DEBUG security.UserGroupInformation: PrivilegedAction
as:nn/hadoop-master@platalyticsrealm (auth:KERBEROS)
from:org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:724)
16/08/02 18:34:13 DEBUG security.SaslRpcClient: Sending sasl message state:
NEGOTIATE

16/08/02 18:34:13 DEBUG security.SaslRpcClient: Received SASL message
state: NEGOTIATE
auths {
  method: "TOKEN"
  mechanism: "DIGEST-MD5"
  protocol: ""
  serverId: "default"
  challenge:
"realm=\"default\",nonce=\"xHi0jI3ZHzKXd2aQ0Gqx4N1qcgbdJAWBCa36ZeSO\",qop=\"auth\",charset=utf-8,algorithm=md5-sess"
}
auths {
  method: "KERBEROS"
  mechanism: "GSSAPI"
  protocol: "nn"
  serverId: "hadoop-master"
}

16/08/02 18:34:13 DEBUG security.SaslRpcClient: Get token info
proto:interface org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB
info:@org.apache.hadoop.security.token.TokenInfo(value=class
org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenSelector)
16/08/02 18:34:13 DEBUG security.SaslRpcClient: Get kerberos info
proto:interface org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB
info:@org.apache.hadoop.security.KerberosInfo(clientPrincipal=,
serverPrincipal=dfs.namenode.kerberos.principal)
16/08/02 18:34:13 DEBUG security.SaslRpcClient: RPC Server's Kerberos
principal name for
protocol=org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB is
nn/hadoop-master@platalyticsrealm
16/08/02 18:34:13 DEBUG security.SaslRpcClient: Creating SASL
GSSAPI(KERBEROS)  client to authenticate to service at hadoop-master
16/08/02 18:34:13 DEBUG security.SaslRpcClient: Use KERBEROS authentication
for protocol ClientNamenodeProtocolPB
Found ticket for nn/hadoop-master@platalyticsrealm to go to
krbtgt/platalyticsrealm@platalyticsrealm expiring on Wed Aug 03 06:23:59
PKT 2016
Entered Krb5Context.initSecContext with state=STATE_NEW
Found ticket for nn/hadoop-master@platalyticsrealm to go to
krbtgt/platalyticsrealm@platalyticsrealm expiring on Wed Aug 03 06:23:59
PKT 2016
Service ticket not found in the subject
>>> Credentials acquireServiceCreds: same realm
Using builtin default etypes for default_tgs_enctypes
default etypes for default_tgs_enctypes: 18 17 16 23 1 3.
>>> CksumType: sun.security.krb5.internal.crypto.RsaMd5CksumType
>>> EType: sun.security.krb5.internal.crypto.Des3CbcHmacSha1KdEType
>>> KdcAccessibility: reset
>>> KrbKdcReq send: kdc=platalytics.com UDP:88, timeout=30000, number of
retries =3, #bytes=727
>>> KDCCommunication: kdc=platalytics.com UDP:88, timeout=30000,Attempt =1,
#bytes=727
>>> KrbKdcReq send: #bytes read=686
>>> KdcAccessibility: remove platalytics.com
>>> EType: sun.security.krb5.internal.crypto.Des3CbcHmacSha1KdEType
>>> KrbApReq: APOptions are 00100000 00000000 00000000 00000000
>>> EType: sun.security.krb5.internal.crypto.Des3CbcHmacSha1KdEType
Krb5Context setting mySeqNumber to: 822249937
Created InitSecContextToken:
0000: 01 00 6E 82 02 67 30 82   02 63 A0 03 02 01 05 A1  ..n..g0..c......
0010: 03 02 01 0E A2 07 03 05   00 20 00 00 00 A3 82 01  ......... ......
0020: 6F 61 82 01 6B 30 82 01   67 A0 03 02 01 05 A1 12  oa..k0..g.......
0030: 1B 10 70 6C 61 74 61 6C   79 74 69 63 73 72 65 61  ..platalyticsrea
0040: 6C 6D A2 1E 30 1C A0 03   02 01 00 A1 15 30 13 1B  lm..0........0..
0050: 02 6E 6E 1B 0D 68 61 64   6F 6F 70 2D 6D 61 73 74  .nn..hadoop-mast
0060: 65 72 A3 82 01 2A 30 82   01 26 A0 03 02 01 10 A1  er...*0..&......
0070: 03 02 01 04 A2 82 01 18   04 82 01 14 25 56 29 BE  ............%V).
0080: 2E AA 50 55 7B 2C 5C AC   BA 64 2D 4D 8D 9C 71 B1  ..PU.,\..d-M..q.
0090: 1A 99 14 81 4C 98 80 B2   65 86 6C 37 61 67 31 D1  ....L...e.l7ag1.
00A0: 6F F6 E7 7A F3 92 A5 9A   F0 BA A5 BE 1C 15 7F 14  o..z............
00B0: 85 7E B0 7A 81 3D 9C B6   00 80 43 00 2A 0C 89 6A  ...z.=....C.*..j
00C0: B1 49 EF 27 F9 97 A1 3E   5C 80 B7 0D 49 6C E0 A3  .I.'...>\...Il..
00D0: 73 BC C2 69 AE 92 88 26   C5 DA FD 6E AB 55 F7 60  s..i...&...n.U.`
00E0: D0 7E 3A A5 5D 78 4E 3F   3D 96 44 6B B9 8F EA D8  ..:.]xN?=.Dk....
00F0: 4E BA 70 F3 5C 25 4E ED   AD E2 76 09 FF 36 D8 6D  N.p.\%N...v..6.m
0100: A4 22 C3 93 10 04 04 F2   6C D4 04 C9 A9 14 95 47  ."......l......G
0110: 16 BA 62 6F 58 5F 4F 8E   38 23 A5 5C 1D 58 F8 D5  ..boX_O.8#.\.X..
0120: 87 23 3D 7F 0B A7 BE 18   25 1F F1 7B 4C 54 EC BD  .#=.....%...LT..
0130: A6 D4 05 4C 82 03 64 FD   5A 4E 24 D8 71 D5 5A 15  ...L..d.ZN$.q.Z.
0140: 4C 2E E3 12 88 19 19 09   C1 F9 31 9D 6E CE D4 6F  L.........1.n..o
0150: 7A 20 F6 82 BB F6 28 D1   ED A3 54 69 01 9E A4 4C  z ....(...Ti...L
0160: 40 E2 E0 FC F5 35 44 C1   25 8C 50 1F C0 01 1D C0  @....5D.%.P.....
0170: 63 A5 45 B8 56 DF F7 F8   CA 86 8B 96 0C 5C 49 EA  c.E.V........\I.
0180: F0 A9 70 9C 2E 0E 36 57   65 47 97 09 8C 24 F1 00  ..p...6WeG...$..
0190: A4 81 DA 30 81 D7 A0 03   02 01 10 A2 81 CF 04 81  ...0............
01A0: CC F1 F6 BE 3A A7 C0 1A   04 D0 72 DE 57 94 D1 FE  ....:.....r.W...
01B0: 16 7E E8 09 72 D7 83 54   B3 1C 98 59 36 86 78 12  ....r..T...Y6.x.
01C0: A5 02 E3 B6 8C C6 83 B5   C9 7C 53 A3 C9 79 AF C8  ..........S..y..
01D0: B8 1A B3 B2 A6 7E 02 1A   A5 9C 41 EA 08 87 A8 E5  ..........A.....
01E0: D1 0E ED 69 5C CA 33 63   24 C8 4B E1 57 D5 C3 AF  ...i\.3c$.K.W...
01F0: 39 0A DE F6 9F 63 3B 44   79 5B 29 F7 9A B0 2E 8B  9....c;Dy[).....
0200: 1C EF 4A 0B D9 3A 55 75   C5 38 B7 5C 50 11 0E 74  ..J..:Uu.8.\P..t
0210: BE 57 DC 70 30 DD AF 14   35 97 1C 14 11 70 46 FD  .W.p0...5....pF.
0220: F9 8C 14 60 DE 35 D8 DC   81 86 C7 31 1F F8 6A 65  ...`.5.....1..je
0230: 2D B7 8A EF F2 61 21 00   2C 8D 4F 3A 49 1E 24 80  -....a!.,.O:I.$.
0240: FA 56 D0 2D 0E 52 AE 29   2B 6A 4A C7 16 8F B5 D8  .V.-.R.)+jJ.....
0250: EC 41 18 03 34 F2 D8 94   79 82 C8 0D E2 10 72 39  .A..4...y.....r9
0260: 85 B9 F7 BB 54 5C 71 21   49 23 A5 4A D0           ....T\q!I#.J.

16/08/02 18:34:13 DEBUG security.SaslRpcClient: Sending sasl message state:
INITIATE
token:
"`\202\002x\006\t*\206H\206\367\022\001\002\002\001\000n\202\002g0\202\002c\240\003\002\001\005\241\003\002\001\016\242\a\003\005\000
\000\000\000\243\202\001oa\202\001k0\202\001g\240\003\002\001\005\241\022\033\020platalyticsrealm\242\0360\034\240\003\002\001\000\241\0250\023\033\002nn\033\rhadoop-master\243\202\001*0\202\001&\240\003\002\001\020\241\003\002\001\004\242\202\001\030\004\202\001\024%V)\276.\252PU{,\\\254\272d-M\215\234q\261\032\231\024\201L\230\200\262e\206l7ag1\321o\366\347z\363\222\245\232\360\272\245\276\034\025
\024\205~\260z\201=\234\266\000\200C\000*\f\211j\261I\357\'\371\227\241>\\\200\267\rIl\340\243s\274\302i\256\222\210&\305\332\375n\253U\367`\320~:\245]xN?=\226Dk\271\217\352\330N\272p\363\\%N\355\255\342v\t\3776\330m\244\"\303\223\020\004\004\362l\324\004\311\251\024\225G\026\272boX_O\2168#\245\\\035X\370\325\207#=
\v\247\276\030%\037\361{LT\354\275\246\324\005L\202\003d\375ZN$\330q\325Z\025L.\343\022\210\031\031\t\301\3711\235n\316\324oz
\366\202\273\366(\321\355\243Ti\001\236\244L@
\342\340\374\3655D\301%\214P\037\300\001\035\300c\245E\270V\337\367\370\312\206\213\226\f\\I\352\360\251p\234.\0166WeG\227\t\214$\361\000\244\201\3320\201\327\240\003\002\001\020\242\201\317\004\201\314\361\366\276:\247\300\032\004\320r\336W\224\321\376\026~\350\tr\327\203T\263\034\230Y6\206x\022\245\002\343\266\214\306\203\265\311|S\243\311y\257\310\270\032\263\262\246~\002\032\245\234A\352\b\207\250\345\321\016\355i\\\3123c$\310K\341W\325\303\2579\n\336\366\237c;Dy[)\367\232\260.\213\034\357J\v\331:Uu\3058\267\\P\021\016t\276W\334p0\335\257\0245\227\034\024\021pF\375\371\214\024`\3365\330\334\201\206\3071\037\370je-\267\212\357\362a!\000,\215O:I\036$\200\372V\320-\016R\256)+jJ\307\026\217\265\330\354A\030\0034\362\330\224y\202\310\r\342\020r9\205\271\367\273T\\q!I#\245J\320"
auths {
  method: "KERBEROS"
  mechanism: "GSSAPI"
  protocol: "nn"
  serverId: "hadoop-master"
}

16/08/02 18:34:13 DEBUG security.SaslRpcClient: Received SASL message
state: CHALLENGE
token:
"`l\006\t*\206H\206\367\022\001\002\002\002\000o]0[\240\003\002\001\005\241\003\002\001\017\242O0M\240\003\002\001\020\242F\004D\337\316\251\336\365\261O@\377
\"\035\203\002\357Z\231e\332\357\364\204>d\325\"\340\263\2302\031\277\023G\342=\355\334)\303\271\t\376\252\225\207\033\000\243\332\252\335{\"\033\025
\fW\225\300\375\272\201\367\216\371\273"

Entered Krb5Context.initSecContext with state=STATE_IN_PROCESS
>>> EType: sun.security.krb5.internal.crypto.Des3CbcHmacSha1KdEType
Krb5Context setting peerSeqNumber to: 766454664
16/08/02 18:34:13 DEBUG security.SaslRpcClient: Sending sasl message state:
RESPONSE
token: ""

16/08/02 18:34:13 DEBUG security.SaslRpcClient: Received SASL message
state: CHALLENGE
token:
"`?\006\t*\206H\206\367\022\001\002\002\002\001\004\000\377\377\377\377\272
\237\354\300\003\367{\207A\267\371\245\327\374\333\021\026\375}\353\035\254\327\305\272\373\305\365L\022\374.A\203\002\001\001\000\000\004\004\004\004"

Krb5Context.unwrap: token=[60 3f 06 09 2a 86 48 86 f7 12 01 02 02 02 01 04
00 ff ff ff ff ba 20 9f ec c0 03 f7 7b 87 41 b7 f9 a5 d7 fc db 11 16 fd 7d
eb 1d ac d7 c5 ba fb c5 f5 4c 12 fc 2e 41 83 02 01 01 00 00 04 04 04 04 ]
Krb5Context.unwrap: data=[01 01 00 00 ]
Krb5Context.wrap: data=[01 01 00 00 ]
Krb5Context.wrap: token=[60 3f 06 09 2a 86 48 86 f7 12 01 02 02 02 01 04 00
ff ff ff ff 33 b9 e5 96 b6 c8 d3 80 4f 8a a1 5b 44 c9 b6 76 ea fe ec 80 be
37 12 e1 04 cc e5 0f 2a f8 16 1b 9e 72 17 dc 01 01 00 00 04 04 04 04 ]
16/08/02 18:34:13 DEBUG security.SaslRpcClient: Sending sasl message state:
RESPONSE
token:
"`?\006\t*\206H\206\367\022\001\002\002\002\001\004\000\377\377\377\3773\271\345\226\266\310\323\200O\212\241[D\311\266v\352\376\354\200\2767\022\341\004\314\345\017*\370\026\033\236r\027\334\001\001\000\000\004\004\004\004"

16/08/02 18:34:13 DEBUG security.SaslRpcClient: Received SASL message
state: SUCCESS

16/08/02 18:34:13 DEBUG ipc.Client: Negotiated QOP is :auth
16/08/02 18:34:13 DEBUG ipc.Client: IPC Client (1594470328) connection to /
192.168.23.206:8020 from nn/hadoop-master@platalyticsrealm: starting,
having connections 1
16/08/02 18:34:13 DEBUG ipc.Client: IPC Client (1594470328) connection to /
192.168.23.206:8020 from nn/hadoop-master@platalyticsrealm sending #0
16/08/02 18:34:13 DEBUG ipc.Client: IPC Client (1594470328) connection to /
192.168.23.206:8020 from nn/hadoop-master@platalyticsrealm got value #0
16/08/02 18:34:13 DEBUG ipc.ProtobufRpcEngine: Call: getFileInfo took 594ms
16/08/02 18:34:14 TRACE ipc.ProtobufRpcEngine: 1: Response <- /
192.168.23.206:8020: getFileInfo {fs { fileType: IS_DIR path: "" length: 0
permission { perm: 493 } owner: "hdfs" group: "supergroup"
modification_time: 1470131070337 access_time: 0 block_replication: 0
blocksize: 0 fileId: 16385 childrenNum: 1 storagePolicy: 0 }}
16/08/02 18:34:14 TRACE ipc.ProtobufRpcEngine: 1: Call -> /
192.168.23.206:8020: getListing {src: "/" startAfter: "" needLocation:
false}
16/08/02 18:34:14 DEBUG ipc.Client: IPC Client (1594470328) connection to /
192.168.23.206:8020 from nn/hadoop-master@platalyticsrealm sending #1
16/08/02 18:34:14 DEBUG ipc.Client: IPC Client (1594470328) connection to /
192.168.23.206:8020 from nn/hadoop-master@platalyticsrealm got value #1
16/08/02 18:34:14 DEBUG ipc.ProtobufRpcEngine: Call: getListing took 7ms
16/08/02 18:34:14 TRACE ipc.ProtobufRpcEngine: 1: Response <- /
192.168.23.206:8020: getListing {dirList { partialListing { fileType:
IS_DIR path: "ranger" length: 0 permission { perm: 493 } owner: "hdfs"
group: "supergroup" modification_time: 1470131070364 access_time: 0
block_replication: 0 blocksize: 0 fileId: 16386 childrenNum: 1
storagePolicy: 0 } remainingEntries: 0 }}
*Found 1 items*
*drwxr-xr-x   - hdfs supergroup          0 2016-08-02 14:44 /ranger*
16/08/02 18:34:14 DEBUG ipc.Client: stopping client from cache:
org.apache.hadoop.ipc.Client@5e0df7af
16/08/02 18:34:14 DEBUG ipc.Client: removing client from cache:
org.apache.hadoop.ipc.Client@5e0df7af
16/08/02 18:34:14 DEBUG ipc.Client: stopping actual client because no more
references remain: org.apache.hadoop.ipc.Client@5e0df7af
16/08/02 18:34:14 DEBUG ipc.Client: Stopping client
16/08/02 18:34:14 DEBUG ipc.Client: IPC Client (1594470328) connection to /
192.168.23.206:8020 from nn/hadoop-master@platalyticsrealm: closed
16/08/02 18:34:14 DEBUG ipc.Client: IPC Client (1594470328) connection to /
192.168.23.206:8020 from nn/hadoop-master@platalyticsrealm: stopped,
remaining connections 0


On Tue, Aug 2, 2016 at 9:03 PM, Dima Spivak <dspi...@cloudera.com> wrote:

> Hm, not sure what to say. The error seems to be pointing at not having a
> TGT...
>
> -Dima
>
> On Tue, Aug 2, 2016 at 12:45 AM, Aneela Saleem <ane...@platalytics.com>
> wrote:
>
> > Yes, I have kinit'd as the service user. But still getting error
> >
> > On Tue, Aug 2, 2016 at 3:05 AM, Dima Spivak <dspi...@cloudera.com>
> wrote:
> >
> > > The stacktrace suggests you don't have a ticket-granting ticket. Have
> you
> > > kinit'd as the service user?
> > >
> > > -Dima
> > >
> > > On Sun, Jul 31, 2016 at 11:19 PM, Aneela Saleem <
> ane...@platalytics.com>
> > > wrote:
> > >
> > > > Hi Dima,
> > > >
> > > > I followed the official reference guide now, but still same error.
> > > > Attached is the hbase-site.xml file, please have a look. What's wrong
> > > there?
> > > >
> > > > On Thu, Jul 28, 2016 at 11:58 PM, Dima Spivak <dspi...@cloudera.com>
> > > > wrote:
> > > >
> > > >> I haven't looked in detail at your hbase-site.xml, but if you're
> > running
> > > >> Apache HBase (and not a CDH release), I might recommend using the
> > > official
> > > >> reference guide [1] to configure your cluster instead of the CDH
> 4.2.0
> > > >> docs
> > > >> since those would correspond to HBase 0.94, and might well have
> > > different
> > > >> steps required to set up security. If you are trying out CDH HBase,
> be
> > > >> sure
> > > >> to use up-to-date documentation for your release.
> > > >>
> > > >> Let us know how it goes.
> > > >>
> > > >> [1] https://hbase.apache.org/book.html#hbase.secure.configuration
> > > >>
> > > >> -Dima
> > > >>
> > > >> On Thu, Jul 28, 2016 at 10:09 AM, Aneela Saleem <
> > ane...@platalytics.com
> > > >
> > > >> wrote:
> > > >>
> > > >> > Hi Dima,
> > > >> >
> > > >> > I'm running Hbase version 1.2.2
> > > >> >
> > > >> > On Thu, Jul 28, 2016 at 8:35 PM, Dima Spivak <
> dspi...@cloudera.com>
> > > >> wrote:
> > > >> >
> > > >> > > Hi Aneela,
> > > >> > >
> > > >> > > What version of HBase are you running?
> > > >> > >
> > > >> > > -Dima
> > > >> > >
> > > >> > > On Thursday, July 28, 2016, Aneela Saleem <
> ane...@platalytics.com
> > >
> > > >> > wrote:
> > > >> > >
> > > >> > > > Hi,
> > > >> > > >
> > > >> > > > I have successfully configured Zookeeper with Kerberos
> > > >> authentication.
> > > >> > > Now
> > > >> > > > i'm facing issue while configuring HBase with Kerberos
> > > >> authentication.
> > > >> > I
> > > >> > > > have followed this link
> > > >> > > > <
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> http://www.cloudera.com/documentation/archive/cdh/4-x/4-2-0/CDH4-Security-Guide/cdh4sg_topic_8_2.html
> > > >> > > >.
> > > >> > > > Attached are the configuration files, i.e., hbase-site.xml and
> > > >> > > > zk-jaas.conf.
> > > >> > > >
> > > >> > > > Following are the logs from regionserver:
> > > >> > > >
> > > >> > > > 016-07-28 17:44:56,881 WARN  [regionserver/hadoop-master/
> > > >> > > > 192.168.23.206:16020] regionserver.HRegionServer: error
> telling
> > > >> master
> > > >> > > we
> > > >> > > > are up
> > > >> > > > com.google.protobuf.ServiceException: java.io.IOException:
> Could
> > > not
> > > >> > set
> > > >> > > > up IO Streams to hadoop-master/192.168.23.206:16000
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:240)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:336)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$BlockingStub.regionServerStartup(RegionServerStatusProtos.java:8982)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.regionserver.HRegionServer.reportForDuty(HRegionServer.java:2284)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:906)
> > > >> > > > at java.lang.Thread.run(Thread.java:745)
> > > >> > > > Caused by: java.io.IOException: Could not set up IO Streams to
> > > >> > > > hadoop-master/192.168.23.206:16000
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:785)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:906)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:873)
> > > >> > > > at
> > > >> > >
> > > >>
> > org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1241)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:227)
> > > >> > > > ... 5 more
> > > >> > > > Caused by: java.lang.RuntimeException: SASL authentication
> > failed.
> > > >> The
> > > >> > > > most likely cause is missing or invalid credentials. Consider
> > > >> 'kinit'.
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection$1.run(RpcClientImpl.java:685)
> > > >> > > > at java.security.AccessController.doPrivileged(Native Method)
> > > >> > > > at javax.security.auth.Subject.doAs(Subject.java:415)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.handleSaslConnectionFailure(RpcClientImpl.java:643)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:751)
> > > >> > > > ... 9 more
> > > >> > > > Caused by: javax.security.sasl.SaslException: GSS initiate
> > failed
> > > >> > [Caused
> > > >> > > > by GSSException: No valid credentials provided (Mechanism
> level:
> > > >> Failed
> > > >> > > to
> > > >> > > > find any Kerberos tgt)]
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> com.sun.security.sasl.gsskerb.GssKrb5Client.evaluateChallenge(GssKrb5Client.java:212)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.security.HBaseSaslRpcClient.saslConnect(HBaseSaslRpcClient.java:179)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupSaslConnection(RpcClientImpl.java:617)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.access$700(RpcClientImpl.java:162)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection$2.run(RpcClientImpl.java:743)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection$2.run(RpcClientImpl.java:740)
> > > >> > > > at java.security.AccessController.doPrivileged(Native Method)
> > > >> > > > at javax.security.auth.Subject.doAs(Subject.java:415)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:740)
> > > >> > > > ... 9 more
> > > >> > > > Caused by: GSSException: No valid credentials provided
> > (Mechanism
> > > >> > level:
> > > >> > > > Failed to find any Kerberos tgt)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> sun.security.jgss.krb5.Krb5InitCredential.getInstance(Krb5InitCredential.java:147)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> sun.security.jgss.krb5.Krb5MechFactory.getCredentialElement(Krb5MechFactory.java:121)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> sun.security.jgss.krb5.Krb5MechFactory.getMechanismContext(Krb5MechFactory.java:187)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> sun.security.jgss.GSSManagerImpl.getMechanismContext(GSSManagerImpl.java:223)
> > > >> > > > at
> > > >> > >
> > > >>
> > sun.security.jgss.GSSContextImpl.initSecContext(GSSContextImpl.java:212)
> > > >> > > > at
> > > >> > >
> > > >>
> > sun.security.jgss.GSSContextImpl.initSecContext(GSSContextImpl.java:179)
> > > >> > > > at
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > >
> >
> com.sun.security.sasl.gsskerb.GssKrb5Client.evaluateChallenge(GssKrb5Client.java:193)
> > > >> > > >
> > > >> > > >
> > > >> > > > Please have a look, whats going wrong here?
> > > >> > > >
> > > >> > > > Thanks
> > > >> > > >
> > > >> > > >
> > > >> > >
> > > >> >
> > > >>
> > > >
> > > >
> > >
> > >
> > > --
> > > -Dima
> > >
> >
>
>
>
> --
> -Dima
>

Reply via email to