Did you check that the security extensions are installed (JCE)? KhajaAsmath Mohammed <mdkhajaasm...@gmail.com> schrieb am Mi., 22. Nov. 2017 um 19:36 Uhr:
> [image: Inline image 1] > > This is what we are on. > > On Wed, Nov 22, 2017 at 12:33 PM, KhajaAsmath Mohammed < > mdkhajaasm...@gmail.com> wrote: > >> We use oracle JDK. we are on unix. >> >> On Wed, Nov 22, 2017 at 12:31 PM, Georg Heiler <georg.kf.hei...@gmail.com >> > wrote: >> >>> Do you use oracle or open jdk? We recently had an issue with open jdk: >>> formerly, java Security extensions were installed by default - no longer so >>> on centos 7.3 >>> >>> Are these installed? >>> >>> KhajaAsmath Mohammed <mdkhajaasm...@gmail.com> schrieb am Mi. 22. Nov. >>> 2017 um 19:29: >>> >>>> I passed keytab, renewal is enabled by running the script every eight >>>> hours. User gets renewed by the script every eight hours. >>>> >>>> On Wed, Nov 22, 2017 at 12:27 PM, Georg Heiler < >>>> georg.kf.hei...@gmail.com> wrote: >>>> >>>>> Did you pass a keytab? Is renewal enabled in your kdc? >>>>> KhajaAsmath Mohammed <mdkhajaasm...@gmail.com> schrieb am Mi. 22. >>>>> Nov. 2017 um 19:25: >>>>> >>>>>> Hi, >>>>>> >>>>>> I have written spark stream job and job is running successfully for >>>>>> more than 36 hours. After around 36 hours job gets failed with kerberos >>>>>> issue. Any solution on how to resolve it. >>>>>> >>>>>> org.apache.spark.SparkException: Task failed while wri\ >>>>>> >>>>>> ting rows. >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.hive.SparkHiveDynamicPartitionWriterContainer.writeToFile(hiveWriterContainers.scala:328) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.hive.execution.InsertIntoHiveTable$$anonfun$saveAsHiveFile$3.apply(InsertIntoHiveTable.scala:210) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.hive.execution.InsertIntoHiveTable$$anonfun$saveAsHiveFile$3.apply(InsertIntoHiveTable.scala:210) >>>>>> >>>>>> at >>>>>> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87) >>>>>> >>>>>> at org.apache.spark.scheduler.Task.run(Task.scala:99) >>>>>> >>>>>> at >>>>>> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:322) >>>>>> >>>>>> at >>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>>>>> >>>>>> at >>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>>>>> >>>>>> at java.lang.Thread.run(Thread.java:745) >>>>>> >>>>>> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: >>>>>> java.io.IOException: org.apache.hadoop.security.authentication.client.\ >>>>>> >>>>>> AuthenticationException: >>>>>> org.apache.hadoop.security.token.SecretManager$InvalidToken: token >>>>>> (kms-dt >>>>>> owner=va_dflt, renewer=yarn, re\ >>>>>> >>>>>> alUser=, issueDate=1511262017635, maxDate=1511866817635, >>>>>> sequenceNumber=1854601, masterKeyId=3392) is expired >>>>>> >>>>>> at org.apache.hadoop.hive.ql.io >>>>>> .HiveFileFormatUtils.getHiveRecordWriter(HiveFileFormatUtils.java:248) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.hive.SparkHiveDynamicPartitionWriterContainer.newOutputWriter$1(hiveWriterContainers.scala:346) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.hive.SparkHiveDynamicPartitionWriterContainer.writeToFile(hiveWriterContainers.scala:304) >>>>>> >>>>>> ... 8 more >>>>>> >>>>>> Caused by: java.io.IOException: >>>>>> org.apache.hadoop.security.authentication.client.AuthenticationException: >>>>>> org.apache.hadoop.securit\ >>>>>> >>>>>> y.token.SecretManager$InvalidToken: token (kms-dt owner=va_dflt, >>>>>> renewer=yarn, realUser=, issueDate=1511262017635, maxDate=15118668\ >>>>>> >>>>>> 17635, sequenceNumber=1854601, masterKeyId=3392) is expired >>>>>> >>>>>> at >>>>>> org.apache.hadoop.crypto.key.kms.LoadBalancingKMSClientProvider.decryptEncryptedKey(LoadBalancingKMSClientProvider.java:216) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.crypto.key.KeyProviderCryptoExtension.decryptEncryptedKey(KeyProviderCryptoExtension.java:388) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(DFSClient.java:1440) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.hdfs.DFSClient.createWrappedOutputStream(DFSClient.java:1542) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.hdfs.DFSClient.createWrappedOutputStream(DFSClient.java:1527) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.hdfs.DistributedFileSystem$7.doCall(DistributedFileSystem.java:428) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.hdfs.DistributedFileSystem$7.doCall(DistributedFileSystem.java:421) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:421) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:362) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.fs.FileSystem.create(FileSystem.java:925) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.fs.FileSystem.create(FileSystem.java:906) >>>>>> >>>>>> at >>>>>> parquet.hadoop.ParquetFileWriter.<init>(ParquetFileWriter.java:220) >>>>>> >>>>>> at >>>>>> parquet.hadoop.ParquetOutputFormat.getRecordWriter(ParquetOutputFormat.java:311) >>>>>> >>>>>> at >>>>>> parquet.hadoop.ParquetOutputFormat.getRecordWriter(ParquetOutputFormat.java:287) >>>>>> >>>>>> at org.apache.hadoop.hive.ql.io >>>>>> .parquet.write.ParquetRecordWriterWrapper.<init>(ParquetRecordWriterWrapper.java:65) >>>>>> >>>>>> at org.apache.hadoop.hive.ql.io >>>>>> .parquet.MapredParquetOutputFormat.getParquerRecordWriterWrapper(MapredParquetOutputFormat.java:125) >>>>>> >>>>>> at org.apache.hadoop.hive.ql.io >>>>>> .parquet.MapredParquetOutputFormat.getHiveRecordWriter(MapredParquetOutputFormat.java:114) >>>>>> >>>>>> at org.apache.hadoop.hive.ql.io >>>>>> .HiveFileFormatUtils.getRecordWriter(HiveFileFormatUtils.java:260) >>>>>> >>>>>> at org.apache.hadoop.hive.ql.io >>>>>> .HiveFileFormatUtils.getHiveRecordWriter(HiveFileFormatUtils.java:245) >>>>>> >>>>>> ... 10 more >>>>>> >>>>>> Caused by: >>>>>> org.apache.hadoop.security.authentication.client.AuthenticationException: >>>>>> org.apache.hadoop.security.token.SecretManager\ >>>>>> >>>>>> $InvalidToken: token (kms-dt owner=va_dflt, renewer=yarn, realUser=, >>>>>> issueDate=1511262017635, maxDate=1511866817635, sequenceNumber\ >>>>>> >>>>>> =1854601, masterKeyId=3392) is expired >>>>>> >>>>>> at >>>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) >>>>>> >>>>>> at >>>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) >>>>>> >>>>>> at >>>>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >>>>>> >>>>>> at >>>>>> java.lang.reflect.Constructor.newInstance(Constructor.java:526) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.util.HttpExceptionUtils.validateResponse(HttpExceptionUtils.java:157) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.crypto.key.kms.KMSClientProvider.call(KMSClientProvider.java:627) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.crypto.key.kms.KMSClientProvider.call(KMSClientProvider.java:585) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.crypto.key.kms.KMSClientProvider.decryptEncryptedKey(KMSClientProvider.java:852) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.crypto.key.kms.LoadBalancingKMSClientProvider$5.call(LoadBalancingKMSClientProvider.java:209) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.crypto.key.kms.LoadBalancingKMSClientProvider$5.call(LoadBalancingKMSClientProvider.java:205) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.crypto.key.kms.LoadBalancingKMSClientProvider.doOp(LoadBalancingKMSClientProvider.java:94) >>>>>> >>>>>> at >>>>>> org.apache.hadoop.crypto.key.kms.LoadBalancingKMSClientProvider.decryptEncryptedKey(LoadBalancingKMSClientProvider.java:205) >>>>>> >>>>>> ... 29 more >>>>>> >>>>>> >>>>>> Thanks, >>>>>> >>>>>> Asmath >>>>>> >>>>> >>>> >> >