AlexNilone opened a new issue, #3076:
URL: https://github.com/apache/incubator-seatunnel/issues/3076

   ### Search before asking
   
   - [X] I had searched in the 
[issues](https://github.com/apache/incubator-seatunnel/issues?q=is%3Aissue+label%3A%22bug%22)
 and found no similar issues.
   
   
   ### What happened
   
   
因为一些特殊的原因,需要以jdbc方式来访问hiveserver2服务,集群环境是CDH5.13的。因为涉及到hive版本的问题,用的Spark_home是CDH
 通过parcels安装的Spark2.4.0的目录。驱动包另在plugins目录下建的子目录放的是集群的hive-jdbc包
   
   两个问题:
   1、查询不到数据
   2、若以yarn client方式提交作业,报告gss票据问题(集群添加了kerberos认证)
   
   ### SeaTunnel Version
   
   2.1.3
   
   ### SeaTunnel Config
   
   ```conf
   env {
     spark.sql.catalogImplementation = "hive"
        spark.app.name = "SeaTunnel"
        spark.executor.instances = 1
        spark.executor.cores = 1
        spark.num.executors=1
        spark.executor.memory = "1g"
        execution.parallelism = 1
        spark.yarn.keytab=/hdfs.keytab
        spark.yarn.principal="hdfs/server001@MYCDH"
   }
   
   source {
        jdbc {
                        driver = org.apache.hive.jdbc.HiveDriver ,
            url = 
"jdbc:hive2://server001:10000/;principal=hive/server001@MYCDH",
            user = "hive",
            password = "hive",
            table = "test_seatunnel_source"
            result_table_name = "test_seatunnel_source"
   
        }
   }
   
   transform {
   }
   
   sink{
        Console {}
   
   }
   ```
   
   
   ### Running Command
   
   ```shell
   start-seatunnel-spark.sh --master local --deploy-mode client \
        --config 
/data/apache-seatunnel-incubating-2.1.3/config/hivejdbc-console.conf
   
   
   start-seatunnel-spark.sh --master yarn --deploy-mode client \
        --config 
/data/apache-seatunnel-incubating-2.1.3/config/hivejdbc-console.conf
   ```
   
   
   ### Error Exception
   
   ```log
   能打印出来表头,但是没有具体的数据内容查询到。
   
   22/10/12 17:20:33 INFO jdbc.Utils: Resolved authority: cdh129135:10000
   22/10/12 17:20:33 INFO jdbc.JDBCRDD: closed connection
   22/10/12 17:20:33 INFO executor.Executor: Finished task 0.0 in stage 0.0 
(TID 0). 1069 bytes result sent to driver
   22/10/12 17:20:33 INFO scheduler.TaskSetManager: Finished task 0.0 in stage 
0.0 (TID 0) in 876 ms on localhost (executor driver) (1/1)
   22/10/12 17:20:33 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 0.0, 
whose tasks have all completed, from pool 
   22/10/12 17:20:33 INFO scheduler.DAGScheduler: ResultStage 0 (show at 
Console.scala:38) finished in 1.295 s
   22/10/12 17:20:33 INFO scheduler.DAGScheduler: Job 0 finished: show at 
Console.scala:38, took 1.344624 s
   +------------------------+--------------------------+
   |test_seatunnel_source.id|test_seatunnel_source.name|
   +------------------------+--------------------------+
   +------------------------+--------------------------+
   
   22/10/12 17:20:33 INFO spark.SparkContext: Invoking stop() from shutdown hook
   22/10/12 17:20:33 INFO server.AbstractConnector: Stopped 
Spark@1f52eb6f{HTTP/1.1,[http/1.1]}{0.0.0.0:4040}
   22/10/12 17:20:33 INFO ui.SparkUI: Stopped Spark web UI at 
http://cdh129135:4040
   22/10/12 17:20:33 INFO spark.MapOutputTrackerMasterEndpoint: 
MapOutputTrackerMasterEndpoint stopped!
   22/10/12 17:20:33 INFO memory.MemoryStore: MemoryStore cleared
   22/10/12 17:20:33 INFO storage.BlockManager: BlockManager stopped
   22/10/12 17:20:33 INFO storage.BlockManagerMaster: BlockManagerMaster stopped
   22/10/12 17:20:33 INFO 
scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: 
OutputCommitCoordinator stopped!
   22/10/12 17:20:33 INFO spark.SparkContext: Successfully stopped SparkContext
   22/10/12 17:20:33 INFO util.ShutdownHookManager: Shutdown hook called
   22/10/12 17:20:33 INFO util.ShutdownHookManager: Deleting directory 
/tmp/spark-c931a4e1-6abc-476b-813a-718773b5e110
   22/10/12 17:20:33 INFO util.ShutdownHookManager: Deleting directory 
/tmp/spark-7add2b21-6a02-4943-9cf1-5349f3fc3c37
   
   
   
   ---yarn方式运行报gss报错
   Caused by: org.apache.thrift.transport.TTransportException: GSS initiate 
failed
           at 
org.apache.thrift.transport.TSaslTransport.sendAndThrowMessage(TSaslTransport.java:232)
           at 
org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:316)
           at 
org.apache.thrift.transport.TSaslClientTransport.open(TSaslClientTransport.java:37)
           at 
org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$1.run(TUGIAssumingTransport.java:52)
           at 
org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$1.run(TUGIAssumingTransport.java:49)
           at java.security.AccessController.doPrivileged(Native Method)
           at javax.security.auth.Subject.doAs(Subject.java:422)
           at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1920)
           at 
org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport.open(TUGIAssumingTransport.java:49)
           at 
org.apache.hive.jdbc.HiveConnection.openTransport(HiveConnection.java:204)
   ```
   
   
   ### Flink or Spark Version
   
   Spark2.4.0( CDH5.13官方parcels安装包)
   
   ### Java or Scala Version
   
   1.8
   
   ### Screenshots
   
   1
   
   ### Are you willing to submit PR?
   
   - [X] Yes I am willing to submit a PR!
   
   ### Code of Conduct
   
   - [X] I agree to follow this project's [Code of 
Conduct](https://www.apache.org/foundation/policies/conduct)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to