[ 
https://issues.apache.org/jira/browse/FLINK-27191?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17521188#comment-17521188
 ] 

Yuan Zhu commented on FLINK-27191:
----------------------------------

But when job starts running, Hive source need to connect hdfs/hiveMetaStore 
with specific principal. IMHO, TM will install HadoopModule when it starts. If 
we try to connect hdfs/hiveMetaStore with other principal, the configuration 
will conflict with TM's. How to avoid it?

> Support multi kerberos-enabled Hive clusters 
> ---------------------------------------------
>
>                 Key: FLINK-27191
>                 URL: https://issues.apache.org/jira/browse/FLINK-27191
>             Project: Flink
>          Issue Type: Improvement
>          Components: Connectors / Hive
>            Reporter: luoyuxia
>            Priority: Major
>             Fix For: 1.16.0
>
>
> Currently, to access kerberos-enabled Hive cluster, users are expected to add 
> key/secret in flink-conf. But it can only access one Hive cluster in one 
> Flink cluster, we are also expected to support multi kerberos-enabled Hive 
> clusters  in one Flink cluster.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

Reply via email to