Right, create three spark interpreters for your 3 yarn cluster.
Serega Sheypak <serega.shey...@gmail.com>于2017年6月30日周五 下午10:33写道: > Hi, thanks for your reply! > What do you mean by that? > I can have only one env variable HADOOP_CONF_DIR... > And how can user pick which env to run? > > Or you mean I have to create three Spark interpreters and each of them > would have it's own HADOOP_CONF_DIR pointed to single cluster config? > > 2017-06-30 16:21 GMT+02:00 Jeff Zhang <zjf...@gmail.com>: > >> >> Try set HADOOP_CONF_DIR for each yarn conf in interpreter setting. >> >> Serega Sheypak <serega.shey...@gmail.com>于2017年6月30日周五 下午10:11写道: >> >>> Hi I have several different hadoop clusters, each of them has it's own >>> YARN. >>> Is it possible to configure single Zeppelin instance to work with >>> different clusters? >>> I want to run spark on cluster A if data is there. Right now my Zeppelin >>> runs on single cluster and it sucks data from remote clusters which is >>> inefficient. Zeppelin can easily access any HDFS cluster, but what about >>> YARN? >>> >>> What are the correct approaches to solve the problem? >>> >> >