This is not a bug, because these 2 processes use the same SPARK_PID_DIR which is /tmp by default. Although you can resolve this by using different SPARK_PID_DIR, I suspect you would still have other issues like port conflict. I would suggest you to deploy one spark thrift server per machine for now. If stick to deploy multiple spark thrift server on one machine, then define different SPARK_CONF_DIR, SPARK_LOG_DIR and SPARK_PID_DIR for your 2 instances of spark thrift server. Not sure if there's other conflicts. but please try first.
On Fri, Jul 1, 2016 at 10:47 AM, Egor Pahomov <pahomov.e...@gmail.com> wrote: > I get > > "org.apache.spark.sql.hive.thriftserver.HiveThriftServer2 running as > process 28989. Stop it first." > > Is it a bug? > > 2016-07-01 10:10 GMT-07:00 Jeff Zhang <zjf...@gmail.com>: > >> I don't think the one instance per machine is true. As long as you >> resolve the conflict issue such as port conflict, pid file, log file and >> etc, you can run multiple instances of spark thrift server. >> >> On Fri, Jul 1, 2016 at 9:32 AM, Egor Pahomov <pahomov.e...@gmail.com> >> wrote: >> >>> Hi, I'm using Spark Thrift JDBC server and 2 limitations are really >>> bother me - >>> >>> 1) One instance per machine >>> 2) Yarn client only(not yarn cluster) >>> >>> Are there any architectural reasons for such limitations? About >>> yarn-client I might understand in theory - master is the same process as a >>> server, so it makes some sense, but it's really inconvenient - I need a lot >>> of memory on my driver machine. Reasons for one instance per machine I do >>> not understand. >>> >>> -- >>> >>> >>> *Sincerely yoursEgor Pakhomov* >>> >> >> >> >> -- >> Best Regards >> >> Jeff Zhang >> > > > > -- > > > *Sincerely yoursEgor Pakhomov* > -- Best Regards Jeff Zhang