Disclaimer: CDH questions are better handled at [email protected]. But the question I'd like to ask is: why do you need your own Spark build? What's wrong with CDH's Spark that it doesn't work for you?
On Thu, Jan 8, 2015 at 3:01 PM, freedafeng <[email protected]> wrote: > Could anyone come up with your experience on how to do this? > > I have created a cluster and installed cdh5.3.0 on it with basically core + > Hbase. but cloudera installed and configured the spark in its parcels > anyway. I'd like to install our custom spark on this cluster to use the > hadoop and hbase service there. There could be potentially conflicts if this > is not done correctly. Library conflicts are what I worry most. > > I understand this is a special case. but if you know how to do it, please > let me know. Thanks. > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/correct-best-way-to-install-custom-spark1-2-on-cdh5-3-0-tp21045.html > Sent from the Apache Spark User List mailing list archive at Nabble.com. > > --------------------------------------------------------------------- > To unsubscribe, e-mail: [email protected] > For additional commands, e-mail: [email protected] > -- Marcelo --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
