Thanks.
In addition??I want to know, if I can use spark-1.6.1-bin-hadoop2.6.tgz which is a pre-built package on hadoop 2.7.2?? ------------------ ???????? ------------------ ??????: "Ted Yu";<yuzhih...@gmail.com>; ????????: 2016??6??22??(??????) ????11:51 ??????: "??????"<251922...@qq.com>; ????: "user"<user@spark.apache.org>; ????: Re: spark-1.6.1-bin-without-hadoop can not use spark-sql build/mvn clean -Phive -Phive-thriftserver -Pyarn -Phadoop-2.6 -Psparkr -Dhadoop.version=2.7.2 package On Wed, Jun 22, 2016 at 8:00 AM, 251922566 <251922...@qq.com> wrote: ok,i will rebuild myself. if i want to use spark with hadoop 2.7.2, when i build spark, i should put what on param --hadoop, 2.7.2 or others? ???????????????? -------- ???????? -------- ??????Re: spark-1.6.1-bin-without-hadoop can not use spark-sql ????????Ted Yu ?????????????? <251922...@qq.com> ??????user I wonder if the tar ball was built with: -Phive -Phive-thriftserver Maybe rebuild by yourself with the above ? FYI On Wed, Jun 22, 2016 at 4:38 AM, ?????? <251922...@qq.com> wrote: Hi all. I download spark-1.6.1-bin-without-hadoop.tgz from website. And I configured "SPARK_DIST_CLASSPATH" in spark-env.sh. Now spark-shell run well. But spark-sql can not run. My hadoop version is 2.7.2. This is error infos: bin/spark-sql java.lang.ClassNotFoundException: org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver at java.net.URLClassLoader$1.run(URLClassLoader.java:366) at java.net.URLClassLoader$1.run(URLClassLoader.java:355) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:354) at java.lang.ClassLoader.loadClass(ClassLoader.java:425) at java.lang.ClassLoader.loadClass(ClassLoader.java:358) at java.lang.Class.forName0(Native Method) at java.lang.Class.forName(Class.java:278) at org.apache.spark.util.Utils$.classForName(Utils.scala:174) at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:689) at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) Failed to load main class org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver. You need to build Spark with -Phive and -Phive-thriftserver. Do I need configure something else in spark-env.sh or spark-default.conf? Suggestions are appreciated ,thanks.