Can you try running the example like this ./bin/run-example sql.RDDRelation <opts>
I know there are some jars in the example folders, and running them this way adds them to the classpath On Jul 7, 2016 3:47 AM, "kevin" <kiss.kevin...@gmail.com> wrote: > hi,all: > I build spark use: > > ./make-distribution.sh --name "hadoop2.7.1" --tgz > "-Pyarn,hadoop-2.6,parquet-provided,hive,hive-thriftserver" -DskipTests > -Dhadoop.version=2.7.1 > > I can run example : > ./bin/spark-submit --class org.apache.spark.examples.SparkPi \ > --master spark://master1:7077 \ > --driver-memory 1g \ > --executor-memory 512m \ > --executor-cores 1 \ > lib/spark-examples*.jar \ > 10 > > but can't run example : > org.apache.spark.examples.sql.RDDRelation > > *I got error:* > 16/07/07 18:28:45 INFO client.AppClient$ClientEndpoint: Executor updated: > app-20160707182845-0003/2 is now RUNNING > 16/07/07 18:28:45 INFO client.AppClient$ClientEndpoint: Executor updated: > app-20160707182845-0003/4 is now RUNNING > 16/07/07 18:28:45 INFO client.AppClient$ClientEndpoint: Executor updated: > app-20160707182845-0003/3 is now RUNNING > 16/07/07 18:28:45 INFO client.AppClient$ClientEndpoint: Executor updated: > app-20160707182845-0003/0 is now RUNNING > 16/07/07 18:28:45 INFO client.AppClient$ClientEndpoint: Executor updated: > app-20160707182845-0003/1 is now RUNNING > 16/07/07 18:28:45 INFO client.AppClient$ClientEndpoint: Executor updated: > app-20160707182845-0003/5 is now RUNNING > 16/07/07 18:28:46 INFO cluster.SparkDeploySchedulerBackend: > SchedulerBackend is ready for scheduling beginning after reached > minRegisteredResourcesRatio: 0.0 > Exception in thread "main" java.lang.NoClassDefFoundError: > org/apache/parquet/hadoop/ParquetOutputCommitter > at org.apache.spark.sql.SQLConf$.<init>(SQLConf.scala:319) > at org.apache.spark.sql.SQLConf$.<clinit>(SQLConf.scala) > at org.apache.spark.sql.SQLContext.<init>(SQLContext.scala:85) > at org.apache.spark.sql.SQLContext.<init>(SQLContext.scala:77) > at main.RDDRelation$.main(RDDRelation.scala:13) > at main.RDDRelation.main(RDDRelation.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) > at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > Caused by: java.lang.ClassNotFoundException: > org.apache.parquet.hadoop.ParquetOutputCommitter > at java.net.URLClassLoader.findClass(URLClassLoader.java:381) > at java.lang.ClassLoader.loadClass(ClassLoader.java:424) > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331) > at java.lang.ClassLoader.loadClass(ClassLoader.java:357) > ... 15 more > >