Hi, Can you please check if you have spark-catalyst jar in $SPARK_HOME/jars folder for your cloudera version, if its not there please try to include it and retry.
Thanks and regards Bhavya On Sun, Nov 5, 2017 at 7:24 PM, Lionel CL <whuca...@outlook.com> wrote: > I have the same problem in CDH 5.8.0 > spark2 version is 2.1.0.cloudera1 > carbondata version 1.2.0. > > There's no error occurred when using open source version spark. > > <hadoop.version>2.6.0-cdh5.8.0</hadoop.version> > <spark.version>2.1.0.cloudera1</spark.version> > <scala.binary.version>2.11</scala.binary.version> > <scala.version>2.11.8</scala.version> > > > scala> cc.sql("create table t111(vin string) stored by 'carbondata'") > 17/11/03 10:22:03 AUDIT command.CreateTable: [][][Thread-1]Creating Table > with Database name [default] and Table name [t111] > java.lang.NoSuchMethodError: org.apache.spark.sql.catalyst. > catalog.CatalogTable.copy(Lorg/apache/spark/sql/catalyst/ > TableIdentifier;Lorg/apache/spark/sql/catalyst/catalog/ > CatalogTableType;Lorg/apache/spark/sql/catalyst/catalog/ > CatalogStorageFormat;Lorg/apache/spark/sql/types/StructT > ype;Lscala/Option;Lscala/collection/Seq;Lscala/Option; > Ljava/lang/String;JJLscala/collection/immutable/Map; > Lscala/Option;Lscala/Option;Lscala/Option;Lscala/Option; > Lscala/collection/Seq;Z)Lorg/apache/spark/sql/catalyst/ > catalog/CatalogTable; > at org.apache.spark.sql.CarbonSource$.updateCatalogTableWithCar > bonSchema(CarbonSource.scala:253) > at org.apache.spark.sql.execution.command.DDLStrategy.apply( > DDLStrategy.scala:135) > at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun > $1.apply(QueryPlanner.scala:62) > at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun > $1.apply(QueryPlanner.scala:62) > at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434) > at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440) > at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:439) > > > 在 2017/11/1 上午1:58,“chenliang613”<chenliang6...@gmail.com<mailto:chenlia > ng6...@gmail.com>> 写入: > > Hi > > Did you use open source spark version? > > Can you provide more detail info : > 1. which carbondata version and spark version, you used ? > 2. Can you share with us , reproduce script and steps. > > Regards > Liang > > > hujianjun wrote > scala> carbon.sql("CREATE TABLE IF NOT EXISTS carbon_table(id string,name > string,city string,age Int)STORED BY 'carbondata'") > 17/10/23 19:13:52 AUDIT command.CarbonCreateTableCommand: > [master][root][Thread-1]Creating Table with Database name [clb_carbon] and > Table name [carbon_table] > java.lang.NoSuchMethodError: > org.apache.spark.sql.catalyst.catalog.CatalogTable.copy(Lorg > /apache/spark/sql/catalyst/TableIdentifier;Lorg/apache/ > spark/sql/catalyst/catalog/CatalogTableType;Lorg/apache/ > spark/sql/catalyst/catalog/CatalogStorageFormat;Lorg/ > apache/spark/sql/types/StructType;Lscala/Option;Lscala/ > collection/Seq;Lscala/Option;Ljava/lang/String;JJLscala/ > collection/immutable/Map;Lscala/Option;Lscala/Option; > Lscala/Option;Lscala/Option;Lscala/collection/Seq;Z)Lorg/ > apache/spark/sql/catalyst/catalog/CatalogTable; > at > org.apache.spark.sql.CarbonSource$.updateCatalogTableWithCar > bonSchema(CarbonSource.scala:253) > at > org.apache.spark.sql.execution.strategy.DDLStrategy.apply( > DDLStrategy.scala:154) > at > org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun > $1.apply(QueryPlanner.scala:62) > at > org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun > $1.apply(QueryPlanner.scala:62) > at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434) > at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440) > at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:439) > at > org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(Que > ryPlanner.scala:92) > at > org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun > $2$$anonfun$apply$2.apply(QueryPlanner.scala:77) > at > org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun > $2$$anonfun$apply$2.apply(QueryPlanner.scala:74) > at > scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(T > raversableOnce.scala:157) > at > scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(T > raversableOnce.scala:157) > at scala.collection.Iterator$class.foreach(Iterator.scala:893) > at scala.collection.AbstractIterator.foreach(Iterator.scala:1336) > at > scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.scala:157) > at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1336) > at > org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun > $2.apply(QueryPlanner.scala:74) > at > org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun > $2.apply(QueryPlanner.scala:66) > at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434) > at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440) > at > org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(Que > ryPlanner.scala:92) > at > org.apache.spark.sql.execution.QueryExecution.sparkPlan$ > lzycompute(QueryExecution.scala:79) > at > org.apache.spark.sql.execution.QueryExecution.sparkPlan( > QueryExecution.scala:75) > at > org.apache.spark.sql.execution.QueryExecution.executedPlan$ > lzycompute(QueryExecution.scala:84) > at > org.apache.spark.sql.execution.QueryExecution.executedPlan( > QueryExecution.scala:84) > at > org.apache.spark.sql.execution.QueryExecution.toRdd$ > lzycompute(QueryExecution.scala:87) > at > org.apache.spark.sql.execution.QueryExecution.toRdd( > QueryExecution.scala:87) > at org.apache.spark.sql.Dataset. > <init> > (Dataset.scala:185) > at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64) > at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592) > at > org.apache.spark.sql.execution.command.CarbonCreateTableComm > and.processSchema(CarbonCreateTableCommand.scala:84) > at > org.apache.spark.sql.execution.command.CarbonCreateTableComm > and.run(CarbonCreateTableCommand.scala:36) > at > org.apache.spark.sql.execution.command.ExecutedCommandExec.s > ideEffectResult$lzycompute(commands.scala:58) > at > org.apache.spark.sql.execution.command.ExecutedCommandExec.s > ideEffectResult(commands.scala:56) > at > org.apache.spark.sql.execution.command.ExecutedCommandExec. > doExecute(commands.scala:74) > at > org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1. > apply(SparkPlan.scala:114) > at > org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1. > apply(SparkPlan.scala:114) > at > org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQue > ry$1.apply(SparkPlan.scala:135) > at > org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperati > onScope.scala:151) > at > org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:132) > at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan. > scala:113) > at > org.apache.spark.sql.execution.QueryExecution.toRdd$ > lzycompute(QueryExecution.scala:87) > at > org.apache.spark.sql.execution.QueryExecution.toRdd( > QueryExecution.scala:87) > at org.apache.spark.sql.Dataset. > <init> > (Dataset.scala:185) > at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64) > at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592) > ... 52 elided > -- > Sent from: http://apache-carbondata-user-mailing-list.3231.n8.nabble.com/ > > > > > > -- > Sent from: http://apache-carbondata-user-mailing-list.3231.n8.nabble.com/ > >