mvn -DskipTests -Pspark-2.1 clean package The pom file was changed as which provided in former email.
在 2017/11/6 下午7:47,“Bhavya Aggarwal”<bha...@knoldus.com> 写入: >Hi, > >Can you please let me know how are you building the Carbondata assembly >jar, or which command you are running to build carbondata. > >Regards >Bhavya > >On Mon, Nov 6, 2017 at 2:18 PM, Lionel CL <whuca...@outlook.com> wrote: > >> Yes, there is a catalyst jar under the path /opt/cloudera/parcels/SPARK2/ >> lib/spark2/jars/ >> >> spark-catalyst_2.11-2.1.0.cloudera1.jar >> >> >> >> >> >> >> >> 在 2017/11/6 下午4:12,“Bhavya Aggarwal”<bha...@knoldus.com> 写入: >> >> >Hi, >> > >> >Can you please check if you have spark-catalyst jar in $SPARK_HOME/jars >> >folder for your cloudera version, if its not there please try to include >> >it and retry. >> > >> >Thanks and regards >> >Bhavya >> > >> >On Sun, Nov 5, 2017 at 7:24 PM, Lionel CL <whuca...@outlook.com> wrote: >> > >> >> I have the same problem in CDH 5.8.0 >> >> spark2 version is 2.1.0.cloudera1 >> >> carbondata version 1.2.0. >> >> >> >> There's no error occurred when using open source version spark. >> >> >> >> <hadoop.version>2.6.0-cdh5.8.0</hadoop.version> >> >> <spark.version>2.1.0.cloudera1</spark.version> >> >> <scala.binary.version>2.11</scala.binary.version> >> >> <scala.version>2.11.8</scala.version> >> >> >> >> >> >> scala> cc.sql("create table t111(vin string) stored by 'carbondata'") >> >> 17/11/03 10:22:03 AUDIT command.CreateTable: [][][Thread-1]Creating >> Table >> >> with Database name [default] and Table name [t111] >> >> java.lang.NoSuchMethodError: org.apache.spark.sql.catalyst. >> >> catalog.CatalogTable.copy(Lorg/apache/spark/sql/catalyst/ >> >> TableIdentifier;Lorg/apache/spark/sql/catalyst/catalog/ >> >> CatalogTableType;Lorg/apache/spark/sql/catalyst/catalog/ >> >> CatalogStorageFormat;Lorg/apache/spark/sql/types/StructT >> >> ype;Lscala/Option;Lscala/collection/Seq;Lscala/Option; >> >> Ljava/lang/String;JJLscala/collection/immutable/Map; >> >> Lscala/Option;Lscala/Option;Lscala/Option;Lscala/Option; >> >> Lscala/collection/Seq;Z)Lorg/apache/spark/sql/catalyst/ >> >> catalog/CatalogTable; >> >> at org.apache.spark.sql.CarbonSource$.updateCatalogTableWithCar >> >> bonSchema(CarbonSource.scala:253) >> >> at org.apache.spark.sql.execution.command.DDLStrategy.apply( >> >> DDLStrategy.scala:135) >> >> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun >> >> $1.apply(QueryPlanner.scala:62) >> >> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun >> >> $1.apply(QueryPlanner.scala:62) >> >> at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434) >> >> at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440) >> >> at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:439) >> >> >> >> >> >> 在 2017/11/1 上午1:58,“chenliang613”<chenliang6...@gmail.com<mailto: >> chenlia >> >> ng6...@gmail.com>> 写入: >> >> >> >> Hi >> >> >> >> Did you use open source spark version? >> >> >> >> Can you provide more detail info : >> >> 1. which carbondata version and spark version, you used ? >> >> 2. Can you share with us , reproduce script and steps. >> >> >> >> Regards >> >> Liang >> >> >> >> >> >> hujianjun wrote >> >> scala> carbon.sql("CREATE TABLE IF NOT EXISTS carbon_table(id >> string,name >> >> string,city string,age Int)STORED BY 'carbondata'") >> >> 17/10/23 19:13:52 AUDIT command.CarbonCreateTableCommand: >> >> [master][root][Thread-1]Creating Table with Database name [clb_carbon] >> and >> >> Table name [carbon_table] >> >> java.lang.NoSuchMethodError: >> >> org.apache.spark.sql.catalyst.catalog.CatalogTable.copy(Lorg >> >> /apache/spark/sql/catalyst/TableIdentifier;Lorg/apache/ >> >> spark/sql/catalyst/catalog/CatalogTableType;Lorg/apache/ >> >> spark/sql/catalyst/catalog/CatalogStorageFormat;Lorg/ >> >> apache/spark/sql/types/StructType;Lscala/Option;Lscala/ >> >> collection/Seq;Lscala/Option;Ljava/lang/String;JJLscala/ >> >> collection/immutable/Map;Lscala/Option;Lscala/Option; >> >> Lscala/Option;Lscala/Option;Lscala/collection/Seq;Z)Lorg/ >> >> apache/spark/sql/catalyst/catalog/CatalogTable; >> >> at >> >> org.apache.spark.sql.CarbonSource$.updateCatalogTableWithCar >> >> bonSchema(CarbonSource.scala:253) >> >> at >> >> org.apache.spark.sql.execution.strategy.DDLStrategy.apply( >> >> DDLStrategy.scala:154) >> >> at >> >> org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun >> >> $1.apply(QueryPlanner.scala:62) >> >> at >> >> org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun >> >> $1.apply(QueryPlanner.scala:62) >> >> at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434) >> >> at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440) >> >> at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:439) >> >> at >> >> org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(Que >> >> ryPlanner.scala:92) >> >> at >> >> org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun >> >> $2$$anonfun$apply$2.apply(QueryPlanner.scala:77) >> >> at >> >> org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun >> >> $2$$anonfun$apply$2.apply(QueryPlanner.scala:74) >> >> at >> >> scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(T >> >> raversableOnce.scala:157) >> >> at >> >> scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(T >> >> raversableOnce.scala:157) >> >> at scala.collection.Iterator$class.foreach(Iterator.scala:893) >> >> at scala.collection.AbstractIterator.foreach(Iterator.scala:1336) >> >> at >> >> scala.collection.TraversableOnce$class.foldLeft(TraversableOnce. >> scala:157) >> >> at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1336) >> >> at >> >> org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun >> >> $2.apply(QueryPlanner.scala:74) >> >> at >> >> org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun >> >> $2.apply(QueryPlanner.scala:66) >> >> at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434) >> >> at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440) >> >> at >> >> org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(Que >> >> ryPlanner.scala:92) >> >> at >> >> org.apache.spark.sql.execution.QueryExecution.sparkPlan$ >> >> lzycompute(QueryExecution.scala:79) >> >> at >> >> org.apache.spark.sql.execution.QueryExecution.sparkPlan( >> >> QueryExecution.scala:75) >> >> at >> >> org.apache.spark.sql.execution.QueryExecution.executedPlan$ >> >> lzycompute(QueryExecution.scala:84) >> >> at >> >> org.apache.spark.sql.execution.QueryExecution.executedPlan( >> >> QueryExecution.scala:84) >> >> at >> >> org.apache.spark.sql.execution.QueryExecution.toRdd$ >> >> lzycompute(QueryExecution.scala:87) >> >> at >> >> org.apache.spark.sql.execution.QueryExecution.toRdd( >> >> QueryExecution.scala:87) >> >> at org.apache.spark.sql.Dataset. >> >> <init> >> >> (Dataset.scala:185) >> >> at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64) >> >> at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592) >> >> at >> >> org.apache.spark.sql.execution.command.CarbonCreateTableComm >> >> and.processSchema(CarbonCreateTableCommand.scala:84) >> >> at >> >> org.apache.spark.sql.execution.command.CarbonCreateTableComm >> >> and.run(CarbonCreateTableCommand.scala:36) >> >> at >> >> org.apache.spark.sql.execution.command.ExecutedCommandExec.s >> >> ideEffectResult$lzycompute(commands.scala:58) >> >> at >> >> org.apache.spark.sql.execution.command.ExecutedCommandExec.s >> >> ideEffectResult(commands.scala:56) >> >> at >> >> org.apache.spark.sql.execution.command.ExecutedCommandExec. >> >> doExecute(commands.scala:74) >> >> at >> >> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1. >> >> apply(SparkPlan.scala:114) >> >> at >> >> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1. >> >> apply(SparkPlan.scala:114) >> >> at >> >> org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQue >> >> ry$1.apply(SparkPlan.scala:135) >> >> at >> >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperati >> >> onScope.scala:151) >> >> at >> >> org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala: >> 132) >> >> at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan. >> >> scala:113) >> >> at >> >> org.apache.spark.sql.execution.QueryExecution.toRdd$ >> >> lzycompute(QueryExecution.scala:87) >> >> at >> >> org.apache.spark.sql.execution.QueryExecution.toRdd( >> >> QueryExecution.scala:87) >> >> at org.apache.spark.sql.Dataset. >> >> <init> >> >> (Dataset.scala:185) >> >> at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64) >> >> at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592) >> >> ... 52 elided >> >> -- >> >> Sent from: http://apache-carbondata-user-mailing-list.3231.n8.nabble. >> com/ >> >> >> >> >> >> >> >> >> >> >> >> -- >> >> Sent from: http://apache-carbondata-user-mailing-list.3231.n8.nabble. >> com/ >> >> >> >> >>