Hi,

Can you please check if you have spark-catalyst jar in $SPARK_HOME/jars
folder for your  cloudera version, if its not there please try to include
it and retry.

Thanks and regards
Bhavya

On Sun, Nov 5, 2017 at 7:24 PM, Lionel CL <whuca...@outlook.com> wrote:

> I have the same problem in CDH 5.8.0
> spark2 version is 2.1.0.cloudera1
> carbondata version 1.2.0.
>
> There's no error occurred when using open source version spark.
>
> <hadoop.version>2.6.0-cdh5.8.0</hadoop.version>
> <spark.version>2.1.0.cloudera1</spark.version>
> <scala.binary.version>2.11</scala.binary.version>
> <scala.version>2.11.8</scala.version>
>
>
> scala> cc.sql("create table t111(vin string) stored by 'carbondata'")
> 17/11/03 10:22:03 AUDIT command.CreateTable: [][][Thread-1]Creating Table
> with Database name [default] and Table name [t111]
> java.lang.NoSuchMethodError: org.apache.spark.sql.catalyst.
> catalog.CatalogTable.copy(Lorg/apache/spark/sql/catalyst/
> TableIdentifier;Lorg/apache/spark/sql/catalyst/catalog/
> CatalogTableType;Lorg/apache/spark/sql/catalyst/catalog/
> CatalogStorageFormat;Lorg/apache/spark/sql/types/StructT
> ype;Lscala/Option;Lscala/collection/Seq;Lscala/Option;
> Ljava/lang/String;JJLscala/collection/immutable/Map;
> Lscala/Option;Lscala/Option;Lscala/Option;Lscala/Option;
> Lscala/collection/Seq;Z)Lorg/apache/spark/sql/catalyst/
> catalog/CatalogTable;
>   at org.apache.spark.sql.CarbonSource$.updateCatalogTableWithCar
> bonSchema(CarbonSource.scala:253)
>   at org.apache.spark.sql.execution.command.DDLStrategy.apply(
> DDLStrategy.scala:135)
>   at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun
> $1.apply(QueryPlanner.scala:62)
>   at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun
> $1.apply(QueryPlanner.scala:62)
>   at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
>   at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
>   at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:439)
>
>
> 在 2017/11/1 上午1:58,“chenliang613”<chenliang6...@gmail.com<mailto:chenlia
> ng6...@gmail.com>> 写入:
>
> Hi
>
> Did you use open source spark version?
>
> Can you provide more detail info :
> 1. which carbondata version and spark version, you used ?
> 2. Can you share with us , reproduce script and steps.
>
> Regards
> Liang
>
>
> hujianjun wrote
> scala> carbon.sql("CREATE TABLE IF NOT EXISTS carbon_table(id string,name
> string,city string,age Int)STORED BY 'carbondata'")
> 17/10/23 19:13:52 AUDIT command.CarbonCreateTableCommand:
> [master][root][Thread-1]Creating Table with Database name [clb_carbon] and
> Table name [carbon_table]
> java.lang.NoSuchMethodError:
> org.apache.spark.sql.catalyst.catalog.CatalogTable.copy(Lorg
> /apache/spark/sql/catalyst/TableIdentifier;Lorg/apache/
> spark/sql/catalyst/catalog/CatalogTableType;Lorg/apache/
> spark/sql/catalyst/catalog/CatalogStorageFormat;Lorg/
> apache/spark/sql/types/StructType;Lscala/Option;Lscala/
> collection/Seq;Lscala/Option;Ljava/lang/String;JJLscala/
> collection/immutable/Map;Lscala/Option;Lscala/Option;
> Lscala/Option;Lscala/Option;Lscala/collection/Seq;Z)Lorg/
> apache/spark/sql/catalyst/catalog/CatalogTable;
>    at
> org.apache.spark.sql.CarbonSource$.updateCatalogTableWithCar
> bonSchema(CarbonSource.scala:253)
>    at
> org.apache.spark.sql.execution.strategy.DDLStrategy.apply(
> DDLStrategy.scala:154)
>    at
> org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun
> $1.apply(QueryPlanner.scala:62)
>    at
> org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun
> $1.apply(QueryPlanner.scala:62)
>    at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
>    at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
>    at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:439)
>    at
> org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(Que
> ryPlanner.scala:92)
>    at
> org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun
> $2$$anonfun$apply$2.apply(QueryPlanner.scala:77)
>    at
> org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun
> $2$$anonfun$apply$2.apply(QueryPlanner.scala:74)
>    at
> scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(T
> raversableOnce.scala:157)
>    at
> scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(T
> raversableOnce.scala:157)
>    at scala.collection.Iterator$class.foreach(Iterator.scala:893)
>    at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
>    at
> scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.scala:157)
>    at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1336)
>    at
> org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun
> $2.apply(QueryPlanner.scala:74)
>    at
> org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun
> $2.apply(QueryPlanner.scala:66)
>    at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
>    at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
>    at
> org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(Que
> ryPlanner.scala:92)
>    at
> org.apache.spark.sql.execution.QueryExecution.sparkPlan$
> lzycompute(QueryExecution.scala:79)
>    at
> org.apache.spark.sql.execution.QueryExecution.sparkPlan(
> QueryExecution.scala:75)
>    at
> org.apache.spark.sql.execution.QueryExecution.executedPlan$
> lzycompute(QueryExecution.scala:84)
>    at
> org.apache.spark.sql.execution.QueryExecution.executedPlan(
> QueryExecution.scala:84)
>    at
> org.apache.spark.sql.execution.QueryExecution.toRdd$
> lzycompute(QueryExecution.scala:87)
>    at
> org.apache.spark.sql.execution.QueryExecution.toRdd(
> QueryExecution.scala:87)
>    at org.apache.spark.sql.Dataset.
> <init>
> (Dataset.scala:185)
>    at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64)
>    at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592)
>    at
> org.apache.spark.sql.execution.command.CarbonCreateTableComm
> and.processSchema(CarbonCreateTableCommand.scala:84)
>    at
> org.apache.spark.sql.execution.command.CarbonCreateTableComm
> and.run(CarbonCreateTableCommand.scala:36)
>    at
> org.apache.spark.sql.execution.command.ExecutedCommandExec.s
> ideEffectResult$lzycompute(commands.scala:58)
>    at
> org.apache.spark.sql.execution.command.ExecutedCommandExec.s
> ideEffectResult(commands.scala:56)
>    at
> org.apache.spark.sql.execution.command.ExecutedCommandExec.
> doExecute(commands.scala:74)
>    at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.
> apply(SparkPlan.scala:114)
>    at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.
> apply(SparkPlan.scala:114)
>    at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQue
> ry$1.apply(SparkPlan.scala:135)
>    at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperati
> onScope.scala:151)
>    at
> org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:132)
>    at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.
> scala:113)
>    at
> org.apache.spark.sql.execution.QueryExecution.toRdd$
> lzycompute(QueryExecution.scala:87)
>    at
> org.apache.spark.sql.execution.QueryExecution.toRdd(
> QueryExecution.scala:87)
>    at org.apache.spark.sql.Dataset.
> <init>
> (Dataset.scala:185)
>    at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64)
>    at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592)
>    ... 52 elided
> --
> Sent from: http://apache-carbondata-user-mailing-list.3231.n8.nabble.com/
>
>
>
>
>
> --
> Sent from: http://apache-carbondata-user-mailing-list.3231.n8.nabble.com/
>
>

Reply via email to