When I change the SaveMode.Append to Override,then the error is more weird:
INFO 25-11 20:19:46,572 - streaming-job-executor-0 Query [ CREATE TABLE IF NOT EXISTS DEFAULT.CARBON2 (A STRING, B STRING) STORED BY 'ORG.APACHE.CARBONDATA.FORMAT' ] INFO 25-11 20:19:46,656 - Parsing command: CREATE TABLE IF NOT EXISTS default.carbon2 (a STRING, b STRING) STORED BY 'org.apache.carbondata.format' INFO 25-11 20:19:46,663 - Parse Completed AUDIT 25-11 20:19:46,860 - [allwefantasy][allwefantasy][Thread-100]Creating Table with Database name [default] and Table name [carbon2] INFO 25-11 20:19:46,889 - 1: get_tables: db=default pat=.* INFO 25-11 20:19:46,889 - ugi=allwefantasy ip=unknown-ip-addr cmd=get_tables: db=default pat=.* INFO 25-11 20:19:46,889 - 1: Opening raw store with implemenation class:org.apache.hadoop.hive.metastore.ObjectStore INFO 25-11 20:19:46,891 - ObjectStore, initialize called INFO 25-11 20:19:46,897 - Reading in results for query "org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is closing INFO 25-11 20:19:46,898 - Using direct SQL, underlying DB is MYSQL INFO 25-11 20:19:46,898 - Initialized ObjectStore INFO 25-11 20:19:46,954 - streaming-job-executor-0 Table block size not specified for default_carbon2. Therefore considering the default value 1024 MB INFO 25-11 20:19:46,978 - Table carbon2 for Database default created successfully. INFO 25-11 20:19:46,978 - streaming-job-executor-0 Table carbon2 for Database default created successfully. AUDIT 25-11 20:19:46,978 - [allwefantasy][allwefantasy][Thread-100]Creating timestamp file for default.carbon2 INFO 25-11 20:19:46,979 - streaming-job-executor-0 Query [CREATE TABLE DEFAULT.CARBON2 USING CARBONDATA OPTIONS (TABLENAME "DEFAULT.CARBON2", TABLEPATH "FILE:///TMP/CARBONDATA/STORE/DEFAULT/CARBON2") ] INFO 25-11 20:19:47,033 - 1: get_table : db=default tbl=carbon2 INFO 25-11 20:19:47,034 - ugi=allwefantasy ip=unknown-ip-addr cmd=get_table : db=default tbl=carbon2 WARN 25-11 20:19:47,062 - Couldn't find corresponding Hive SerDe for data source provider carbondata. Persisting data source relation `default`.`carbon2` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive. INFO 25-11 20:19:47,247 - 1: create_table: Table(tableName:carbon2, dbName:default, owner:allwefantasy, createTime:1480076387, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:col, type:array<string>, comment:from deserializer)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe, parameters:{tableName=default.carbon2, serialization.format=1, tablePath=file:///tmp/carbondata/store/default/carbon2}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{EXTERNAL=TRUE, spark.sql.sources.provider=carbondata}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE, privileges:PrincipalPrivilegeSet(userPrivileges:{}, groupPrivileges:null, rolePrivileges:null)) INFO 25-11 20:19:47,247 - ugi=allwefantasy ip=unknown-ip-addr cmd=create_table: Table(tableName:carbon2, dbName:default, owner:allwefantasy, createTime:1480076387, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:col, type:array<string>, comment:from deserializer)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe, parameters:{tableName=default.carbon2, serialization.format=1, tablePath=file:///tmp/carbondata/store/default/carbon2}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{EXTERNAL=TRUE, spark.sql.sources.provider=carbondata}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE, privileges:PrincipalPrivilegeSet(userPrivileges:{}, groupPrivileges:null, rolePrivileges:null)) INFO 25-11 20:19:47,257 - Creating directory if it doesn't exist: file:/tmp/user/hive/warehouse/carbon2 AUDIT 25-11 20:19:47,564 - [allwefantasy][allwefantasy][Thread-100]Table created with Database name [default] and Table name [carbon2] org.apache.spark.sql.catalyst.analysis.NoSuchTableException at org.apache.spark.sql.hive.CarbonMetastoreCatalog.lookupRelation1(CarbonMetastoreCatalog.scala:141) at org.apache.spark.sql.hive.CarbonMetastoreCatalog.lookupRelation1(CarbonMetastoreCatalog.scala:127) at org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1044) at org.apache.carbondata.spark.CarbonDataFrameWriter.loadDataFrame(CarbonDataFrameWriter.scala:132) at org.apache.carbondata.spark.CarbonDataFrameWriter.writeToCarbonFile(CarbonDataFrameWriter.scala:52) at org.apache.carbondata.spark.CarbonDataFrameWriter.saveAsCarbonFile(CarbonDataFrameWriter.scala:37) at org.apache.spark.sql.CarbonSource.createRelation(CarbonDatasourceRelation.scala:110) at org.apache.spark.sql.execution.datasources.ResolvedDataSource$.apply(ResolvedDataSource.scala:222) at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:148) at streaming.core.compositor.spark.streaming.output.SQLOutputCompositor$$anonfun$result$1.apply(SQLOutputCompositor.scala:60) at streaming.core.compositor.spark.streaming.output.SQLOutputCompositor$$anonfun$result$1.apply(SQLOutputCompositor.scala:53) at org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:661) at org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:661) at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ForEachDStream.scala:50) at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:50) at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:50) at org.apache.spark.streaming.dstream.DStream.createRDDWithLocalProperties(DStream.scala:426) at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply$mcV$sp(ForEachDStream.scala:49) at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:49) at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:49) at scala.util.Try$.apply(Try.scala:161) at org.apache.spark.streaming.scheduler.Job.run(Job.scala:39) at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply$mcV$sp(JobScheduler.scala:224) at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:224) at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:224) at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57) at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler.run(JobScheduler.scala:223) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) AUDIT 25-11 20:19:47,584 - [allwefantasy][allwefantasy][Thread-100]Table Not Found: carbon2 INFO 25-11 20:19:47,588 - Finished job streaming job 1480076380000 ms.0 from job set of time 1480076380000 ms INFO 25-11 20:19:47,590 - Total delay: 7.586 s for time 1480076380000 ms (execution: 7.547 s) -- View this message in context: http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/Using-DataFrame-to-write-carbondata-file-cause-no-table-found-error-tp3203p3212.html Sent from the Apache CarbonData Mailing List archive mailing list archive at Nabble.com.