When I change the SaveMode.Append to Override,then the error is more weird:



INFO  25-11 20:19:46,572 - streaming-job-executor-0 Query [
          CREATE TABLE IF NOT EXISTS DEFAULT.CARBON2
          (A STRING, B STRING)
          STORED BY 'ORG.APACHE.CARBONDATA.FORMAT'
      ]
INFO  25-11 20:19:46,656 - Parsing command: 
          CREATE TABLE IF NOT EXISTS default.carbon2
          (a STRING, b STRING)
          STORED BY 'org.apache.carbondata.format'
      
INFO  25-11 20:19:46,663 - Parse Completed
AUDIT 25-11 20:19:46,860 - [allwefantasy][allwefantasy][Thread-100]Creating
Table with Database name [default] and Table name [carbon2]
INFO  25-11 20:19:46,889 - 1: get_tables: db=default pat=.*
INFO  25-11 20:19:46,889 - ugi=allwefantasy     ip=unknown-ip-addr
cmd=get_tables: db=default pat=.*       
INFO  25-11 20:19:46,889 - 1: Opening raw store with implemenation
class:org.apache.hadoop.hive.metastore.ObjectStore
INFO  25-11 20:19:46,891 - ObjectStore, initialize called
INFO  25-11 20:19:46,897 - Reading in results for query
"org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is
closing
INFO  25-11 20:19:46,898 - Using direct SQL, underlying DB is MYSQL
INFO  25-11 20:19:46,898 - Initialized ObjectStore
INFO  25-11 20:19:46,954 - streaming-job-executor-0 Table block size not
specified for default_carbon2. Therefore considering the default value 1024
MB
INFO  25-11 20:19:46,978 - Table carbon2 for Database default created
successfully.
INFO  25-11 20:19:46,978 - streaming-job-executor-0 Table carbon2 for
Database default created successfully.
AUDIT 25-11 20:19:46,978 - [allwefantasy][allwefantasy][Thread-100]Creating
timestamp file for default.carbon2
INFO  25-11 20:19:46,979 - streaming-job-executor-0 Query [CREATE TABLE
DEFAULT.CARBON2 USING CARBONDATA OPTIONS (TABLENAME "DEFAULT.CARBON2",
TABLEPATH "FILE:///TMP/CARBONDATA/STORE/DEFAULT/CARBON2") ]
INFO  25-11 20:19:47,033 - 1: get_table : db=default tbl=carbon2
INFO  25-11 20:19:47,034 - ugi=allwefantasy     ip=unknown-ip-addr      
cmd=get_table
: db=default tbl=carbon2        
WARN  25-11 20:19:47,062 - Couldn't find corresponding Hive SerDe for data
source provider carbondata. Persisting data source relation
`default`.`carbon2` into Hive metastore in Spark SQL specific format, which
is NOT compatible with Hive.
INFO  25-11 20:19:47,247 - 1: create_table: Table(tableName:carbon2,
dbName:default, owner:allwefantasy, createTime:1480076387, lastAccessTime:0,
retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:col,
type:array<string>, comment:from deserializer)], location:null,
inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat,
outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat,
compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null,
serializationLib:org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe,
parameters:{tableName=default.carbon2, serialization.format=1,
tablePath=file:///tmp/carbondata/store/default/carbon2}), bucketCols:[],
sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[],
skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[],
parameters:{EXTERNAL=TRUE, spark.sql.sources.provider=carbondata},
viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE,
privileges:PrincipalPrivilegeSet(userPrivileges:{}, groupPrivileges:null,
rolePrivileges:null))
INFO  25-11 20:19:47,247 - ugi=allwefantasy     ip=unknown-ip-addr
cmd=create_table: Table(tableName:carbon2, dbName:default,
owner:allwefantasy, createTime:1480076387, lastAccessTime:0, retention:0,
sd:StorageDescriptor(cols:[FieldSchema(name:col, type:array<string>,
comment:from deserializer)], location:null,
inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat,
outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat,
compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null,
serializationLib:org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe,
parameters:{tableName=default.carbon2, serialization.format=1,
tablePath=file:///tmp/carbondata/store/default/carbon2}), bucketCols:[],
sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[],
skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[],
parameters:{EXTERNAL=TRUE, spark.sql.sources.provider=carbondata},
viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE,
privileges:PrincipalPrivilegeSet(userPrivileges:{}, groupPrivileges:null,
rolePrivileges:null))   
INFO  25-11 20:19:47,257 - Creating directory if it doesn't exist:
file:/tmp/user/hive/warehouse/carbon2
AUDIT 25-11 20:19:47,564 - [allwefantasy][allwefantasy][Thread-100]Table
created with Database name [default] and Table name [carbon2]
org.apache.spark.sql.catalyst.analysis.NoSuchTableException
        at
org.apache.spark.sql.hive.CarbonMetastoreCatalog.lookupRelation1(CarbonMetastoreCatalog.scala:141)
        at
org.apache.spark.sql.hive.CarbonMetastoreCatalog.lookupRelation1(CarbonMetastoreCatalog.scala:127)
        at
org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1044)
        at
org.apache.carbondata.spark.CarbonDataFrameWriter.loadDataFrame(CarbonDataFrameWriter.scala:132)
        at
org.apache.carbondata.spark.CarbonDataFrameWriter.writeToCarbonFile(CarbonDataFrameWriter.scala:52)
        at
org.apache.carbondata.spark.CarbonDataFrameWriter.saveAsCarbonFile(CarbonDataFrameWriter.scala:37)
        at
org.apache.spark.sql.CarbonSource.createRelation(CarbonDatasourceRelation.scala:110)
        at
org.apache.spark.sql.execution.datasources.ResolvedDataSource$.apply(ResolvedDataSource.scala:222)
        at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:148)
        at
streaming.core.compositor.spark.streaming.output.SQLOutputCompositor$$anonfun$result$1.apply(SQLOutputCompositor.scala:60)
        at
streaming.core.compositor.spark.streaming.output.SQLOutputCompositor$$anonfun$result$1.apply(SQLOutputCompositor.scala:53)
        at
org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:661)
        at
org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:661)
        at
org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ForEachDStream.scala:50)
        at
org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:50)
        at
org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:50)
        at
org.apache.spark.streaming.dstream.DStream.createRDDWithLocalProperties(DStream.scala:426)
        at
org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply$mcV$sp(ForEachDStream.scala:49)
        at
org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:49)
        at
org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:49)
        at scala.util.Try$.apply(Try.scala:161)
        at org.apache.spark.streaming.scheduler.Job.run(Job.scala:39)
        at
org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply$mcV$sp(JobScheduler.scala:224)
        at
org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:224)
        at
org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:224)
        at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57)
        at
org.apache.spark.streaming.scheduler.JobScheduler$JobHandler.run(JobScheduler.scala:223)
        at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)
AUDIT 25-11 20:19:47,584 - [allwefantasy][allwefantasy][Thread-100]Table Not
Found: carbon2
INFO  25-11 20:19:47,588 - Finished job streaming job 1480076380000 ms.0
from job set of time 1480076380000 ms
INFO  25-11 20:19:47,590 - Total delay: 7.586 s for time 1480076380000 ms
(execution: 7.547 s)



--
View this message in context: 
http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/Using-DataFrame-to-write-carbondata-file-cause-no-table-found-error-tp3203p3212.html
Sent from the Apache CarbonData Mailing List archive mailing list archive at 
Nabble.com.

Reply via email to