Re: Removing of kettle code from Carbondata
+1 在 2017/3/11 9:51, Ravindra Pesala 写道: Hi All, I guess it is time to remove the kettle flow from Carbondata loading. Now there are two flows to load the data and becomes difficult to maintain the code.Bug fixing or any feature implementation needs to be done in both the places so it becomes difficult for developer to implement and test. Please comment and vote on it. -- 金铸 技术发展部(TDD) 东软集团股份有限公司 沈阳浑南新区新秀街2号东软软件园A2-105A Postcode:110179 Tel: (86 24)8366 2049 Mobile:13897999526 --- Confidentiality Notice: The information contained in this e-mail and any accompanying attachment(s) is intended only for the use of the intended recipient and may be confidential and/or privileged of Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader of this communication is not the intended recipient, unauthorized use, forwarding, printing, storing, disclosure or copying is strictly prohibited, and may be unlawful.If you have received this communication in error,please immediately notify the sender by return e-mail, and delete the original message and all copies from your system. Thank you. ---
Re: Please vote and advise on building thrift files
+1 for proposal 1 在 2016/11/17 12:13, 邢冰 写道: +1 for proposal 1 thx 发自网易邮箱大师 On 11/17/2016 12:09, Ravindra Pesala wrote: +1 for proposal 1 On 17 November 2016 at 08:23, Xiaoqiao Hewrote: +1 for proposal 1. On Thu, Nov 17, 2016 at 10:31 AM, ZhuWilliam wrote: +1 for proposal 1 . Auto generated code should not be added to project. Also most the of time ,people who dive into carbondata may not touch format code. -- View this message in context: http://apache-carbondata- mailing-list-archive.1130556.n5.nabble.com/Please-vote-and- advise-on-building-thrift-files-tp2952p2957.html Sent from the Apache CarbonData Mailing List archive mailing list archive at Nabble.com. -- Thanks & Regards, Ravi --- Confidentiality Notice: The information contained in this e-mail and any accompanying attachment(s) is intended only for the use of the intended recipient and may be confidential and/or privileged of Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader of this communication is not the intended recipient, unauthorized use, forwarding, printing, storing, disclosure or copying is strictly prohibited, and may be unlawful.If you have received this communication in error,please immediately notify the sender by return e-mail, and delete the original message and all copies from your system. Thank you. ---
Re: [VOTE] Apache CarbonData 0.2.0-incubating release
+1 bingding jinzhu 在 2016/11/11 0:33, Jacky Li 写道: +1 binding Regards, Jacky ---Original--- From: "Aniket Adnaik"Date: 2016/11/10 14:43:49 To: "dev" ;"chenliang613" ; Subject: Re: [VOTE] Apache CarbonData 0.2.0-incubating release +1 Regards, Aniket On 9 Nov 2016 3:17 p.m., "Liang Chen" wrote: Hi all, I submit the CarbonData 0.2.0-incubating to your vote. Release Notes: https://issues.apache.org/jira/secure/ReleaseNote.jspa? projectId=12320220=12337896 Staging Repository: https://repository.apache.org/content/repositories/ orgapachecarbondata-1006 Git Tag: carbondata-0.2.0-incubating Please vote to approve this release: [ ] +1 Approve the release [ ] -1 Don't approve the release (please provide specific comments) This vote will be open for at least 72 hours. If this vote passes (we need at least 3 binding votes, meaning three votes from the PPMC), I will forward to gene...@incubator.apache.org for the IPMC votes. Here is my vote : +1 (binding) Regards Liang > --- Confidentiality Notice: The information contained in this e-mail and any accompanying attachment(s) is intended only for the use of the intended recipient and may be confidential and/or privileged of Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader of this communication is not the intended recipient, unauthorized use, forwarding, printing, storing, disclosure or copying is strictly prohibited, and may be unlawful.If you have received this communication in error,please immediately notify the sender by return e-mail, and delete the original message and all copies from your system. Thank you. ---
Re: [Discuss]Set block_size for table on table level
+1,agree with jihong. 在 2016/9/27 5:12, chenliang613 写道: +1, agree with Jihong's comment : make it as optional, usually the default block size will be used if user don't specially define it. Regards Liang Jihong Ma wrote +1, To avoid potential compatibility issue, we could introduce this param as an optional field, as long as it is not a required field, we are fine with a defined default block size. Regards. Jihong -Original Message- From: Jacky Li [mailto: jacky.likun@ ] Sent: Monday, September 26, 2016 7:29 AM To: dev@.apache Subject: Re: [Discuss]Set block_size for table on table level I am OK with this feature, the only thing I am worrying about is the compatibility of CarbonData file reader. Can you make it compatible when you reading old CarbonData file without this property. We have encountered many times that user need to delete the store and re-load the data. Regards, Jacky 在 2016年9月26日,下午2:15,Ravindra Pesala ravi.pesala@ 写道: +1 At same time max and min block size should be restricted and validated while creating table. On 26 September 2016 at 07:36, Zhangshunyu zhangshunyu1990@ wrote: Purpose: To configure block file size for each table on column level, so that each table could has its own blocksize. My solution: Add a new parameter in table properties, when create a table, the user can set it in ddl. Add a parameter in thrift format just like other properties, and write this info into thrift file so that this info would not lost when cluster is restarted. What's your opinion? -- View this message in context: http://apache-carbondata- mailing-list-archive.1130556.n5.nabble.com/Discuss-Set- block-size-for-table-on-table-level-tp1472.html Sent from the Apache CarbonData Mailing List archive mailing list archive at Nabble.com. -- Thanks & Regards, Ravi -- --- Confidentiality Notice: The information contained in this e-mail and any accompanying attachment(s) is intended only for the use of the intended recipient and may be confidential and/or privileged of Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader of this communication is not the intended recipient, unauthorized use, forwarding, printing, storing, disclosure or copying is strictly prohibited, and may be unlawful.If you have received this communication in error,please immediately notify the sender by return e-mail, and delete the original message and all copies from your system. Thank you. ---
Re: [VOTE] Apache CarbonData 0.1.0-incubating release
+1 在 2016/8/20 2:57, Jean-Baptiste Onofré 写道: Hi all, I submit the first CarbonData release to your vote. Release Notes: https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12320220=12337895 Staging Repository: https://repository.apache.org/content/repositories/orgapachecarbondata-1000/ Git Tag: carbondata-0.1.0-incubating Please vote to approve this release: [ ] +1 Approve the release [ ] -1 Don't approve the release (please provide specific comments) This vote will be open for at least 72 hours. If this vote passes (we need at least 3 binding votes, meaning three votes from the PPMC), I will forward to gene...@incubator.apache.org for the IPMC votes. Thanks Regards JB -- 金铸 技术发展部(TDD) 东软集团股份有限公司 沈阳浑南新区新秀街2号东软软件园A2-105A Postcode:110179 Tel: (86 24)8366 2049 Mobile:13897999526 --- Confidentiality Notice: The information contained in this e-mail and any accompanying attachment(s) is intended only for the use of the intended recipient and may be confidential and/or privileged of Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader of this communication is not the intended recipient, unauthorized use, forwarding, printing, storing, disclosure or copying is strictly prohibited, and may be unlawful.If you have received this communication in error,please immediately notify the sender by return e-mail, and delete the original message and all copies from your system. Thank you. ---
Re: Open discussion and Vote: What kind of JIRA issue events need send mail to dev@carbondata.incubator.apache.org
Option2:+1 在 2016/8/18 15:57, Liang Big data 写道: Hi all Please discuss and vote, do you think what kind of JIRA issue events need send mails to dev@carbondata.incubator.apache.org? Option1: None, all JIRA issue events don't need send any mails to dev@carbondata.incubator.apache.org, users can directly go to apache jira to check issues' content. Option2:Issue Created and Issue Commented, the two JIRA issue events send mails to dev@carbondata.incubator.apache.org Option3:Keep the current notification schema, the below events send mails to dev@carbondata.incubator.apache.org. or any other option? The below is current notification schema EventsNotifications Issue Created - All Watchers - Current Assignee - Reporter - Single Email Address (dev@carbondata.incubator.apache.org) Issue Updated - All Watchers - Current Assignee - Reporter - Single Email Address (dev@carbondata.incubator.apache.org) Issue Assigned - All Watchers - Current Assignee - Reporter - Single Email Address (dev@carbondata.incubator.apache.org) Issue Resolved - All Watchers - Current Assignee - Reporter - Single Email Address (dev@carbondata.incubator.apache.org) Issue Closed - All Watchers - Current Assignee - Reporter - Single Email Address (dev@carbondata.incubator.apache.org) Issue Commented - All Watchers - Current Assignee - Reporter - Single Email Address (dev@carbondata.incubator.apache.org) Issue Comment Edited - All Watchers - Current Assignee - Reporter - Single Email Address (dev@carbondata.incubator.apache.org) Issue Comment Deleted - All Watchers - Current Assignee - Reporter - Single Email Address (dev@carbondata.incubator.apache.org) Issue Reopened - All Watchers - Current Assignee - Reporter - Single Email Address (dev@carbondata.incubator.apache.org) Issue Deleted - All Watchers - Current Assignee - Reporter - Single Email Address (dev@carbondata.incubator.apache.org) Issue Moved - Single Email Address (dev@carbondata.incubator.apache.org) Regards Liang --- Confidentiality Notice: The information contained in this e-mail and any accompanying attachment(s) is intended only for the use of the intended recipient and may be confidential and/or privileged of Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader of this communication is not the intended recipient, unauthorized use, forwarding, printing, storing, disclosure or copying is strictly prohibited, and may be unlawful.If you have received this communication in error,please immediately notify the sender by return e-mail, and delete the original message and all copies from your system. Thank you. ---
Re: 答复: load data fail
i drop table,use $hdc_home/hive/conf/hive-sie.xml replace $hdc_home/spark/conf/hive-site.xml,fixed it. but i do not know the principle inside. if t4 is exist in hive's defualt,in other words create table t4 in hive,then create table in carbondata do not reported exception。 在 2016/8/17 10:35, Chenliang (Liang, CarbonData) 写道: Can you share the case experience: how did you solve it. Regards Liang -邮件原件- 发件人: 金铸 [mailto:jin...@neusoft.com] 发送时间: 2016年8月17日 10:31 收件人: dev@carbondata.incubator.apache.org 主题: Re: load data fail thanks a lot,I solve this。 在 2016/8/17 0:53, Eason 写道: hi jinzhu, whether this happen on multiple instance loading the same table? currently ,it is no support concurrent load on same table. for this exception 1.please check if any locks are created under system temp folder with//lockfile, if it exists please delete. 2.try to change the lock ype: carbon.lock.type = ZOOKEEPERLOCK Regards, Eason 在 2016年08月12日 14:25, 金铸 写道: hi : /usr/hdp/2.4.0.0-169/spark/bin/spark-shell --master yarn-client --jars /opt/incubator-carbondata/assembly/target/scala-2.10/carbondata_2.10- 0.1.0-incubating-SNAPSHOT-shade-hadoop2.2.0.jar,/usr/hdp/2.4.0.0-169/ spark/lib/datanucleus-api-jdo-3.2.6.jar,/usr/hdp/2.4.0.0-169/spark/li b/datanucleus-rdbms-3.2.9.jar,/usr/hdp/2.4.0.0-169/spark/lib/datanucl eus-core-3.2.10.jar,/opt//mysql-connector-java-5.1.37.jar scala>import org.apache.spark.sql.CarbonContext scala>import java.io.File scala>import org.apache.hadoop.hive.conf.HiveConf scala>val cc = new CarbonContext(sc, "hdfs://hadoop01/data/carbondata01/store") scala>cc.setConf("hive.metastore.warehouse.dir", "/apps/hive/warehouse") scala>cc.setConf(HiveConf.ConfVars.HIVECHECKFILEFORMAT.varname, "false") scala>cc.setConf("carbon.kettle.home","/usr/hdp/2.4.0.0-169/spark/car scala>bonlib/carbonplugins") scala> cc.setConf("carbon.kettle.home","/usr/hdp/2.4.0.0-169/spark/carbonlib /carbonplugins") scala> cc.sql(s"load data local inpath 'hdfs://hadoop01/sample.csv' into table t4 options('FILEHEADER'='id,name,city,age')") INFO 12-08 14:21:24,461 - main Query [LOAD DATA LOCAL INPATH 'HDFS://HADOOP01/SAMPLE.CSV' INTO TABLE T4 OPTIONS('FILEHEADER'='ID,NAME,CITY,AGE')] INFO 12-08 14:21:39,475 - Table MetaData Unlocked Successfully after data load java.lang.RuntimeException: Table is locked for updation. Please try after some time at scala.sys.package$.error(package.scala:27) at org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1049) at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58) at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56) at org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.s cala:150) thanks a lot - --- Confidentiality Notice: The information contained in this e-mail and any accompanying attachment(s) is intended only for the use of the intended recipient and may be confidential and/or privileged of Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader of this communication is not the intended recipient, unauthorized use, forwarding, printing, storing, disclosure or copying is strictly prohibited, and may be unlawful.If you have received this communication in error,please immediately notify the sender by return e-mail, and delete the original message and all copies from your system. Thank you. ---
load data fail
hi : /usr/hdp/2.4.0.0-169/spark/bin/spark-shell --master yarn-client --jars /opt/incubator-carbondata/assembly/target/scala-2.10/carbondata_2.10-0.1.0-incubating-SNAPSHOT-shade-hadoop2.2.0.jar,/usr/hdp/2.4.0.0-169/spark/lib/datanucleus-api-jdo-3.2.6.jar,/usr/hdp/2.4.0.0-169/spark/lib/datanucleus-rdbms-3.2.9.jar,/usr/hdp/2.4.0.0-169/spark/lib/datanucleus-core-3.2.10.jar,/opt//mysql-connector-java-5.1.37.jar scala>import org.apache.spark.sql.CarbonContext scala>import java.io.File scala>import org.apache.hadoop.hive.conf.HiveConf scala>val cc = new CarbonContext(sc, "hdfs://hadoop01/data/carbondata01/store") scala>cc.setConf("hive.metastore.warehouse.dir", "/apps/hive/warehouse") scala>cc.setConf(HiveConf.ConfVars.HIVECHECKFILEFORMAT.varname, "false") scala>cc.setConf("carbon.kettle.home","/usr/hdp/2.4.0.0-169/spark/carbonlib/carbonplugins") scala> cc.setConf("carbon.kettle.home","/usr/hdp/2.4.0.0-169/spark/carbonlib/carbonplugins") scala> cc.sql(s"load data local inpath 'hdfs://hadoop01/sample.csv' into table t4 options('FILEHEADER'='id,name,city,age')") INFO 12-08 14:21:24,461 - main Query [LOAD DATA LOCAL INPATH 'HDFS://HADOOP01/SAMPLE.CSV' INTO TABLE T4 OPTIONS('FILEHEADER'='ID,NAME,CITY,AGE')] INFO 12-08 14:21:39,475 - Table MetaData Unlocked Successfully after data load java.lang.RuntimeException: Table is locked for updation. Please try after some time at scala.sys.package$.error(package.scala:27) at org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1049) at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58) at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56) at org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150) thanks a lot --- Confidentiality Notice: The information contained in this e-mail and any accompanying attachment(s) is intended only for the use of the intended recipient and may be confidential and/or privileged of Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader of this communication is not the intended recipient, unauthorized use, forwarding, printing, storing, disclosure or copying is strictly prohibited, and may be unlawful.If you have received this communication in error,please immediately notify the sender by return e-mail, and delete the original message and all copies from your system. Thank you. ---
Re: Discussing to add carbondata-tools module
+1 在 2016/8/5 17:15, Jean-Baptiste Onofré 写道: I guess it's where we can put the checkstyle, ... (what we have in the dev folder right now), correct ? Regards JB On 08/05/2016 08:31 AM, QiangCai wrote: Hi all, To improve the CarbonData system's usability and maintainability, I suggest to add carbondata-tools module. I think this module should provide some command tools as following. 1. import import a data file/folder to any existing table 2. export export the given columns to a file 3. schema show the detail information of the specified table schema file 4. metadata show tablestatus metadata show the history track for dataloading and compaction 5. footer show blocklet metadata list show start/end key, min/max value, row number, total size for the specified blocklet 6. blocklet show blocket list show blocket data, RLE map, Inverted index for the given columns 7. index show BTree node list show node information:start/end key, min/max value) 8. dictionary show key-value list of specified gloabl/local dictionary file show sort index show dictionary metadata Thank you and look forward to having your opinion on this carbon-tools module. David Cai -- View this message in context: http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/Discussing-to-add-carbondata-tools-module-tp4.html Sent from the Apache CarbonData Mailing List archive mailing list archive at Nabble.com. -- 金铸 技术发展部(TDD) 东软集团股份有限公司 沈阳浑南新区新秀街2号东软软件园A2-105A Postcode:110179 Tel: (86 24)8366 2049 Mobile:13897999526 --- Confidentiality Notice: The information contained in this e-mail and any accompanying attachment(s) is intended only for the use of the intended recipient and may be confidential and/or privileged of Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader of this communication is not the intended recipient, unauthorized use, forwarding, printing, storing, disclosure or copying is strictly prohibited, and may be unlawful.If you have received this communication in error,please immediately notify the sender by return e-mail, and delete the original message and all copies from your system. Thank you. ---
Re: load data error
hi Liang: Thanks your repay。 I have already used the “/opt/incubator-carbondata/sample.csv” ,Reported the same error。 在 2016/7/30 22:44, Liang Big data 写道: Hi jinzhu金铸: please check the below error:the input path having some issues. Please use the absolute path to try it again. --- ERROR 29-07 16:39:46,904 - main org.apache.hadoop.mapreduce.lib.input.InvalidInputException: Input path does not exist: /opt/incubator-carbondata/sample.csv Regards Liang 2016-07-29 8:47 GMT+08:00 金铸 <jin...@neusoft.com>: [hadoop@slave2 ~]$ cat /opt/incubator-carbondata/sample.csv id,name,city,age 1,david,shenzhen,31 2,eason,shenzhen,27 3,jarry,wuhan,35 [hadoop@slave2 ~]$ > load data inpath '../sample.csv' into table test_table; INFO 29-07 16:39:46,087 - main Property file path: /opt/incubator-carbondata/bin/../../../conf/carbon.properties INFO 29-07 16:39:46,087 - main --Using Carbon.properties INFO 29-07 16:39:46,087 - main {} INFO 29-07 16:39:46,088 - main Query [LOAD DATA INPATH '../SAMPLE.CSV' INTO TABLE TEST_TABLE] INFO 29-07 16:39:46,527 - Successfully able to get the table metadata file lock INFO 29-07 16:39:46,537 - main Initiating Direct Load for the Table : (default.test_table) INFO 29-07 16:39:46,541 - Generate global dictionary from source data files! INFO 29-07 16:39:46,569 - [Block Distribution] INFO 29-07 16:39:46,569 - totalInputSpaceConsumed : 74 , defaultParallelism : 6 INFO 29-07 16:39:46,569 - mapreduce.input.fileinputformat.split.maxsize : 16777216 INFO 29-07 16:39:46,689 - Block broadcast_0 stored as values in memory (estimated size 232.6 KB, free 232.6 KB) INFO 29-07 16:39:46,849 - Block broadcast_0_piece0 stored as bytes in memory (estimated size 19.7 KB, free 252.3 KB) INFO 29-07 16:39:46,850 - Added broadcast_0_piece0 in memory on 192.168.241.223:41572 (size: 19.7 KB, free: 511.5 MB) INFO 29-07 16:39:46,856 - Created broadcast 0 from NewHadoopRDD at CarbonTextFile.scala:45 ERROR 29-07 16:39:46,904 - generate global dictionary failed ERROR 29-07 16:39:46,904 - main org.apache.hadoop.mapreduce.lib.input.InvalidInputException: Input path does not exist: /opt/incubator-carbondata/sample.csv at org.apache.hadoop.mapreduce.lib.input.FileInputFormat.singleThreadedListStatus(FileInputFormat.java:321) at org.apache.hadoop.mapreduce.lib.input.FileInputFormat.listStatus(FileInputFormat.java:264) at org.apache.hadoop.mapreduce.lib.input.FileInputFormat.getSplits(FileInputFormat.java:385) at org.apache.spark.rdd.NewHadoopRDD.getPartitions(NewHadoopRDD.scala:120) at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239) at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237) at scala.Option.getOrElse(Option.scala:120) at org.apache.spark.rdd.RDD.partitions(RDD.scala:237) at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35) at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239) at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237) at scala.Option.getOrElse(Option.scala:120) at org.apache.spark.rdd.RDD.partitions(RDD.scala:237) at org.apache.spark.rdd.RDD$$anonfun$take$1.apply(RDD.scala:1307) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111) at org.apache.spark.rdd.RDD.withScope(RDD.scala:316) at org.apache.spark.rdd.RDD.take(RDD.scala:1302) at com.databricks.spark.csv.CarbonCsvRelation.firstLine$lzycompute(CarbonCsvRelation.scala:175) at com.databricks.spark.csv.CarbonCsvRelation.firstLine(CarbonCsvRelation.scala:170) at com.databricks.spark.csv.CarbonCsvRelation.inferSchema(CarbonCsvRelation.scala:141) at com.databricks.spark.csv.CarbonCsvRelation.(CarbonCsvRelation.scala:71) at com.databricks.spark.csv.newapi.DefaultSource.createRelation(DefaultSource.scala:142) at com.databricks.spark.csv.newapi.DefaultSource.createRelation(DefaultSource.scala:44) at org.apache.spark.sql.execution.datasources.ResolvedDataSource$.apply(ResolvedDataSource.scala:158) at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:119) at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:109) at org.carbondata.spark.util.GlobalDictionaryUtil$.loadDataFrame(GlobalDictionaryUtil.scala:365) at org.carbondata.spark.util.GlobalDictionaryUtil$.generateGlobalDictionary(GlobalDictionaryUtil.scala:676) at org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1159) at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58) at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56) at org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands
load data error
ame.(DataFrame.scala:145) at org.apache.spark.sql.DataFrame.(DataFrame.scala:130) at org.carbondata.spark.rdd.CarbonDataFrameRDD.(CarbonDataFrameRDD.scala:23) at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:131) at org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:63) at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:311) at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376) at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:226) at org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver$.main(CarbonSQLCLIDriver.scala:40) at org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver.main(CarbonSQLCLIDriver.scala) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) -- 金铸 --- Confidentiality Notice: The information contained in this e-mail and any accompanying attachment(s) is intended only for the use of the intended recipient and may be confidential and/or privileged of Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader of this communication is not the intended recipient, unauthorized use, forwarding, printing, storing, disclosure or copying is strictly prohibited, and may be unlawful.If you have received this communication in error,please immediately notify the sender by return e-mail, and delete the original message and all copies from your system. Thank you. ---