Hi, Can you set the store location in carbon properties. *carbon.storelocation=hdfs://master:9000/carbondata/<store_location>*
-Regards Kumar Vishal On Tue, Dec 27, 2016 at 4:27 PM, 251469031 <251469...@qq.com> wrote: > I'm using spark 1.6.0, and the carbondata is the latest master branch in > the github. > > > my carbon.properties is configured as : > > > carbon.ddl.base.hdfs.url=hdfs://master:9000/carbondata/data > carbon.badRecords.location=/opt/Carbon/Spark/badrecords > carbon.kettle.home=/opt/spark-1.6.0/carbonlib/carbonplugins > > carbon.lock.type=HDFSLOCK > .... > > > my spark-default.conf is configured as : > > > spark.master spark://master:7077 > spark.yarn.dist.files > /opt/spark-1.6.0/conf/carbon.properties > spark.yarn.dist.archives /opt/spark-1.6.0/carbonlib/ > carbondata_2.10-1.0.0-incubating-SNAPSHOT-shade-hadoop2.7.2.jar > spark.executor.extraJavaOptions > -Dcarbon.properties.filepath=carbon.properties > #spark.executor.extraClassPath /opt/spark-1.6.0/carbonlib/ > carbondata_2.10-1.0.0-incubating-SNAPSHOT-shade-hadoop2.2.0.jar > #spark.driver.extraClassPath /opt/spark-1.6.0/carbonlib/ > carbondata_2.10-1.0.0-incubating-SNAPSHOT-shade-hadoop2.2.0.jar > spark.driver.extraJavaOptions > -Dcarbon.properties.filepath=/opt/spark-1.6.0/conf/carbon.properties > carbon.kettle.home /opt/spark-1.6.0/carbonlib/ > carbonplugins > > > > > > ------------------ 原始邮件 ------------------ > 发件人: "Ravindra Pesala";<ravi.pes...@gmail.com>; > 发送时间: 2016年12月27日(星期二) 下午4:15 > 收件人: "dev"<dev@carbondata.incubator.apache.org>; > > 主题: Re: Dictionary file is locked for updation > > > > Hi, > > It seems the store path location is taking default location. Did you set > the store location properly? Which spark version you are using? > > Regards, > Ravindra > > On Tue, Dec 27, 2016, 1:38 PM 251469031 <251469...@qq.com> wrote: > > > Hi Kumar, > > > > > > thx to your repley, the full logs is as follows: > > > > > > 16/12/27 12:30:17 INFO locks.HdfsFileLock: Executor task launch worker-0 > > HDFS lock > > path:hdfs://master:9000../carbon.store/default/test_ > table/2e9b7efa-2934-463a-9280-ff50c5129268.lock > > 16/12/27 12:30:17 INFO storage.ShuffleBlockFetcherIterator: Getting 1 > > non-empty blocks out of 1 blocks > > 16/12/27 12:30:17 INFO storage.ShuffleBlockFetcherIterator: Started 1 > > remote fetches in 1 ms > > 16/12/27 12:30:32 ERROR rdd.CarbonGlobalDictionaryGenerateRDD: Executor > > task launch worker-0 > > java.lang.RuntimeException: Dictionary file name is locked for updation. > > Please try after some time > > at scala.sys.package$.error(package.scala:27) > > at > > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:364) > > at > > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate > RDD.compute(CarbonGlobalDictionaryRDD.scala:302) > > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD. > scala:306) > > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) > > at > > org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) > > at org.apache.spark.scheduler.Task.run(Task.scala:89) > > at > > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) > > at > > java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPoolExecutor.java:1142) > > at > > java.util.concurrent.ThreadPoolExecutor$Worker.run( > ThreadPoolExecutor.java:617) > > at java.lang.Thread.run(Thread.java:745) > > > > > > > > as u see, the lock file path > > is:hdfs://master:9000../carbon.store/default/test_ > table/2e9b7efa-2934-463a-9280-ff50c5129268.lock > > > > > > > > > > ------------------ 原始邮件 ------------------ > > 发件人: "Kumar Vishal";<kumarvishal1...@gmail.com>; > > 发送时间: 2016年12月27日(星期二) 下午3:25 > > 收件人: "dev"<dev@carbondata.incubator.apache.org>; > > > > 主题: Re: Dictionary file is locked for updation > > > > > > > > Hi, > > can you please find *"HDFS lock path"* string in executor log and let me > > know the complete log message. > > > > -Regards > > Kumar Vishal > > > > On Tue, Dec 27, 2016 at 12:45 PM, 251469031 <251469...@qq.com> wrote: > > > > > Hi all, > > > > > > > > > when I run the following script: > > > scala> cc.sql(s"load data inpath > > 'hdfs://master:9000/carbondata/sample.csv' > > > into table test_table") > > > > > > > > > it turns out that: > > > WARN 27-12 12:37:58,044 - Lost task 1.3 in stage 2.0 (TID 13, slave1): > > > java.lang.RuntimeException: Dictionary file name is locked for > updation. > > > Please try after some time > > > > > > > > > what I have done are: > > > 1.in carbon.properties, set carbon.lock.type=HDFSLOCK > > > 2.send carbon.properties & spark-defaults.conf to all nodes of the > > clusters > > > > > > > > > if any of you have any idea, looking forward to your replay, thx~ >