Another question, when kylin execute the codes blew ( In SparkBatchCubingJobBuilder2.addLayerCubingSteps() ): ------------------------------------------------ StringUtil.appendWithSeparator(jars, findJar("org.htrace.HTraceConfiguration", null)); // htrace-core.jar StringUtil.appendWithSeparator(jars, findJar("org.apache.htrace.Trace", null)); // htrace-core.jar StringUtil.appendWithSeparator(jars, findJar("org.cloudera.htrace.HTraceConfiguration", null)); // htrace-core.jar ------------------------------------------------ It found three related jars: ------------------------------------------------ /xxx/kylin-deploy/kylin-2.0.0/lib/kylin-job-2.0.0.jar /xxx/hadoop-2.7.3/share/hadoop/common/lib/htrace-core-3.1.0-incubating.jar /xxx/hbase-0.98.8-hadoop2/lib/htrace-core-2.04.jar ------------------------------------------------ And these three jars were add into the "--jars" params when execute spark-submit in step "Build Cube with Spark". Is that correct?
2017-06-22 skyyws 发件人:"skyyws"<sky...@163.com> 发送时间:2017-06-22 17:21 主题:Re: Re: Re: Build sample error with spark on kylin 2.0.0 收件人:"dev"<dev@kylin.apache.org> 抄送: Hi, ShaoFeng, sorry for trouble you again. I fixed the "NoclassDefFoundError"(just replace soft link with copy these jars), and there is another problem: -------------------------------------------------------------------------------------------------------------- 17/06/22 17:04:40 INFO TaskSetManager: Starting task 1.3 in stage 0.0 (TID 7, hadoop759.lt.163.org, partition 1,RACK_LOCAL, 3275 bytes) 17/06/22 17:05:14 WARN TaskSetManager: Lost task 0.3 in stage 0.0 (TID 6, hadoop759.lt.163.org): java.lang.IllegalArgumentException: Failed to find metadata store by url: kylin_metadata_2_0_0@hbase at org.apache.kylin.common.persistence.ResourceStore.createResourceStore(ResourceStore.java:99) at org.apache.kylin.common.persistence.ResourceStore.getStore(ResourceStore.java:110) at org.apache.kylin.cube.CubeDescManager.getStore(CubeDescManager.java:370) at org.apache.kylin.cube.CubeDescManager.reloadAllCubeDesc(CubeDescManager.java:298) at org.apache.kylin.cube.CubeDescManager.<init>(CubeDescManager.java:109) at org.apache.kylin.cube.CubeDescManager.getInstance(CubeDescManager.java:81) at org.apache.kylin.cube.CubeInstance.getDescriptor(CubeInstance.java:114) at org.apache.kylin.cube.CubeSegment.getCubeDesc(CubeSegment.java:119) at org.apache.kylin.cube.CubeSegment.isEnableSharding(CubeSegment.java:477) at org.apache.kylin.cube.kv.RowKeyEncoder.<init>(RowKeyEncoder.java:48) at org.apache.kylin.cube.kv.AbstractRowKeyEncoder.createInstance(AbstractRowKeyEncoder.java:48) at org.apache.kylin.engine.spark.SparkCubingByLayer$2.call(SparkCubingByLayer.java:205) at org.apache.kylin.engine.spark.SparkCubingByLayer$2.call(SparkCubingByLayer.java:193) at org.apache.spark.api.java.JavaPairRDD$$anonfun$pairFunToScalaFun$1.apply(JavaPairRDD.scala:1018) at org.apache.spark.api.java.JavaPairRDD$$anonfun$pairFunToScalaFun$1.apply(JavaPairRDD.scala:1018) at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:191) at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) at org.apache.spark.scheduler.Task.run(Task.scala:89) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) 17/06/22 17:05:14 ERROR TaskSetManager: Task 0 in stage 0.0 failed 4 times; aborting job 17/06/22 17:05:14 INFO YarnScheduler: Cancelling stage 0 17/06/22 17:05:14 INFO YarnScheduler: Stage 0 was cancelled 17/06/22 17:05:14 INFO DAGScheduler: ShuffleMapStage 0 (mapToPair at SparkCubingByLayer.java:193) failed in 265.116 s 17/06/22 17:05:14 INFO DAGScheduler: Job 0 failed: saveAsNewAPIHadoopFile at SparkCubingByLayer.java:288, took 265.278439 s Exception in thread "main" java.lang.RuntimeException: error execute org.apache.kylin.engine.spark.SparkCubingByLayer at org.apache.kylin.common.util.AbstractApplication.execute(AbstractApplication.java:42) at org.apache.kylin.common.util.SparkEntry.main(SparkEntry.java:44) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage 0.0 (TID 6, hadoop759.lt.163.org): java.lang.IllegalArgumentException: Failed to find metadata store by url: kylin_metadata_2_0_0@hbase at org.apache.kylin.common.persistence.ResourceStore.createResourceStore(ResourceStore.java:99) at org.apache.kylin.common.persistence.ResourceStore.getStore(ResourceStore.java:110) at org.apache.kylin.cube.CubeDescManager.getStore(CubeDescManager.java:370) at org.apache.kylin.cube.CubeDescManager.reloadAllCubeDesc(CubeDescManager.java:298) at org.apache.kylin.cube.CubeDescManager.<init>(CubeDescManager.java:109) at org.apache.kylin.cube.CubeDescManager.getInstance(CubeDescManager.java:81) at org.apache.kylin.cube.CubeInstance.getDescriptor(CubeInstance.java:114) at org.apache.kylin.cube.CubeSegment.getCubeDesc(CubeSegment.java:119) at org.apache.kylin.cube.CubeSegment.isEnableSharding(CubeSegment.java:477) at org.apache.kylin.cube.kv.RowKeyEncoder.<init>(RowKeyEncoder.java:48) at org.apache.kylin.cube.kv.AbstractRowKeyEncoder.createInstance(AbstractRowKeyEncoder.java:48) at org.apache.kylin.engine.spark.SparkCubingByLayer$2.call(SparkCubingByLayer.java:205) at org.apache.kylin.engine.spark.SparkCubingByLayer$2.call(SparkCubingByLayer.java:193) at org.apache.spark.api.java.JavaPairRDD$$anonfun$pairFunToScalaFun$1.apply(JavaPairRDD.scala:1018) at org.apache.spark.api.java.JavaPairRDD$$anonfun$pairFunToScalaFun$1.apply(JavaPairRDD.scala:1018) at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:191) at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) at org.apache.spark.scheduler.Task.run(Task.scala:89) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) Driver stacktrace: at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431) at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419) at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418) at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799) at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799) at scala.Option.foreach(Option.scala:236) at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1640) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1599) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1588) at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620) at org.apache.spark.SparkContext.runJob(SparkContext.scala:1832) at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845) at org.apache.spark.SparkContext.runJob(SparkContext.scala:1922) at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply$mcV$sp(PairRDDFunctions.scala:1144) at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:1074) at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:1074) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111) at org.apache.spark.rdd.RDD.withScope(RDD.scala:316) at org.apache.spark.rdd.PairRDDFunctions.saveAsNewAPIHadoopDataset(PairRDDFunctions.scala:1074) at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopFile$2.apply$mcV$sp(PairRDDFunctions.scala:994) at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopFile$2.apply(PairRDDFunctions.scala:985) at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopFile$2.apply(PairRDDFunctions.scala:985) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111) at org.apache.spark.rdd.RDD.withScope(RDD.scala:316) at org.apache.spark.rdd.PairRDDFunctions.saveAsNewAPIHadoopFile(PairRDDFunctions.scala:985) at org.apache.spark.api.java.JavaPairRDD.saveAsNewAPIHadoopFile(JavaPairRDD.scala:800) at org.apache.kylin.engine.spark.SparkCubingByLayer.saveToHDFS(SparkCubingByLayer.java:288) at org.apache.kylin.engine.spark.SparkCubingByLayer.execute(SparkCubingByLayer.java:257) at org.apache.kylin.common.util.AbstractApplication.execute(AbstractApplication.java:37) ... 10 more Caused by: java.lang.IllegalArgumentException: Failed to find metadata store by url: kylin_metadata_2_0_0@hbase at org.apache.kylin.common.persistence.ResourceStore.createResourceStore(ResourceStore.java:99) at org.apache.kylin.common.persistence.ResourceStore.getStore(ResourceStore.java:110) at org.apache.kylin.cube.CubeDescManager.getStore(CubeDescManager.java:370) at org.apache.kylin.cube.CubeDescManager.reloadAllCubeDesc(CubeDescManager.java:298) at org.apache.kylin.cube.CubeDescManager.<init>(CubeDescManager.java:109) at org.apache.kylin.cube.CubeDescManager.getInstance(CubeDescManager.java:81) at org.apache.kylin.cube.CubeInstance.getDescriptor(CubeInstance.java:114) at org.apache.kylin.cube.CubeSegment.getCubeDesc(CubeSegment.java:119) at org.apache.kylin.cube.CubeSegment.isEnableSharding(CubeSegment.java:477) at org.apache.kylin.cube.kv.RowKeyEncoder.<init>(RowKeyEncoder.java:48) at org.apache.kylin.cube.kv.AbstractRowKeyEncoder.createInstance(AbstractRowKeyEncoder.java:48) at org.apache.kylin.engine.spark.SparkCubingByLayer$2.call(SparkCubingByLayer.java:205) at org.apache.kylin.engine.spark.SparkCubingByLayer$2.call(SparkCubingByLayer.java:193) at org.apache.spark.api.java.JavaPairRDD$$anonfun$pairFunToScalaFun$1.apply(JavaPairRDD.scala:1018) at org.apache.spark.api.java.JavaPairRDD$$anonfun$pairFunToScalaFun$1.apply(JavaPairRDD.scala:1018) at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:191) at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) at org.apache.spark.scheduler.Task.run(Task.scala:89) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) ------------------------------------------------------------------------------------------------------------------------------------------------- I'm not sure it's the problem of my kylin or spark? 2017-06-22 skyyws 发件人:ShaoFeng Shi <shaofeng...@apache.org> 发送时间:2017-06-22 12:52 主题:Re: Re: Re: Build sample error with spark on kylin 2.0.0 收件人:"dev"<dev@kylin.apache.org> 抄送: The root cause should be " java.lang.NoClassDefFoundError: org/cloudera/htrace/Trace". Please locate the "htrace-core.jar" in local disk first, and then re-run the spark-submit command with adding this jar in the param "--jars". If it works this time, then you can configure this jar path in kylin.properties: kylin.engine.spark.additional-jars=/path/to/htrace-core.jar Then restart Kylin and resume the job. 2017-06-22 12:02 GMT+08:00 skyyws <sky...@163.com>: > Hi ShaoFeng , no more other error msg before or after this in kylin.log, > but I try to execute cmd by spart-submit directly like this: > -------------------------------------------------------------------- > ./spark-submit --class org.apache.kylin.common.util.SparkEntry --conf > spark.executor.instances=1 --conf spark.yarn.queue=default --conf > spark.history.fs.logDirectory=hdfs://xxx/user/user1/kylin_2_0_0_test/spark-history > > --conf spark.master=yarn --conf spark.executor.memory=1G --conf > spark.eventLog.enabled=true --conf spark.eventLog.dir=hdfs://xxx/ > user/user1/kylin_2_0_0_test/spark-history --conf spark.executor.cores=2 > --conf spark.submit.deployMode=cluster --files > /xxx/hbase-0.98.8-hadoop2/conf/hbase-site.xml > --jars /xxx/kylin-deploy/kylin-2.0.0/lib/kylin-job-2.0.0.jar,/user/ > user1/ext_lib/htrace-core-2.04.jar,/user/user1/ext_lib/ > hbase-client-0.98.8-hadoop2.jar,/user/user1/ext_lib/hbase- > common-0.98.8-hadoop2.jar,/user/user1/ext_lib/hbase- > protocol-0.98.8-hadoop2.jar,/user/user1/ext_lib/metrics- > core-2.2.0.jar,/user/user1/ext_lib/guava-12.0.1.jar, > /xxx/kylin-deploy/kylin-2.0.0/lib/kylin-job-2.0.0.jar -className > org.apache.kylin.engine.spark.SparkCubingByLayer -hiveTable > default.kylin_intermediate_kylin_sales_test_cube_a0cd9950_cddc_4c3b_aaa5_fddf87d1fdaa > > -segmentId a0cd9950-cddc-4c3b-aaa5-fddf87d1fdaa -confPath > /xxx/kylin-deploy/kylin-2.0.0/conf -output hdfs:///user/user1/kylin_2_0_ > 0_test/kylin_metadata_2_0_0/kylin-7a376cb7-7ee7-43fd-95dd- > 79c2c1999f40/kylin_sales_test_cube/cuboid/ -cubename kylin_sales_test_cube > -------------------------------------------------------------------- > Then, I got some other msg from spark, here is the full error msg: > -------------------------------------------------------------------- > 17/06/22 11:35:19 ERROR HBaseConnection: Error when open connection hbase > java.io.IOException: java.lang.reflect.InvocationTargetException > at org.apache.hadoop.hbase.client.HConnectionManager.createConnection( > HConnectionManager.java:413) > at org.apache.hadoop.hbase.client.HConnectionManager.createConnection( > HConnectionManager.java:306) > at org.apache.kylin.storage.hbase.HBaseConnection.get( > HBaseConnection.java:229) > at org.apache.kylin.storage.hbase.HBaseResourceStore.getConnection( > HBaseResourceStore.java:72) > at org.apache.kylin.storage.hbase.HBaseResourceStore. > createHTableIfNeeded(HBaseResourceStore.java:89) > at org.apache.kylin.storage.hbase.HBaseResourceStore.< > init>(HBaseResourceStore.java:85) > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at sun.reflect.NativeConstructorAccessorImpl.newInstance( > NativeConstructorAccessorImpl.java:57) > at sun.reflect.DelegatingConstructorAccessorImpl.newInstance( > DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:526) > at org.apache.kylin.common.persistence.ResourceStore.createResourceStore( > ResourceStore.java:91) > at org.apache.kylin.common.persistence.ResourceStore. > getStore(ResourceStore.java:110) > at org.apache.kylin.cube.CubeManager.getStore(CubeManager.java:820) > at org.apache.kylin.cube.CubeManager.loadAllCubeInstance( > CubeManager.java:740) > at org.apache.kylin.cube.CubeManager.<init>(CubeManager.java:145) > at org.apache.kylin.cube.CubeManager.getInstance(CubeManager.java:109) > at org.apache.kylin.engine.spark.SparkCubingByLayer.execute( > SparkCubingByLayer.java:160) > at org.apache.kylin.common.util.AbstractApplication.execute( > AbstractApplication.java:37) > at org.apache.kylin.common.util.SparkEntry.main(SparkEntry.java:44) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:57) > at sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$ > deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) > at org.apache.spark.deploy.SparkSubmit$.doRunMain$1( > SparkSubmit.scala:181) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > Caused by: java.lang.reflect.InvocationTargetException > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at sun.reflect.NativeConstructorAccessorImpl.newInstance( > NativeConstructorAccessorImpl.java:57) > at sun.reflect.DelegatingConstructorAccessorImpl.newInstance( > DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:526) > at org.apache.hadoop.hbase.client.HConnectionManager.createConnection( > HConnectionManager.java:411) > ... 27 more > Caused by: java.lang.NoClassDefFoundError: org/cloudera/htrace/Trace > at org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.exists( > RecoverableZooKeeper.java:218) > at org.apache.hadoop.hbase.zookeeper.ZKUtil.checkExists(ZKUtil.java:479) > at org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode( > ZKClusterId.java:65) > at org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId( > ZooKeeperRegistry.java:83) > at org.apache.hadoop.hbase.client.HConnectionManager$ > HConnectionImplementation.retrieveClusterId(HConnectionManager.java:839) > at org.apache.hadoop.hbase.client.HConnectionManager$ > HConnectionImplementation.<init>(HConnectionManager.java:642) > ... 32 more > Caused by: java.lang.ClassNotFoundException: org.cloudera.htrace.Trace > at java.net.URLClassLoader$1.run(URLClassLoader.java:366) > at java.net.URLClassLoader$1.run(URLClassLoader.java:355) > at java.security.AccessController.doPrivileged(Native Method) > at java.net.URLClassLoader.findClass(URLClassLoader.java:354) > at java.lang.ClassLoader.loadClass(ClassLoader.java:425) > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) > at java.lang.ClassLoader.loadClass(ClassLoader.java:358) > ... 38 more > 17/06/22 11:35:19 ERROR ResourceStore: Create new store instance failed > java.lang.reflect.InvocationTargetException > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at sun.reflect.NativeConstructorAccessorImpl.newInstance( > NativeConstructorAccessorImpl.java:57) > at sun.reflect.DelegatingConstructorAccessorImpl.newInstance( > DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:526) > at org.apache.kylin.common.persistence.ResourceStore.createResourceStore( > ResourceStore.java:91) > at org.apache.kylin.common.persistence.ResourceStore. > getStore(ResourceStore.java:110) > at org.apache.kylin.cube.CubeManager.getStore(CubeManager.java:820) > at org.apache.kylin.cube.CubeManager.loadAllCubeInstance( > CubeManager.java:740) > at org.apache.kylin.cube.CubeManager.<init>(CubeManager.java:145) > at org.apache.kylin.cube.CubeManager.getInstance(CubeManager.java:109) > at org.apache.kylin.engine.spark.SparkCubingByLayer.execute( > SparkCubingByLayer.java:160) > at org.apache.kylin.common.util.AbstractApplication.execute( > AbstractApplication.java:37) > at org.apache.kylin.common.util.SparkEntry.main(SparkEntry.java:44) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:57) > at sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$ > deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) > at org.apache.spark.deploy.SparkSubmit$.doRunMain$1( > SparkSubmit.scala:181) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > Caused by: java.lang.IllegalArgumentException: File not exist by > 'kylin_metadata_2_0_0@hbase': /xxx/spark-1.6.2-bin-hadoop2. > 7/kylin_metadata_2_0_0@hbase > at org.apache.kylin.common.persistence.FileResourceStore. > <init>(FileResourceStore.java:49) > ... 22 more > 17/06/22 11:35:19 ERROR ResourceStore: Create new store instance failed > java.lang.reflect.InvocationTargetException > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at sun.reflect.NativeConstructorAccessorImpl.newInstance( > NativeConstructorAccessorImpl.java:57) > at sun.reflect.DelegatingConstructorAccessorImpl.newInstance( > DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:526) > at org.apache.kylin.common.persistence.ResourceStore.createResourceStore( > ResourceStore.java:91) > at org.apache.kylin.common.persistence.ResourceStore. > getStore(ResourceStore.java:110) > at org.apache.kylin.cube.CubeManager.getStore(CubeManager.java:820) > at org.apache.kylin.cube.CubeManager.loadAllCubeInstance( > CubeManager.java:740) > at org.apache.kylin.cube.CubeManager.<init>(CubeManager.java:145) > at org.apache.kylin.cube.CubeManager.getInstance(CubeManager.java:109) > at org.apache.kylin.engine.spark.SparkCubingByLayer.execute( > SparkCubingByLayer.java:160) > at org.apache.kylin.common.util.AbstractApplication.execute( > AbstractApplication.java:37) > at org.apache.kylin.common.util.SparkEntry.main(SparkEntry.java:44) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:57) > at sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$ > deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) > at org.apache.spark.deploy.SparkSubmit$.doRunMain$1( > SparkSubmit.scala:181) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > Caused by: org.apache.kylin.common.persistence.StorageException: Error > when open connection hbase > at org.apache.kylin.storage.hbase.HBaseConnection.get( > HBaseConnection.java:242) > at org.apache.kylin.storage.hbase.HBaseResourceStore.getConnection( > HBaseResourceStore.java:72) > at org.apache.kylin.storage.hbase.HBaseResourceStore. > createHTableIfNeeded(HBaseResourceStore.java:89) > at org.apache.kylin.storage.hbase.HBaseResourceStore.< > init>(HBaseResourceStore.java:85) > ... 22 more > Caused by: java.io.IOException: java.lang.reflect. > InvocationTargetException > at org.apache.hadoop.hbase.client.HConnectionManager.createConnection( > HConnectionManager.java:413) > at org.apache.hadoop.hbase.client.HConnectionManager.createConnection( > HConnectionManager.java:306) > at org.apache.kylin.storage.hbase.HBaseConnection.get( > HBaseConnection.java:229) > ... 25 more > Caused by: java.lang.reflect.InvocationTargetException > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at sun.reflect.NativeConstructorAccessorImpl.newInstance( > NativeConstructorAccessorImpl.java:57) > at sun.reflect.DelegatingConstructorAccessorImpl.newInstance( > DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:526) > at org.apache.hadoop.hbase.client.HConnectionManager.createConnection( > HConnectionManager.java:411) > ... 27 more > Caused by: java.lang.NoClassDefFoundError: org/cloudera/htrace/Trace > at org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.exists( > RecoverableZooKeeper.java:218) > at org.apache.hadoop.hbase.zookeeper.ZKUtil.checkExists(ZKUtil.java:479) > at org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode( > ZKClusterId.java:65) > at org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId( > ZooKeeperRegistry.java:83) > at org.apache.hadoop.hbase.client.HConnectionManager$ > HConnectionImplementation.retrieveClusterId(HConnectionManager.java:839) > at org.apache.hadoop.hbase.client.HConnectionManager$ > HConnectionImplementation.<init>(HConnectionManager.java:642) > ... 32 more > Caused by: java.lang.ClassNotFoundException: org.cloudera.htrace.Trace > at java.net.URLClassLoader$1.run(URLClassLoader.java:366) > at java.net.URLClassLoader$1.run(URLClassLoader.java:355) > at java.security.AccessController.doPrivileged(Native Method) > at java.net.URLClassLoader.findClass(URLClassLoader.java:354) > at java.lang.ClassLoader.loadClass(ClassLoader.java:425) > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) > at java.lang.ClassLoader.loadClass(ClassLoader.java:358) > ... 38 more > 17/06/22 11:35:19 INFO ClientCnxn: Opening socket connection to server > 127.0.0.1/127.0.0.1:2181. Will not attempt to authenticate using SASL > (unknown error) > Exception in thread "main" java.lang.RuntimeException: error execute > org.apache.kylin.engine.spark.SparkCubingByLayer > at org.apache.kylin.common.util.AbstractApplication.execute( > AbstractApplication.java:42) > at org.apache.kylin.common.util.SparkEntry.main(SparkEntry.java:44) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:57) > at sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$ > deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) > at org.apache.spark.deploy.SparkSubmit$.doRunMain$1( > SparkSubmit.scala:181) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > Caused by: java.lang.IllegalArgumentException: Failed to find metadata > store by url: kylin_metadata_2_0_0@hbase > at org.apache.kylin.common.persistence.ResourceStore.createResourceStore( > ResourceStore.java:99) > at org.apache.kylin.common.persistence.ResourceStore. > getStore(ResourceStore.java:110) > at org.apache.kylin.cube.CubeManager.getStore(CubeManager.java:820) > at org.apache.kylin.cube.CubeManager.loadAllCubeInstance( > CubeManager.java:740) > at org.apache.kylin.cube.CubeManager.<init>(CubeManager.java:145) > at org.apache.kylin.cube.CubeManager.getInstance(CubeManager.java:109) > at org.apache.kylin.engine.spark.SparkCubingByLayer.execute( > SparkCubingByLayer.java:160) > at org.apache.kylin.common.util.AbstractApplication.execute( > AbstractApplication.java:37) > ... 10 more > 17/06/22 11:35:19 INFO ClientCnxn: Socket connection established to > 127.0.0.1/127.0.0.1:2181, initiating session > 17/06/22 11:35:19 INFO SparkContext: Invoking stop() from shutdown hook > -------------------------------------------------------------------- > Thanks for your attention! > 2017-06-22 > > skyyws > > > > 发件人:ShaoFeng Shi <shaofeng...@apache.org> > 发送时间:2017-06-22 11:42 > 主题:Re: Re: Re: Build sample error with spark on kylin 2.0.0 > 收件人:"dev"<dev@kylin.apache.org> > 抄送: > > hi Sky, glad to see it moves forward. The "Failed to find metadata store by > url: kylin_metadata_2_0 > _0@hbase" is not root cause. Could you check more with the log files, is > there any other error before or after this? > > 2017-06-21 20:43 GMT+08:00 skyyws <sky...@163.com>: > > > Thank you for your suggestion, Shaofeng Shi, I try to use hadoop client > > 2.7.3, it worked. But I met another probelm: > > -------------------------- -------------------------- > > -------------------------- -------------------------- > > 17/06/21 20:20:39 WARN TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0, > > hadoop645.lt.163.org): java.lang.IllegalArgumentException: Failed to > find > > metadata store by url: kylin_metadata_2_0 > > _0@hbase > > at org.apache.kylin.common.persistence.ResourceStore. > > createResourceStore(ResourceStore.java:99) > > at org.apache.kylin.common.persistence.ResourceStore. > > getStore(ResourceStore.java:110) > > at org.apache.kylin.cube.CubeDescManager.getStore( > > CubeDescManager.java:370) > > at org.apache.kylin.cube.CubeDescManager.reloadAllCubeDesc( > > CubeDescManager.java:298) > > at org.apache.kylin.cube.CubeDescManager.<init>( > > CubeDescManager.java:109) > > at org.apache.kylin.cube.CubeDescManager.getInstance( > > CubeDescManager.java:81) > > at org.apache.kylin.cube.CubeInstance.getDescriptor( > > CubeInstance.java:114) > > at org.apache.kylin.cube.CubeSegment.getCubeDesc( > > CubeSegment.java:119) > > at org.apache.kylin.cube.kv.RowKeyEncoder.<init>( > > RowKeyEncoder.java:50) > > at org.apache.kylin.cube.kv.AbstractRowKeyEncoder. > createInstance( > > AbstractRowKeyEncoder.java:48) > > at org.apache.kylin.engine.spark.SparkCubingByLayer$2.call( > > SparkCubingByLayer.java:205) > > at org.apache.kylin.engine.spark.SparkCubingByLayer$2.call( > > SparkCubingByLayer.java:193) > > at org.apache.spark.api.java.JavaPairRDD$$anonfun$ > > pairFunToScalaFun$1.apply(JavaPairRDD.scala:1018) > > at org.apache.spark.api.java.JavaPairRDD$$anonfun$ > > pairFunToScalaFun$1.apply(JavaPairRDD.scala:1018) > > at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) > > at org.apache.spark.util.collection.ExternalSorter. > > insertAll(ExternalSorter.scala:191) > > at org.apache.spark.shuffle.sort.SortShuffleWriter.write( > > SortShuffleWriter.scala:64) > > at org.apache.spark.scheduler.ShuffleMapTask.runTask( > > ShuffleMapTask.scala:73) > > at org.apache.spark.scheduler.ShuffleMapTask.runTask( > > ShuffleMapTask.scala:41) > > at org.apache.spark.scheduler.Task.run(Task.scala:89) > > at org.apache.spark.executor.Executor$TaskRunner.run( > > Executor.scala:227) > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > > ThreadPoolExecutor.java:1145) > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > > ThreadPoolExecutor.java:615) > > at java.lang.Thread.run(Thread.java:745) > > 17/06/21 20:20:39 INFO TaskSetManager: Starting task 0.1 in stage 0.0 > (TID > > 2, hadoop645.lt.163.org, partition 0,RACK_LOCAL, 3276 bytes) > > 17/06/21 20:21:14 WARN TaskSetManager: Lost task 1.0 in stage 0.0 (TID 1, > > hadoop645.lt.163.org): java.lang.IllegalArgumentException: Failed to > find > > metadata store by url: kylin_metadata_2_0 > > _0@hbase > > at org.apache.kylin.common.persistence.ResourceStore. > > createResourceStore(ResourceStore.java:99) > > at org.apache.kylin.common.persistence.ResourceStore. > > getStore(ResourceStore.java:110) > > at org.apache.kylin.cube.CubeDescManager.getStore( > > CubeDescManager.java:370) > > at org.apache.kylin.cube.CubeDescManager.reloadAllCubeDesc( > > CubeDescManager.java:298) > > at org.apache.kylin.cube.CubeDescManager.<init>( > > CubeDescManager.java:109) > > at org.apache.kylin.cube.CubeDescManager.getInstance( > > CubeDescManager.java:81) > > at org.apache.kylin.cube.CubeInstance.getDescriptor( > > CubeInstance.java:114) > > at org.apache.kylin.cube.CubeSegment.getCubeDesc( > > CubeSegment.java:119) > > at org.apache.kylin.cube.kv.RowKeyEncoder.<init>( > > RowKeyEncoder.java:50) > > at org.apache.kylin.cube.kv.AbstractRowKeyEncoder. > createInstance( > > AbstractRowKeyEncoder.java:48) > > at org.apache.kylin.engine.spark.SparkCubingByLayer$2.call( > > SparkCubingByLayer.java:205) > > at org.apache.kylin.engine.spark.SparkCubingByLayer$2.call( > > SparkCubingByLayer.java:193) > > at org.apache.spark.api.java.JavaPairRDD$$anonfun$ > > pairFunToScalaFun$1.apply(JavaPairRDD.scala:1018) > > at org.apache.spark.api.java.JavaPairRDD$$anonfun$ > > pairFunToScalaFun$1.apply(JavaPairRDD.scala:1018) > > at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) > > at org.apache.spark.util.collection.ExternalSorter. > > insertAll(ExternalSorter.scala:191) > > at org.apache.spark.shuffle.sort.SortShuffleWriter.write( > > SortShuffleWriter.scala:64) > > at org.apache.spark.scheduler.ShuffleMapTask.runTask( > > ShuffleMapTask.scala:73) > > at org.apache.spark.scheduler.ShuffleMapTask.runTask( > > ShuffleMapTask.scala:41) > > at org.apache.spark.scheduler.Task.run(Task.scala:89) > > at org.apache.spark.executor.Executor$TaskRunner.run( > > Executor.scala:227) > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > > ThreadPoolExecutor.java:1145) > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > > ThreadPoolExecutor.java:615) > > at java.lang.Thread.run(Thread.java:745) > > -------------------------- -------------------------- > > -------------------------- -------------------------- > > But I can use the kylin in-build spark-shell to get data from hive and > > hbase successfully, just like this: > > -------------------------- -------------------------- > > -------------------------- -------------------------- > > sqlContext.sql("show tables").take(1) > > -------------------------- -------------------------- > > -------------------------- -------------------------- > > import org.apache.spark._ > > import org.apache.spark.rdd.NewHadoopRDD > > import org.apache.hadoop.fs.Path > > import org.apache.hadoop.hbase.util.Bytes > > import org.apache.hadoop.hbase.HColumnDescriptor > > import org.apache.hadoop.hbase.{HBaseConfiguration, HTableDescriptor} > > import org.apache.hadoop.hbase.client.{HBaseAdmin, Put, HTable, Result} > > import org.apache.hadoop.hbase.mapreduce.TableInputFormat > > import org.apache.hadoop.hbase.io.ImmutableBytesWritable > > val conf = HBaseConfiguration.create() > > conf.set("hbase.zookeeper.quorum", "localhost") > > conf.set(TableInputFormat.INPUT_TABLE, "test_table") > > val hBaseRDD = sc.newAPIHadoopRDD(conf, classOf[TableInputFormat], > > classOf[org.apache.hadoop.hbase.io.ImmutableBytesWritable], > > classOf[org.apache.hadoop.hbase.client.Result]) > > > > val res = hBaseRDD.take(1) > > val rs = res(0)._2 > > val kv = rs.raw > > for(keyvalue <- kv) println("rowkey:"+ new String(keyvalue.getRow)+ " > > cf:"+new String(keyvalue.getFamily()) + " column:" + new > > String(keyvalue.getQualifier) + " " + "value:"+new > > String(keyvalue.getValue())) > > -------------------------- -------------------------- > > -------------------------- -------------------------- > > By the way, I've already put hive-site.xml and hbase-site.xml into the > > HADOOP_CONF_DIR and $SPARK_HOME/conf(which is acually > > $KYLIN_HOME/spark/conf), and I also set spark.driver.extraClassPath in > > spark-defaults.conf to attachs some related jars(hbase-client.jar, > > hbase-common.jar and so on). > > I don't know why, anyone could give me some advice? > > 2017-06-21 > > > > skyyws > > > > > > > > 发件人:ShaoFeng Shi <shaofeng...@apache.org> > > 发送时间:2017-06-20 15:13 > > 主题:Re: Re: Build sample error with spark on kylin 2.0.0 > > 收件人:"dev"<dev@kylin.apache.org> > > 抄送: > > > > Or you can check whether there is old hadoop jars on your cluster, > > according to https://issues.apache.org/jira/browse/HADOOP-11064 > > > > > > 2017-06-20 9:33 GMT+08:00 skyyws <sky...@163.com>: > > > > > No, I deploy kylin on linux, this is my machine info: > > > -------------------------- > > > 3.2.0-4-amd64 #1 SMP Debian 3.2.82-1 x86_64 GNU/Linux > > > ------------------------- > > > > > > 2017-06-20 > > > > > > skyyws > > > > > > > > > > > > 发件人:ShaoFeng Shi <shaofeng...@apache.org> > > > 发送时间:2017-06-20 00:10 > > > 主题:Re: Build sample error with spark on kylin 2.0.0 > > > 收件人:"dev"<dev@kylin.apache.org> > > > 抄送: > > > > > > Are you running Kylin on windows? If yes, check: > > > https://stackoverflow.com/questions/33211599/hadoop- > > > error-on-windows-java-lang-unsatisfiedlinkerror > > > > > > 2017-06-19 21:55 GMT+08:00 skyyws <sky...@163.com>: > > > > > > > Hi all, > > > > I met an error when using spark engine build kylin sample on step > > "Build > > > > Cube with Spark", here is the exception log: > > > > ------------------------------------------------------------ > > > > ----------------------------- > > > > Exception in thread "main" java.lang.UnsatisfiedLinkError: > > > > org.apache.hadoop.util.NativeCrc32.nativeComputeChunkedSumsByteAr > > > > ray(II[BI[BIILjava/lang/String;JZ)V > > > > at org.apache.hadoop.util.NativeCrc32. > > > > nativeComputeChunkedSumsByteArray(Native Method) > > > > at org.apache.hadoop.util.NativeCrc32. > > > > calculateChunkedSumsByteArray(NativeCrc32.java:86) > > > > at org.apache.hadoop.util.DataChecksum.calculateChunkedSums( > > > > DataChecksum.java:430) > > > > at org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunks( > > > > FSOutputSummer.java:202) > > > > at org.apache.hadoop.fs.FSOutputSummer.write1( > > > > FSOutputSummer.java:124) > > > > at org.apache.hadoop.fs.FSOutputSummer.write( > > > > FSOutputSummer.java:110) > > > > at org.apache.hadoop.fs.FSDataOutputStream$ > > PositionCache.write( > > > > FSDataOutputStream.java:58) > > > > at java.io.DataOutputStream.write(DataOutputStream.java:107) > > > > at org.apache.hadoop.io.IOUtils.copyBytes(IOUtils.java:80) > > > > at org.apache.hadoop.io.IOUtils.copyBytes(IOUtils.java:52) > > > > at org.apache.hadoop.io.IOUtils.copyBytes(IOUtils.java:112) > > > > at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:366) > > > > at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:338) > > > > at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:289) > > > > at org.apache.spark.deploy.yarn.Client.copyFileToRemote( > > > > Client.scala:317) > > > > at org.apache.spark.deploy.yarn.Client.org$apache$spark$ > > > > deploy$yarn$Client$$distribute$1(Client.scala:407) > > > > at org.apache.spark.deploy.yarn.Client$$anonfun$ > > > > prepareLocalResources$5.apply(Client.scala:446) > > > > at org.apache.spark.deploy.yarn.Client$$anonfun$ > > > > prepareLocalResources$5.apply(Client.scala:444) > > > > at scala.collection.immutable.List.foreach(List.scala:318) > > > > at org.apache.spark.deploy.yarn. > Client.prepareLocalResources( > > > > Client.scala:444) > > > > at org.apache.spark.deploy.yarn.Client. > > > > createContainerLaunchContext(Client.scala:727) > > > > at org.apache.spark.deploy.yarn.Client.submitApplication( > > > > Client.scala:142) > > > > at org.apache.spark.scheduler.cluster. > > > YarnClientSchedulerBackend. > > > > start(YarnClientSchedulerBackend.scala:57) > > > > at org.apache.spark.scheduler.TaskSchedulerImpl.start( > > > > TaskSchedulerImpl.scala:144) > > > > at org.apache.spark.SparkContext. > > <init>(SparkContext.scala:530) > > > > at org.apache.spark.api.java.JavaSparkContext.<init>( > > > > JavaSparkContext.scala:59) > > > > at org.apache.kylin.engine.spark.SparkCubingByLayer.execute( > > > > SparkCubingByLayer.java:150) > > > > at org.apache.kylin.common.util.AbstractApplication.execute( > > > > AbstractApplication.java:37) > > > > at org.apache.kylin.common.util.SparkEntry.main(SparkEntry. > > > > java:44) > > > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method) > > > > at sun.reflect.NativeMethodAccessorImpl.invoke( > > > > NativeMethodAccessorImpl.java:57) > > > > at sun.reflect.DelegatingMethodAccessorImpl.invoke( > > > > DelegatingMethodAccessorImpl.java:43) > > > > at java.lang.reflect.Method.invoke(Method.java:606) > > > > at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$ > > > > deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) > > > > at org.apache.spark.deploy.SparkSubmit$.doRunMain$1( > > > > SparkSubmit.scala:181) > > > > at org.apache.spark.deploy.SparkSubmit$.submit( > > > > SparkSubmit.scala:206) > > > > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit. > > > > scala:121) > > > > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit. > scala) > > > > 17/06/19 21:22:06 INFO storage.DiskBlockManager: Shutdown hook called > > > > 17/06/19 21:22:06 INFO util.ShutdownHookManager: Shutdown hook called > > > > 17/06/19 21:22:06 INFO util.ShutdownHookManager: Deleting directory > > > > /tmp/spark-0d1d3709-86cd-446c-b728-5070f168de28 > > > > 17/06/19 21:22:06 INFO util.ShutdownHookManager: Deleting directory > > > > /tmp/spark-0d1d3709-86cd-446c-b728-5070f168de28/httpd- > > > > 9bcb9a5d-569f-4f28-ad89-038a9020eda8 > > > > 17/06/19 21:22:06 INFO util.ShutdownHookManager: Deleting directory > > > > /tmp/spark-0d1d3709-86cd-446c-b728-5070f168de28/userFiles- > > > > 2e9ff265-3d37-40e0-8894-6fd4d1a3ad8b > > > > > > > > at org.apache.kylin.common.util.CliCommandExecutor.execute( > > > > CliCommandExecutor.java:92) > > > > at org.apache.kylin.engine.spark.SparkExecutable.doWork( > > > > SparkExecutable.java:124) > > > > at org.apache.kylin.job.execution.AbstractExecutable. > > > > execute(AbstractExecutable.java:124) > > > > at org.apache.kylin.job.execution.DefaultChainedExecutable. > > > doWork( > > > > DefaultChainedExecutable.java:64) > > > > at org.apache.kylin.job.execution.AbstractExecutable. > > > > execute(AbstractExecutable.java:124) > > > > at org.apache.kylin.job.impl.threadpool.DefaultScheduler$ > > > > JobRunner.run(DefaultScheduler.java:142) > > > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > > > > ThreadPoolExecutor.java:1145) > > > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > > > > ThreadPoolExecutor.java:615) > > > > at java.lang.Thread.run(Thread.java:745) > > > > > > > > ------------------------------------------------------------ > > > > ----------------------------- > > > > I can use the kylin in-build spark-shell to do some operations like: > > > > ------------------------------------------------------------ > > > > ----------------------------- > > > > var textFile = sc.textFile("hdfs://xxxx/xxxx/README.md") > > > > textFile.count() > > > > textFile.first() > > > > textFile.filter(line => line.contains("hello")).count() > > > > ------------------------------------------------------------ > > > > ----------------------------- > > > > Here is the env info: > > > > kylin version is 2.0.0 > > > > hadoop version is 2.7.* > > > > spark version is 1.6.* > > > > ------------------------------------------------------------ > > > > ----------------------------- > > > > Anyone can help me?THX > > > > > > > > > > > > 2017-06-19 > > > > skyyws > > > > > > > > > > > > > > > -- > > > Best regards, > > > > > > Shaofeng Shi 史少锋 > > > > > > > > > > > -- > > Best regards, > > > > Shaofeng Shi 史少锋 > > > > > > -- > Best regards, > > Shaofeng Shi 史少锋 > -- Best regards, Shaofeng Shi 史少锋