Hi ShaoFeng , no more other error msg before or after this in kylin.log, but I try to execute cmd by spart-submit directly like this: -------------------------------------------------------------------- ./spark-submit --class org.apache.kylin.common.util.SparkEntry --conf spark.executor.instances=1 --conf spark.yarn.queue=default --conf spark.history.fs.logDirectory=hdfs://xxx/user/user1/kylin_2_0_0_test/spark-history --conf spark.master=yarn --conf spark.executor.memory=1G --conf spark.eventLog.enabled=true --conf spark.eventLog.dir=hdfs://xxx/user/user1/kylin_2_0_0_test/spark-history --conf spark.executor.cores=2 --conf spark.submit.deployMode=cluster --files /xxx/hbase-0.98.8-hadoop2/conf/hbase-site.xml --jars /xxx/kylin-deploy/kylin-2.0.0/lib/kylin-job-2.0.0.jar,/user/user1/ext_lib/htrace-core-2.04.jar,/user/user1/ext_lib/hbase-client-0.98.8-hadoop2.jar,/user/user1/ext_lib/hbase-common-0.98.8-hadoop2.jar,/user/user1/ext_lib/hbase-protocol-0.98.8-hadoop2.jar,/user/user1/ext_lib/metrics-core-2.2.0.jar,/user/user1/ext_lib/guava-12.0.1.jar, /xxx/kylin-deploy/kylin-2.0.0/lib/kylin-job-2.0.0.jar -className org.apache.kylin.engine.spark.SparkCubingByLayer -hiveTable default.kylin_intermediate_kylin_sales_test_cube_a0cd9950_cddc_4c3b_aaa5_fddf87d1fdaa -segmentId a0cd9950-cddc-4c3b-aaa5-fddf87d1fdaa -confPath /xxx/kylin-deploy/kylin-2.0.0/conf -output hdfs:///user/user1/kylin_2_0_0_test/kylin_metadata_2_0_0/kylin-7a376cb7-7ee7-43fd-95dd-79c2c1999f40/kylin_sales_test_cube/cuboid/ -cubename kylin_sales_test_cube -------------------------------------------------------------------- Then, I got some other msg from spark, here is the full error msg: -------------------------------------------------------------------- 17/06/22 11:35:19 ERROR HBaseConnection: Error when open connection hbase java.io.IOException: java.lang.reflect.InvocationTargetException at org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:413) at org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:306) at org.apache.kylin.storage.hbase.HBaseConnection.get(HBaseConnection.java:229) at org.apache.kylin.storage.hbase.HBaseResourceStore.getConnection(HBaseResourceStore.java:72) at org.apache.kylin.storage.hbase.HBaseResourceStore.createHTableIfNeeded(HBaseResourceStore.java:89) at org.apache.kylin.storage.hbase.HBaseResourceStore.<init>(HBaseResourceStore.java:85) at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:526) at org.apache.kylin.common.persistence.ResourceStore.createResourceStore(ResourceStore.java:91) at org.apache.kylin.common.persistence.ResourceStore.getStore(ResourceStore.java:110) at org.apache.kylin.cube.CubeManager.getStore(CubeManager.java:820) at org.apache.kylin.cube.CubeManager.loadAllCubeInstance(CubeManager.java:740) at org.apache.kylin.cube.CubeManager.<init>(CubeManager.java:145) at org.apache.kylin.cube.CubeManager.getInstance(CubeManager.java:109) at org.apache.kylin.engine.spark.SparkCubingByLayer.execute(SparkCubingByLayer.java:160) at org.apache.kylin.common.util.AbstractApplication.execute(AbstractApplication.java:37) at org.apache.kylin.common.util.SparkEntry.main(SparkEntry.java:44) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) Caused by: java.lang.reflect.InvocationTargetException at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:526) at org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:411) ... 27 more Caused by: java.lang.NoClassDefFoundError: org/cloudera/htrace/Trace at org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.exists(RecoverableZooKeeper.java:218) at org.apache.hadoop.hbase.zookeeper.ZKUtil.checkExists(ZKUtil.java:479) at org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:65) at org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:83) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.retrieveClusterId(HConnectionManager.java:839) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.<init>(HConnectionManager.java:642) ... 32 more Caused by: java.lang.ClassNotFoundException: org.cloudera.htrace.Trace at java.net.URLClassLoader$1.run(URLClassLoader.java:366) at java.net.URLClassLoader$1.run(URLClassLoader.java:355) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:354) at java.lang.ClassLoader.loadClass(ClassLoader.java:425) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) at java.lang.ClassLoader.loadClass(ClassLoader.java:358) ... 38 more 17/06/22 11:35:19 ERROR ResourceStore: Create new store instance failed java.lang.reflect.InvocationTargetException at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:526) at org.apache.kylin.common.persistence.ResourceStore.createResourceStore(ResourceStore.java:91) at org.apache.kylin.common.persistence.ResourceStore.getStore(ResourceStore.java:110) at org.apache.kylin.cube.CubeManager.getStore(CubeManager.java:820) at org.apache.kylin.cube.CubeManager.loadAllCubeInstance(CubeManager.java:740) at org.apache.kylin.cube.CubeManager.<init>(CubeManager.java:145) at org.apache.kylin.cube.CubeManager.getInstance(CubeManager.java:109) at org.apache.kylin.engine.spark.SparkCubingByLayer.execute(SparkCubingByLayer.java:160) at org.apache.kylin.common.util.AbstractApplication.execute(AbstractApplication.java:37) at org.apache.kylin.common.util.SparkEntry.main(SparkEntry.java:44) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) Caused by: java.lang.IllegalArgumentException: File not exist by 'kylin_metadata_2_0_0@hbase': /xxx/spark-1.6.2-bin-hadoop2.7/kylin_metadata_2_0_0@hbase at org.apache.kylin.common.persistence.FileResourceStore.<init>(FileResourceStore.java:49) ... 22 more 17/06/22 11:35:19 ERROR ResourceStore: Create new store instance failed java.lang.reflect.InvocationTargetException at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:526) at org.apache.kylin.common.persistence.ResourceStore.createResourceStore(ResourceStore.java:91) at org.apache.kylin.common.persistence.ResourceStore.getStore(ResourceStore.java:110) at org.apache.kylin.cube.CubeManager.getStore(CubeManager.java:820) at org.apache.kylin.cube.CubeManager.loadAllCubeInstance(CubeManager.java:740) at org.apache.kylin.cube.CubeManager.<init>(CubeManager.java:145) at org.apache.kylin.cube.CubeManager.getInstance(CubeManager.java:109) at org.apache.kylin.engine.spark.SparkCubingByLayer.execute(SparkCubingByLayer.java:160) at org.apache.kylin.common.util.AbstractApplication.execute(AbstractApplication.java:37) at org.apache.kylin.common.util.SparkEntry.main(SparkEntry.java:44) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) Caused by: org.apache.kylin.common.persistence.StorageException: Error when open connection hbase at org.apache.kylin.storage.hbase.HBaseConnection.get(HBaseConnection.java:242) at org.apache.kylin.storage.hbase.HBaseResourceStore.getConnection(HBaseResourceStore.java:72) at org.apache.kylin.storage.hbase.HBaseResourceStore.createHTableIfNeeded(HBaseResourceStore.java:89) at org.apache.kylin.storage.hbase.HBaseResourceStore.<init>(HBaseResourceStore.java:85) ... 22 more Caused by: java.io.IOException: java.lang.reflect.InvocationTargetException at org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:413) at org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:306) at org.apache.kylin.storage.hbase.HBaseConnection.get(HBaseConnection.java:229) ... 25 more Caused by: java.lang.reflect.InvocationTargetException at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:526) at org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:411) ... 27 more Caused by: java.lang.NoClassDefFoundError: org/cloudera/htrace/Trace at org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.exists(RecoverableZooKeeper.java:218) at org.apache.hadoop.hbase.zookeeper.ZKUtil.checkExists(ZKUtil.java:479) at org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:65) at org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:83) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.retrieveClusterId(HConnectionManager.java:839) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.<init>(HConnectionManager.java:642) ... 32 more Caused by: java.lang.ClassNotFoundException: org.cloudera.htrace.Trace at java.net.URLClassLoader$1.run(URLClassLoader.java:366) at java.net.URLClassLoader$1.run(URLClassLoader.java:355) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:354) at java.lang.ClassLoader.loadClass(ClassLoader.java:425) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) at java.lang.ClassLoader.loadClass(ClassLoader.java:358) ... 38 more 17/06/22 11:35:19 INFO ClientCnxn: Opening socket connection to server 127.0.0.1/127.0.0.1:2181. Will not attempt to authenticate using SASL (unknown error) Exception in thread "main" java.lang.RuntimeException: error execute org.apache.kylin.engine.spark.SparkCubingByLayer at org.apache.kylin.common.util.AbstractApplication.execute(AbstractApplication.java:42) at org.apache.kylin.common.util.SparkEntry.main(SparkEntry.java:44) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) Caused by: java.lang.IllegalArgumentException: Failed to find metadata store by url: kylin_metadata_2_0_0@hbase at org.apache.kylin.common.persistence.ResourceStore.createResourceStore(ResourceStore.java:99) at org.apache.kylin.common.persistence.ResourceStore.getStore(ResourceStore.java:110) at org.apache.kylin.cube.CubeManager.getStore(CubeManager.java:820) at org.apache.kylin.cube.CubeManager.loadAllCubeInstance(CubeManager.java:740) at org.apache.kylin.cube.CubeManager.<init>(CubeManager.java:145) at org.apache.kylin.cube.CubeManager.getInstance(CubeManager.java:109) at org.apache.kylin.engine.spark.SparkCubingByLayer.execute(SparkCubingByLayer.java:160) at org.apache.kylin.common.util.AbstractApplication.execute(AbstractApplication.java:37) ... 10 more 17/06/22 11:35:19 INFO ClientCnxn: Socket connection established to 127.0.0.1/127.0.0.1:2181, initiating session 17/06/22 11:35:19 INFO SparkContext: Invoking stop() from shutdown hook -------------------------------------------------------------------- Thanks for your attention! 2017-06-22
skyyws 发件人:ShaoFeng Shi <[email protected]> 发送时间:2017-06-22 11:42 主题:Re: Re: Re: Build sample error with spark on kylin 2.0.0 收件人:"dev"<[email protected]> 抄送: hi Sky, glad to see it moves forward. The "Failed to find metadata store by url: kylin_metadata_2_0 _0@hbase" is not root cause. Could you check more with the log files, is there any other error before or after this? 2017-06-21 20:43 GMT+08:00 skyyws <[email protected]>: > Thank you for your suggestion, Shaofeng Shi, I try to use hadoop client > 2.7.3, it worked. But I met another probelm: > -------------------------- -------------------------- > -------------------------- -------------------------- > 17/06/21 20:20:39 WARN TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0, > hadoop645.lt.163.org): java.lang.IllegalArgumentException: Failed to find > metadata store by url: kylin_metadata_2_0 > _0@hbase > at org.apache.kylin.common.persistence.ResourceStore. > createResourceStore(ResourceStore.java:99) > at org.apache.kylin.common.persistence.ResourceStore. > getStore(ResourceStore.java:110) > at org.apache.kylin.cube.CubeDescManager.getStore( > CubeDescManager.java:370) > at org.apache.kylin.cube.CubeDescManager.reloadAllCubeDesc( > CubeDescManager.java:298) > at org.apache.kylin.cube.CubeDescManager.<init>( > CubeDescManager.java:109) > at org.apache.kylin.cube.CubeDescManager.getInstance( > CubeDescManager.java:81) > at org.apache.kylin.cube.CubeInstance.getDescriptor( > CubeInstance.java:114) > at org.apache.kylin.cube.CubeSegment.getCubeDesc( > CubeSegment.java:119) > at org.apache.kylin.cube.kv.RowKeyEncoder.<init>( > RowKeyEncoder.java:50) > at org.apache.kylin.cube.kv.AbstractRowKeyEncoder.createInstance( > AbstractRowKeyEncoder.java:48) > at org.apache.kylin.engine.spark.SparkCubingByLayer$2.call( > SparkCubingByLayer.java:205) > at org.apache.kylin.engine.spark.SparkCubingByLayer$2.call( > SparkCubingByLayer.java:193) > at org.apache.spark.api.java.JavaPairRDD$$anonfun$ > pairFunToScalaFun$1.apply(JavaPairRDD.scala:1018) > at org.apache.spark.api.java.JavaPairRDD$$anonfun$ > pairFunToScalaFun$1.apply(JavaPairRDD.scala:1018) > at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) > at org.apache.spark.util.collection.ExternalSorter. > insertAll(ExternalSorter.scala:191) > at org.apache.spark.shuffle.sort.SortShuffleWriter.write( > SortShuffleWriter.scala:64) > at org.apache.spark.scheduler.ShuffleMapTask.runTask( > ShuffleMapTask.scala:73) > at org.apache.spark.scheduler.ShuffleMapTask.runTask( > ShuffleMapTask.scala:41) > at org.apache.spark.scheduler.Task.run(Task.scala:89) > at org.apache.spark.executor.Executor$TaskRunner.run( > Executor.scala:227) > at java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPoolExecutor.java:1145) > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > 17/06/21 20:20:39 INFO TaskSetManager: Starting task 0.1 in stage 0.0 (TID > 2, hadoop645.lt.163.org, partition 0,RACK_LOCAL, 3276 bytes) > 17/06/21 20:21:14 WARN TaskSetManager: Lost task 1.0 in stage 0.0 (TID 1, > hadoop645.lt.163.org): java.lang.IllegalArgumentException: Failed to find > metadata store by url: kylin_metadata_2_0 > _0@hbase > at org.apache.kylin.common.persistence.ResourceStore. > createResourceStore(ResourceStore.java:99) > at org.apache.kylin.common.persistence.ResourceStore. > getStore(ResourceStore.java:110) > at org.apache.kylin.cube.CubeDescManager.getStore( > CubeDescManager.java:370) > at org.apache.kylin.cube.CubeDescManager.reloadAllCubeDesc( > CubeDescManager.java:298) > at org.apache.kylin.cube.CubeDescManager.<init>( > CubeDescManager.java:109) > at org.apache.kylin.cube.CubeDescManager.getInstance( > CubeDescManager.java:81) > at org.apache.kylin.cube.CubeInstance.getDescriptor( > CubeInstance.java:114) > at org.apache.kylin.cube.CubeSegment.getCubeDesc( > CubeSegment.java:119) > at org.apache.kylin.cube.kv.RowKeyEncoder.<init>( > RowKeyEncoder.java:50) > at org.apache.kylin.cube.kv.AbstractRowKeyEncoder.createInstance( > AbstractRowKeyEncoder.java:48) > at org.apache.kylin.engine.spark.SparkCubingByLayer$2.call( > SparkCubingByLayer.java:205) > at org.apache.kylin.engine.spark.SparkCubingByLayer$2.call( > SparkCubingByLayer.java:193) > at org.apache.spark.api.java.JavaPairRDD$$anonfun$ > pairFunToScalaFun$1.apply(JavaPairRDD.scala:1018) > at org.apache.spark.api.java.JavaPairRDD$$anonfun$ > pairFunToScalaFun$1.apply(JavaPairRDD.scala:1018) > at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) > at org.apache.spark.util.collection.ExternalSorter. > insertAll(ExternalSorter.scala:191) > at org.apache.spark.shuffle.sort.SortShuffleWriter.write( > SortShuffleWriter.scala:64) > at org.apache.spark.scheduler.ShuffleMapTask.runTask( > ShuffleMapTask.scala:73) > at org.apache.spark.scheduler.ShuffleMapTask.runTask( > ShuffleMapTask.scala:41) > at org.apache.spark.scheduler.Task.run(Task.scala:89) > at org.apache.spark.executor.Executor$TaskRunner.run( > Executor.scala:227) > at java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPoolExecutor.java:1145) > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > -------------------------- -------------------------- > -------------------------- -------------------------- > But I can use the kylin in-build spark-shell to get data from hive and > hbase successfully, just like this: > -------------------------- -------------------------- > -------------------------- -------------------------- > sqlContext.sql("show tables").take(1) > -------------------------- -------------------------- > -------------------------- -------------------------- > import org.apache.spark._ > import org.apache.spark.rdd.NewHadoopRDD > import org.apache.hadoop.fs.Path > import org.apache.hadoop.hbase.util.Bytes > import org.apache.hadoop.hbase.HColumnDescriptor > import org.apache.hadoop.hbase.{HBaseConfiguration, HTableDescriptor} > import org.apache.hadoop.hbase.client.{HBaseAdmin, Put, HTable, Result} > import org.apache.hadoop.hbase.mapreduce.TableInputFormat > import org.apache.hadoop.hbase.io.ImmutableBytesWritable > val conf = HBaseConfiguration.create() > conf.set("hbase.zookeeper.quorum", "localhost") > conf.set(TableInputFormat.INPUT_TABLE, "test_table") > val hBaseRDD = sc.newAPIHadoopRDD(conf, classOf[TableInputFormat], > classOf[org.apache.hadoop.hbase.io.ImmutableBytesWritable], > classOf[org.apache.hadoop.hbase.client.Result]) > > val res = hBaseRDD.take(1) > val rs = res(0)._2 > val kv = rs.raw > for(keyvalue <- kv) println("rowkey:"+ new String(keyvalue.getRow)+ " > cf:"+new String(keyvalue.getFamily()) + " column:" + new > String(keyvalue.getQualifier) + " " + "value:"+new > String(keyvalue.getValue())) > -------------------------- -------------------------- > -------------------------- -------------------------- > By the way, I've already put hive-site.xml and hbase-site.xml into the > HADOOP_CONF_DIR and $SPARK_HOME/conf(which is acually > $KYLIN_HOME/spark/conf), and I also set spark.driver.extraClassPath in > spark-defaults.conf to attachs some related jars(hbase-client.jar, > hbase-common.jar and so on). > I don't know why, anyone could give me some advice? > 2017-06-21 > > skyyws > > > > 发件人:ShaoFeng Shi <[email protected]> > 发送时间:2017-06-20 15:13 > 主题:Re: Re: Build sample error with spark on kylin 2.0.0 > 收件人:"dev"<[email protected]> > 抄送: > > Or you can check whether there is old hadoop jars on your cluster, > according to https://issues.apache.org/jira/browse/HADOOP-11064 > > > 2017-06-20 9:33 GMT+08:00 skyyws <[email protected]>: > > > No, I deploy kylin on linux, this is my machine info: > > -------------------------- > > 3.2.0-4-amd64 #1 SMP Debian 3.2.82-1 x86_64 GNU/Linux > > ------------------------- > > > > 2017-06-20 > > > > skyyws > > > > > > > > 发件人:ShaoFeng Shi <[email protected]> > > 发送时间:2017-06-20 00:10 > > 主题:Re: Build sample error with spark on kylin 2.0.0 > > 收件人:"dev"<[email protected]> > > 抄送: > > > > Are you running Kylin on windows? If yes, check: > > https://stackoverflow.com/questions/33211599/hadoop- > > error-on-windows-java-lang-unsatisfiedlinkerror > > > > 2017-06-19 21:55 GMT+08:00 skyyws <[email protected]>: > > > > > Hi all, > > > I met an error when using spark engine build kylin sample on step > "Build > > > Cube with Spark", here is the exception log: > > > ------------------------------------------------------------ > > > ----------------------------- > > > Exception in thread "main" java.lang.UnsatisfiedLinkError: > > > org.apache.hadoop.util.NativeCrc32.nativeComputeChunkedSumsByteAr > > > ray(II[BI[BIILjava/lang/String;JZ)V > > > at org.apache.hadoop.util.NativeCrc32. > > > nativeComputeChunkedSumsByteArray(Native Method) > > > at org.apache.hadoop.util.NativeCrc32. > > > calculateChunkedSumsByteArray(NativeCrc32.java:86) > > > at org.apache.hadoop.util.DataChecksum.calculateChunkedSums( > > > DataChecksum.java:430) > > > at org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunks( > > > FSOutputSummer.java:202) > > > at org.apache.hadoop.fs.FSOutputSummer.write1( > > > FSOutputSummer.java:124) > > > at org.apache.hadoop.fs.FSOutputSummer.write( > > > FSOutputSummer.java:110) > > > at org.apache.hadoop.fs.FSDataOutputStream$ > PositionCache.write( > > > FSDataOutputStream.java:58) > > > at java.io.DataOutputStream.write(DataOutputStream.java:107) > > > at org.apache.hadoop.io.IOUtils.copyBytes(IOUtils.java:80) > > > at org.apache.hadoop.io.IOUtils.copyBytes(IOUtils.java:52) > > > at org.apache.hadoop.io.IOUtils.copyBytes(IOUtils.java:112) > > > at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:366) > > > at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:338) > > > at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:289) > > > at org.apache.spark.deploy.yarn.Client.copyFileToRemote( > > > Client.scala:317) > > > at org.apache.spark.deploy.yarn.Client.org$apache$spark$ > > > deploy$yarn$Client$$distribute$1(Client.scala:407) > > > at org.apache.spark.deploy.yarn.Client$$anonfun$ > > > prepareLocalResources$5.apply(Client.scala:446) > > > at org.apache.spark.deploy.yarn.Client$$anonfun$ > > > prepareLocalResources$5.apply(Client.scala:444) > > > at scala.collection.immutable.List.foreach(List.scala:318) > > > at org.apache.spark.deploy.yarn.Client.prepareLocalResources( > > > Client.scala:444) > > > at org.apache.spark.deploy.yarn.Client. > > > createContainerLaunchContext(Client.scala:727) > > > at org.apache.spark.deploy.yarn.Client.submitApplication( > > > Client.scala:142) > > > at org.apache.spark.scheduler.cluster. > > YarnClientSchedulerBackend. > > > start(YarnClientSchedulerBackend.scala:57) > > > at org.apache.spark.scheduler.TaskSchedulerImpl.start( > > > TaskSchedulerImpl.scala:144) > > > at org.apache.spark.SparkContext. > <init>(SparkContext.scala:530) > > > at org.apache.spark.api.java.JavaSparkContext.<init>( > > > JavaSparkContext.scala:59) > > > at org.apache.kylin.engine.spark.SparkCubingByLayer.execute( > > > SparkCubingByLayer.java:150) > > > at org.apache.kylin.common.util.AbstractApplication.execute( > > > AbstractApplication.java:37) > > > at org.apache.kylin.common.util.SparkEntry.main(SparkEntry. > > > java:44) > > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > > at sun.reflect.NativeMethodAccessorImpl.invoke( > > > NativeMethodAccessorImpl.java:57) > > > at sun.reflect.DelegatingMethodAccessorImpl.invoke( > > > DelegatingMethodAccessorImpl.java:43) > > > at java.lang.reflect.Method.invoke(Method.java:606) > > > at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$ > > > deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) > > > at org.apache.spark.deploy.SparkSubmit$.doRunMain$1( > > > SparkSubmit.scala:181) > > > at org.apache.spark.deploy.SparkSubmit$.submit( > > > SparkSubmit.scala:206) > > > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit. > > > scala:121) > > > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > > > 17/06/19 21:22:06 INFO storage.DiskBlockManager: Shutdown hook called > > > 17/06/19 21:22:06 INFO util.ShutdownHookManager: Shutdown hook called > > > 17/06/19 21:22:06 INFO util.ShutdownHookManager: Deleting directory > > > /tmp/spark-0d1d3709-86cd-446c-b728-5070f168de28 > > > 17/06/19 21:22:06 INFO util.ShutdownHookManager: Deleting directory > > > /tmp/spark-0d1d3709-86cd-446c-b728-5070f168de28/httpd- > > > 9bcb9a5d-569f-4f28-ad89-038a9020eda8 > > > 17/06/19 21:22:06 INFO util.ShutdownHookManager: Deleting directory > > > /tmp/spark-0d1d3709-86cd-446c-b728-5070f168de28/userFiles- > > > 2e9ff265-3d37-40e0-8894-6fd4d1a3ad8b > > > > > > at org.apache.kylin.common.util.CliCommandExecutor.execute( > > > CliCommandExecutor.java:92) > > > at org.apache.kylin.engine.spark.SparkExecutable.doWork( > > > SparkExecutable.java:124) > > > at org.apache.kylin.job.execution.AbstractExecutable. > > > execute(AbstractExecutable.java:124) > > > at org.apache.kylin.job.execution.DefaultChainedExecutable. > > doWork( > > > DefaultChainedExecutable.java:64) > > > at org.apache.kylin.job.execution.AbstractExecutable. > > > execute(AbstractExecutable.java:124) > > > at org.apache.kylin.job.impl.threadpool.DefaultScheduler$ > > > JobRunner.run(DefaultScheduler.java:142) > > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > > > ThreadPoolExecutor.java:1145) > > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > > > ThreadPoolExecutor.java:615) > > > at java.lang.Thread.run(Thread.java:745) > > > > > > ------------------------------------------------------------ > > > ----------------------------- > > > I can use the kylin in-build spark-shell to do some operations like: > > > ------------------------------------------------------------ > > > ----------------------------- > > > var textFile = sc.textFile("hdfs://xxxx/xxxx/README.md") > > > textFile.count() > > > textFile.first() > > > textFile.filter(line => line.contains("hello")).count() > > > ------------------------------------------------------------ > > > ----------------------------- > > > Here is the env info: > > > kylin version is 2.0.0 > > > hadoop version is 2.7.* > > > spark version is 1.6.* > > > ------------------------------------------------------------ > > > ----------------------------- > > > Anyone can help me?THX > > > > > > > > > 2017-06-19 > > > skyyws > > > > > > > > > > -- > > Best regards, > > > > Shaofeng Shi 史少锋 > > > > > > -- > Best regards, > > Shaofeng Shi 史少锋 > -- Best regards, Shaofeng Shi 史少锋
