Are you running Kylin on windows? If yes, check:
https://stackoverflow.com/questions/33211599/hadoop-error-on-windows-java-lang-unsatisfiedlinkerror

2017-06-19 21:55 GMT+08:00 skyyws <sky...@163.com>:

> Hi all,
> I met an error when using spark engine build kylin sample on step "Build
> Cube with Spark", here is the exception log:
> ------------------------------------------------------------
> -----------------------------
> Exception in thread "main" java.lang.UnsatisfiedLinkError:
> org.apache.hadoop.util.NativeCrc32.nativeComputeChunkedSumsByteAr
> ray(II[BI[BIILjava/lang/String;JZ)V
>         at org.apache.hadoop.util.NativeCrc32.
> nativeComputeChunkedSumsByteArray(Native Method)
>         at org.apache.hadoop.util.NativeCrc32.
> calculateChunkedSumsByteArray(NativeCrc32.java:86)
>         at org.apache.hadoop.util.DataChecksum.calculateChunkedSums(
> DataChecksum.java:430)
>         at org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunks(
> FSOutputSummer.java:202)
>         at org.apache.hadoop.fs.FSOutputSummer.write1(
> FSOutputSummer.java:124)
>         at org.apache.hadoop.fs.FSOutputSummer.write(
> FSOutputSummer.java:110)
>         at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(
> FSDataOutputStream.java:58)
>         at java.io.DataOutputStream.write(DataOutputStream.java:107)
>         at org.apache.hadoop.io.IOUtils.copyBytes(IOUtils.java:80)
>         at org.apache.hadoop.io.IOUtils.copyBytes(IOUtils.java:52)
>         at org.apache.hadoop.io.IOUtils.copyBytes(IOUtils.java:112)
>         at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:366)
>         at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:338)
>         at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:289)
>         at org.apache.spark.deploy.yarn.Client.copyFileToRemote(
> Client.scala:317)
>         at org.apache.spark.deploy.yarn.Client.org$apache$spark$
> deploy$yarn$Client$$distribute$1(Client.scala:407)
>         at org.apache.spark.deploy.yarn.Client$$anonfun$
> prepareLocalResources$5.apply(Client.scala:446)
>         at org.apache.spark.deploy.yarn.Client$$anonfun$
> prepareLocalResources$5.apply(Client.scala:444)
>         at scala.collection.immutable.List.foreach(List.scala:318)
>         at org.apache.spark.deploy.yarn.Client.prepareLocalResources(
> Client.scala:444)
>         at org.apache.spark.deploy.yarn.Client.
> createContainerLaunchContext(Client.scala:727)
>         at org.apache.spark.deploy.yarn.Client.submitApplication(
> Client.scala:142)
>         at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.
> start(YarnClientSchedulerBackend.scala:57)
>         at org.apache.spark.scheduler.TaskSchedulerImpl.start(
> TaskSchedulerImpl.scala:144)
>         at org.apache.spark.SparkContext.<init>(SparkContext.scala:530)
>         at org.apache.spark.api.java.JavaSparkContext.<init>(
> JavaSparkContext.scala:59)
>         at org.apache.kylin.engine.spark.SparkCubingByLayer.execute(
> SparkCubingByLayer.java:150)
>         at org.apache.kylin.common.util.AbstractApplication.execute(
> AbstractApplication.java:37)
>         at org.apache.kylin.common.util.SparkEntry.main(SparkEntry.
> java:44)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:57)
>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:606)
>         at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
> deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
>         at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(
> SparkSubmit.scala:181)
>         at org.apache.spark.deploy.SparkSubmit$.submit(
> SparkSubmit.scala:206)
>         at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.
> scala:121)
>         at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> 17/06/19 21:22:06 INFO storage.DiskBlockManager: Shutdown hook called
> 17/06/19 21:22:06 INFO util.ShutdownHookManager: Shutdown hook called
> 17/06/19 21:22:06 INFO util.ShutdownHookManager: Deleting directory
> /tmp/spark-0d1d3709-86cd-446c-b728-5070f168de28
> 17/06/19 21:22:06 INFO util.ShutdownHookManager: Deleting directory
> /tmp/spark-0d1d3709-86cd-446c-b728-5070f168de28/httpd-
> 9bcb9a5d-569f-4f28-ad89-038a9020eda8
> 17/06/19 21:22:06 INFO util.ShutdownHookManager: Deleting directory
> /tmp/spark-0d1d3709-86cd-446c-b728-5070f168de28/userFiles-
> 2e9ff265-3d37-40e0-8894-6fd4d1a3ad8b
>
>         at org.apache.kylin.common.util.CliCommandExecutor.execute(
> CliCommandExecutor.java:92)
>         at org.apache.kylin.engine.spark.SparkExecutable.doWork(
> SparkExecutable.java:124)
>         at org.apache.kylin.job.execution.AbstractExecutable.
> execute(AbstractExecutable.java:124)
>         at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(
> DefaultChainedExecutable.java:64)
>         at org.apache.kylin.job.execution.AbstractExecutable.
> execute(AbstractExecutable.java:124)
>         at org.apache.kylin.job.impl.threadpool.DefaultScheduler$
> JobRunner.run(DefaultScheduler.java:142)
>         at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
>         at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
>         at java.lang.Thread.run(Thread.java:745)
>
> ------------------------------------------------------------
> -----------------------------
> I can use the kylin in-build spark-shell to do some operations like:
> ------------------------------------------------------------
> -----------------------------
> var textFile = sc.textFile("hdfs://xxxx/xxxx/README.md")
> textFile.count()
> textFile.first()
> textFile.filter(line => line.contains("hello")).count()
> ------------------------------------------------------------
> -----------------------------
> Here is the env info:
> kylin version is 2.0.0
> hadoop version is 2.7.*
> spark version is 1.6.*
> ------------------------------------------------------------
> -----------------------------
> Anyone can help me?THX
>
>
> 2017-06-19
> skyyws




-- 
Best regards,

Shaofeng Shi 史少锋

Reply via email to