[ https://issues.apache.org/jira/browse/MAHOUT-1946?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Andrew Musselman updated MAHOUT-1946: ------------------------------------- Sprint: Jan/Feb-2017 > ViennaCL not being picked up by JNI > ----------------------------------- > > Key: MAHOUT-1946 > URL: https://issues.apache.org/jira/browse/MAHOUT-1946 > Project: Mahout > Issue Type: Bug > Reporter: Andrew Musselman > Priority: Blocker > Fix For: 0.13.0 > > > Using the PR for MAHOUT-1938 but probably in master as well: > scala> :load ./examples/bin/SparseSparseDrmTimer.mscala > Loading ./examples/bin/SparseSparseDrmTimer.mscala... > timeSparseDRMMMul: (m: Int, n: Int, s: Int, para: Int, pctDense: Double, > seed: Long)Long > scala> timeSparseDRMMMul(100,100,100,1,.02,1234L) > [INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver > [INFO] Successfully created org.apache.mahout.viennacl.opencl.GPUMMul solver > gpuRWCW > 17/02/26 13:18:54 ERROR Executor: Exception in task 0.0 in stage 3.0 (TID 3) > java.lang.UnsatisfiedLinkError: no jniViennaCL in java.library.path > at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1867) > at java.lang.Runtime.loadLibrary0(Runtime.java:870) > at java.lang.System.loadLibrary(System.java:1122) > at org.bytedeco.javacpp.Loader.loadLibrary(Loader.java:726) > at org.bytedeco.javacpp.Loader.load(Loader.java:501) > at org.bytedeco.javacpp.Loader.load(Loader.java:434) > at > org.apache.mahout.viennacl.opencl.javacpp.Context$.loadLib(Context.scala:63) > at > org.apache.mahout.viennacl.opencl.javacpp.Context$.<init>(Context.scala:65) > at > org.apache.mahout.viennacl.opencl.javacpp.Context$.<clinit>(Context.scala) > at > org.apache.mahout.viennacl.opencl.GPUMMul$.org$apache$mahout$viennacl$opencl$GPUMMul$$gpuRWCW(GPUMMul.scala:171) > at > org.apache.mahout.viennacl.opencl.GPUMMul$$anonfun$11.apply(GPUMMul.scala:77) > at > org.apache.mahout.viennacl.opencl.GPUMMul$$anonfun$11.apply(GPUMMul.scala:77) > at org.apache.mahout.viennacl.opencl.GPUMMul$.apply(GPUMMul.scala:127) > at org.apache.mahout.viennacl.opencl.GPUMMul$.apply(GPUMMul.scala:33) > at > org.apache.mahout.math.scalabindings.RLikeMatrixOps.$percent$times$percent(RLikeMatrixOps.scala:37) > at > org.apache.mahout.sparkbindings.blas.ABt$.org$apache$mahout$sparkbindings$blas$ABt$$mmulFunc$1(ABt.scala:98) > at > org.apache.mahout.sparkbindings.blas.ABt$$anonfun$6.apply(ABt.scala:113) > at > org.apache.mahout.sparkbindings.blas.ABt$$anonfun$6.apply(ABt.scala:113) > at > org.apache.mahout.sparkbindings.blas.ABt$$anonfun$pairwiseApply$1.apply(ABt.scala:209) > at > org.apache.mahout.sparkbindings.blas.ABt$$anonfun$pairwiseApply$1.apply(ABt.scala:209) > at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) > at > org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:191) > at > org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64) > at > org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73) > at > org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) > at org.apache.spark.scheduler.Task.run(Task.scala:89) > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > 17/02/26 13:18:54 ERROR SparkUncaughtExceptionHandler: Uncaught exception in > thread Thread[Executor task launch worker-0,5,main] > java.lang.UnsatisfiedLinkError: no jniViennaCL in java.library.path > at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1867) > at java.lang.Runtime.loadLibrary0(Runtime.java:870) > at java.lang.System.loadLibrary(System.java:1122) > at org.bytedeco.javacpp.Loader.loadLibrary(Loader.java:726) > at org.bytedeco.javacpp.Loader.load(Loader.java:501) > at org.bytedeco.javacpp.Loader.load(Loader.java:434) > at > org.apache.mahout.viennacl.opencl.javacpp.Context$.loadLib(Context.scala:63) > at > org.apache.mahout.viennacl.opencl.javacpp.Context$.<init>(Context.scala:65) > at > org.apache.mahout.viennacl.opencl.javacpp.Context$.<clinit>(Context.scala) > at > org.apache.mahout.viennacl.opencl.GPUMMul$.org$apache$mahout$viennacl$opencl$GPUMMul$$gpuRWCW(GPUMMul.scala:171) > at > org.apache.mahout.viennacl.opencl.GPUMMul$$anonfun$11.apply(GPUMMul.scala:77) > at > org.apache.mahout.viennacl.opencl.GPUMMul$$anonfun$11.apply(GPUMMul.scala:77) > at org.apache.mahout.viennacl.opencl.GPUMMul$.apply(GPUMMul.scala:127) > at org.apache.mahout.viennacl.opencl.GPUMMul$.apply(GPUMMul.scala:33) > at > org.apache.mahout.math.scalabindings.RLikeMatrixOps.$percent$times$percent(RLikeMatrixOps.scala:37) > at > org.apache.mahout.sparkbindings.blas.ABt$.org$apache$mahout$sparkbindings$blas$ABt$$mmulFunc$1(ABt.scala:98) > at > org.apache.mahout.sparkbindings.blas.ABt$$anonfun$6.apply(ABt.scala:113) > at > org.apache.mahout.sparkbindings.blas.ABt$$anonfun$6.apply(ABt.scala:113) > at > org.apache.mahout.sparkbindings.blas.ABt$$anonfun$pairwiseApply$1.apply(ABt.scala:209) > at > org.apache.mahout.sparkbindings.blas.ABt$$anonfun$pairwiseApply$1.apply(ABt.scala:209) > at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) > at > org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:191) > at > org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64) > at > org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73) > at > org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) > at org.apache.spark.scheduler.Task.run(Task.scala:89) > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > 17/02/26 13:18:54 WARN TaskSetManager: Lost task 0.0 in stage 3.0 (TID 3, > localhost): java.lang.UnsatisfiedLinkError: no jniViennaCL in > java.library.path > at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1867) > at java.lang.Runtime.loadLibrary0(Runtime.java:870) > at java.lang.System.loadLibrary(System.java:1122) > at org.bytedeco.javacpp.Loader.loadLibrary(Loader.java:726) > at org.bytedeco.javacpp.Loader.load(Loader.java:501) > at org.bytedeco.javacpp.Loader.load(Loader.java:434) > at > org.apache.mahout.viennacl.opencl.javacpp.Context$.loadLib(Context.scala:63) > at > org.apache.mahout.viennacl.opencl.javacpp.Context$.<init>(Context.scala:65) > at > org.apache.mahout.viennacl.opencl.javacpp.Context$.<clinit>(Context.scala) > at > org.apache.mahout.viennacl.opencl.GPUMMul$.org$apache$mahout$viennacl$opencl$GPUMMul$$gpuRWCW(GPUMMul.scala:171) > at > org.apache.mahout.viennacl.opencl.GPUMMul$$anonfun$11.apply(GPUMMul.scala:77) > at > org.apache.mahout.viennacl.opencl.GPUMMul$$anonfun$11.apply(GPUMMul.scala:77) > at org.apache.mahout.viennacl.opencl.GPUMMul$.apply(GPUMMul.scala:127) > at org.apache.mahout.viennacl.opencl.GPUMMul$.apply(GPUMMul.scala:33) > at > org.apache.mahout.math.scalabindings.RLikeMatrixOps.$percent$times$percent(RLikeMatrixOps.scala:37) > at > org.apache.mahout.sparkbindings.blas.ABt$.org$apache$mahout$sparkbindings$blas$ABt$$mmulFunc$1(ABt.scala:98) > at > org.apache.mahout.sparkbindings.blas.ABt$$anonfun$6.apply(ABt.scala:113) > at > org.apache.mahout.sparkbindings.blas.ABt$$anonfun$6.apply(ABt.scala:113) > at > org.apache.mahout.sparkbindings.blas.ABt$$anonfun$pairwiseApply$1.apply(ABt.scala:209) > at > org.apache.mahout.sparkbindings.blas.ABt$$anonfun$pairwiseApply$1.apply(ABt.scala:209) > at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) > at > org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:191) > at > org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64) > at > org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73) > at > org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) > at org.apache.spark.scheduler.Task.run(Task.scala:89) > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > 17/02/26 13:18:54 ERROR TaskSetManager: Task 0 in stage 3.0 failed 1 times; > aborting job > 17/02/26 13:18:54 ERROR ShutdownHookManager: Exception while deleting Spark > temp dir: /tmp/spark-34fcf1f5-a8e6-46f8-bdd6-dd1abab2ded1 > java.io.IOException: Failed to delete: > /tmp/spark-34fcf1f5-a8e6-46f8-bdd6-dd1abab2ded1 > at org.apache.spark.util.Utils$.deleteRecursively(Utils.scala:929) > at > org.apache.spark.util.ShutdownHookManager$$anonfun$1$$anonfun$apply$mcV$sp$3.apply(ShutdownHookManager.scala:65) > at > org.apache.spark.util.ShutdownHookManager$$anonfun$1$$anonfun$apply$mcV$sp$3.apply(ShutdownHookManager.scala:62) > at > scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33) > at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108) > at > org.apache.spark.util.ShutdownHookManager$$anonfun$1.apply$mcV$sp(ShutdownHookManager.scala:62) > at > org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:267) > at > org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ShutdownHookManager.scala:239) > at > org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply(ShutdownHookManager.scala:239) > at > org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply(ShutdownHookManager.scala:239) > at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1817) > at > org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply$mcV$sp(ShutdownHookManager.scala:239) > at > org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply(ShutdownHookManager.scala:239) > at > org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply(ShutdownHookManager.scala:239) > at scala.util.Try$.apply(Try.scala:161) > at > org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:239) > at > org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:218) > at > org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54) -- This message was sent by Atlassian JIRA (v6.3.15#6346)