See <https://builds.apache.org/job/Mahout-Quality/3421/display/redirect?page=changes>
Changes: [apalumbo] MAHOUT-1912: MAHOUT-1912: CLI driver tests not working with vienniacl. ------------------------------------------ [...truncated 434.20 KB...] [31m[0m [31mDriver stacktrace:[0m [31m at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)[0m [31m at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)[0m [31m at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)[0m [31m at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)[0m [31m at scala.Option.foreach(Option.scala:236)[0m [31m at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)[0m [31m ...[0m [31m Cause: java.io.FileNotFoundException: /tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/0e/shuffle_45_1_0.index.686b1f33-07ce-4c6a-8675-03f4bf699f82 (No such file or directory)[0m [31m at java.io.FileOutputStream.open(Native Method)[0m [31m at java.io.FileOutputStream.<init>(FileOutputStream.java:221)[0m [31m at java.io.FileOutputStream.<init>(FileOutputStream.java:171)[0m [31m at org.apache.spark.shuffle.IndexShuffleBlockResolver.writeIndexFileAndCommit(IndexShuffleBlockResolver.scala:142)[0m [31m at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:128)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)[0m [31m at org.apache.spark.scheduler.Task.run(Task.scala:89)[0m [31m at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)[0m [31m at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)[0m [31m ...[0m [31m- C = cbind(A, B) with missing rows *** FAILED ***[0m [31m org.apache.spark.SparkException: Job aborted due to stage failure: Task 2 in stage 117.0 failed 1 times, most recent failure: Lost task 2.0 in stage 117.0 (TID 285, localhost): java.io.FileNotFoundException: /tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/2f/shuffle_47_2_0.index.2f3483a4-17ab-42d2-9be9-46438957695b (No such file or directory)[0m [31m at java.io.FileOutputStream.open(Native Method)[0m [31m at java.io.FileOutputStream.<init>(FileOutputStream.java:221)[0m [31m at java.io.FileOutputStream.<init>(FileOutputStream.java:171)[0m [31m at org.apache.spark.shuffle.IndexShuffleBlockResolver.writeIndexFileAndCommit(IndexShuffleBlockResolver.scala:142)[0m [31m at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:128)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)[0m [31m at org.apache.spark.scheduler.Task.run(Task.scala:89)[0m [31m at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)[0m [31m at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)[0m [31m at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)[0m [31m at java.lang.Thread.run(Thread.java:745)[0m [31m[0m [31mDriver stacktrace:[0m [31m at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)[0m [31m at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)[0m [31m at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)[0m [31m at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)[0m [31m at scala.Option.foreach(Option.scala:236)[0m [31m at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)[0m [31m ...[0m [31m Cause: java.io.FileNotFoundException: /tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/2f/shuffle_47_2_0.index.2f3483a4-17ab-42d2-9be9-46438957695b (No such file or directory)[0m [31m at java.io.FileOutputStream.open(Native Method)[0m [31m at java.io.FileOutputStream.<init>(FileOutputStream.java:221)[0m [31m at java.io.FileOutputStream.<init>(FileOutputStream.java:171)[0m [31m at org.apache.spark.shuffle.IndexShuffleBlockResolver.writeIndexFileAndCommit(IndexShuffleBlockResolver.scala:142)[0m [31m at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:128)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)[0m [31m at org.apache.spark.scheduler.Task.run(Task.scala:89)[0m [31m at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)[0m [31m at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)[0m [31m ...[0m collected A = { 0 => {0:1.0,1:2.0,2:3.0} 1 => {} 2 => {} 3 => {0:3.0,1:4.0,2:5.0} } [31m- B = A + 1.0 missing rows *** FAILED ***[0m [31m org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 124.0 failed 1 times, most recent failure: Lost task 0.0 in stage 124.0 (TID 327, localhost): java.io.FileNotFoundException: /tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/2f/shuffle_49_0_0.index.cea4ee67-ee4a-48c8-a29e-6a841c6ed1c8 (No such file or directory)[0m [31m at java.io.FileOutputStream.open(Native Method)[0m [31m at java.io.FileOutputStream.<init>(FileOutputStream.java:221)[0m [31m at java.io.FileOutputStream.<init>(FileOutputStream.java:171)[0m [31m at org.apache.spark.shuffle.IndexShuffleBlockResolver.writeIndexFileAndCommit(IndexShuffleBlockResolver.scala:142)[0m [31m at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:128)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)[0m [31m at org.apache.spark.scheduler.Task.run(Task.scala:89)[0m [31m at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)[0m [31m at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)[0m [31m at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)[0m [31m at java.lang.Thread.run(Thread.java:745)[0m [31m[0m [31mDriver stacktrace:[0m [31m at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)[0m [31m at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)[0m [31m at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)[0m [31m at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)[0m [31m at scala.Option.foreach(Option.scala:236)[0m [31m at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)[0m [31m ...[0m [31m Cause: java.io.FileNotFoundException: /tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/2f/shuffle_49_0_0.index.cea4ee67-ee4a-48c8-a29e-6a841c6ed1c8 (No such file or directory)[0m [31m at java.io.FileOutputStream.open(Native Method)[0m [31m at java.io.FileOutputStream.<init>(FileOutputStream.java:221)[0m [31m at java.io.FileOutputStream.<init>(FileOutputStream.java:171)[0m [31m at org.apache.spark.shuffle.IndexShuffleBlockResolver.writeIndexFileAndCommit(IndexShuffleBlockResolver.scala:142)[0m [31m at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:128)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)[0m [31m at org.apache.spark.scheduler.Task.run(Task.scala:89)[0m [31m at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)[0m [31m at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)[0m [31m ...[0m [INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver [WARN] Unable to create class GPUMMul: attempting OpenMP version [INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver org.apache.mahout.viennacl.openmp.OMPMMul$ [INFO] Unable to create class OMPMMul: falling back to java version in-core mul ms: 1498 [31m- A'B, bigger *** FAILED ***[0m [31m org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 127.0 failed 1 times, most recent failure: Lost task 0.0 in stage 127.0 (TID 329, localhost): java.io.FileNotFoundException: /tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/06/temp_shuffle_27b7da17-0df2-48ae-8277-43460317ed34 (No such file or directory)[0m [31m at java.io.FileOutputStream.open(Native Method)[0m [31m at java.io.FileOutputStream.<init>(FileOutputStream.java:221)[0m [31m at org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:88)[0m [31m at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:140)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)[0m [31m at org.apache.spark.scheduler.Task.run(Task.scala:89)[0m [31m at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)[0m [31m at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)[0m [31m at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)[0m [31m at java.lang.Thread.run(Thread.java:745)[0m [31m[0m [31mDriver stacktrace:[0m [31m at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)[0m [31m at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)[0m [31m at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)[0m [31m at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)[0m [31m at scala.Option.foreach(Option.scala:236)[0m [31m at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)[0m [31m ...[0m [31m Cause: java.io.FileNotFoundException: /tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/06/temp_shuffle_27b7da17-0df2-48ae-8277-43460317ed34 (No such file or directory)[0m [31m at java.io.FileOutputStream.open(Native Method)[0m [31m at java.io.FileOutputStream.<init>(FileOutputStream.java:221)[0m [31m at org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:88)[0m [31m at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:140)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)[0m [31m at org.apache.spark.scheduler.Task.run(Task.scala:89)[0m [31m at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)[0m [31m at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)[0m [31m at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)[0m [31m ...[0m [INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver [WARN] Unable to create class GPUMMul: attempting OpenMP version [INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver org.apache.mahout.viennacl.openmp.OMPMMul$ [INFO] Unable to create class OMPMMul: falling back to java version [INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver [WARN] Unable to create class GPUMMul: attempting OpenMP version [INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver org.apache.mahout.viennacl.openmp.OMPMMul$ [INFO] Unable to create class OMPMMul: falling back to java version [INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver [WARN] Unable to create class GPUMMul: attempting OpenMP version [INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver org.apache.mahout.viennacl.openmp.OMPMMul$ [INFO] Unable to create class OMPMMul: falling back to java version [INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver [WARN] Unable to create class GPUMMul: attempting OpenMP version [INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver org.apache.mahout.viennacl.openmp.OMPMMul$ [INFO] Unable to create class OMPMMul: falling back to java version [INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver [WARN] Unable to create class GPUMMul: attempting OpenMP version [INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver org.apache.mahout.viennacl.openmp.OMPMMul$ [INFO] Unable to create class OMPMMul: falling back to java version [32m- C = At %*% B , zippable[0m [32mTextDelimitedReaderWriterSuite:[0m [32m- indexedDatasetDFSRead should read sparse matrix file with null rows[0m [32mPreprocessorSuite:[0m OpMapBlock(org.apache.mahout.sparkbindings.drm.CheckpointedDrmSpark@70b9d2ee,<function1>,8,-1,true) {1:3.0,2:5.0,3:6.0} [32m- asfactor test[0m {0:2.0,1:5.0,2:-4.0} {0:0.8164965809277263,1:3.2659863237109037,2:8.286535263104035} [32m- standard scaler test[0m [32m- mean center test[0m [32mTFIDFSparkTestSuite:[0m [32m- TF test[0m [32m- TFIDF test[0m [32m- MLlib TFIDF test[0m [36mRun completed in 1 minute, 39 seconds.[0m [36mTotal number of tests run: 128[0m [36mSuites: completed 18, aborted 0[0m [36mTests: succeeded 119, failed 9, canceled 0, ignored 1, pending 0[0m [31m*** 9 TESTS FAILED ***[0m [INFO] [INFO] ------------------------------------------------------------------------ [INFO] Skipping Apache Mahout [INFO] This project has been banned from the build due to previous failures. [INFO] ------------------------------------------------------------------------ [INFO] [INFO] ------------------------------------------------------------------------ [INFO] Skipping Mahout Build Tools [INFO] This project has been banned from the build due to previous failures. [INFO] ------------------------------------------------------------------------ [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary: [INFO] [INFO] Mahout Build Tools ................................. SUCCESS [ 3.068 s] [INFO] Apache Mahout ...................................... SUCCESS [ 0.176 s] [INFO] Mahout Math ........................................ SUCCESS [01:13 min] [INFO] Mahout HDFS ........................................ SUCCESS [ 5.306 s] [INFO] Mahout Map-Reduce .................................. SUCCESS [12:12 min] [INFO] Mahout Integration ................................. SUCCESS [ 45.598 s] [INFO] Mahout Examples .................................... SUCCESS [ 25.099 s] [INFO] Mahout Math Scala bindings ......................... SUCCESS [05:10 min] [INFO] Mahout Spark bindings .............................. FAILURE [02:34 min] [INFO] Mahout Flink bindings .............................. SKIPPED [INFO] Mahout Release Package ............................. SKIPPED [INFO] Mahout H2O backend ................................. SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 22:32 min [INFO] Finished at: 2017-02-26T00:36:46+00:00 [INFO] Final Memory: 65M/606M [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.scalatest:scalatest-maven-plugin:1.0:test (test) on project mahout-spark_2.10: There are test failures -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn <goals> -rf :mahout-spark_2.10 Build step 'Invoke top-level Maven targets' marked build as failure [PMD] Skipping publisher since build result is FAILURE [TASKS] Skipping publisher since build result is FAILURE Archiving artifacts Compressed 153.92 MB of artifacts by 89.6% relative to #3418 Recording test results Publishing Javadoc [JIRA] Updating issue MAHOUT-1912