See
<https://builds.apache.org/job/Mahout-Quality/3421/display/redirect?page=changes>
Changes:
[apalumbo] MAHOUT-1912: MAHOUT-1912: CLI driver tests not working with
vienniacl.
------------------------------------------
[...truncated 434.20 KB...]
[31m[0m
[31mDriver stacktrace:[0m
[31m at
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)[0m
[31m at
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)[0m
[31m at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)[0m
[31m at scala.Option.foreach(Option.scala:236)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)[0m
[31m ...[0m
[31m Cause: java.io.FileNotFoundException:
/tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/0e/shuffle_45_1_0.index.686b1f33-07ce-4c6a-8675-03f4bf699f82
(No such file or directory)[0m
[31m at java.io.FileOutputStream.open(Native Method)[0m
[31m at java.io.FileOutputStream.<init>(FileOutputStream.java:221)[0m
[31m at java.io.FileOutputStream.<init>(FileOutputStream.java:171)[0m
[31m at
org.apache.spark.shuffle.IndexShuffleBlockResolver.writeIndexFileAndCommit(IndexShuffleBlockResolver.scala:142)[0m
[31m at
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:128)[0m
[31m at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)[0m
[31m at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)[0m
[31m at org.apache.spark.scheduler.Task.run(Task.scala:89)[0m
[31m at
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)[0m
[31m at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)[0m
[31m ...[0m
[31m- C = cbind(A, B) with missing rows *** FAILED ***[0m
[31m org.apache.spark.SparkException: Job aborted due to stage failure: Task
2 in stage 117.0 failed 1 times, most recent failure: Lost task 2.0 in stage
117.0 (TID 285, localhost): java.io.FileNotFoundException:
/tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/2f/shuffle_47_2_0.index.2f3483a4-17ab-42d2-9be9-46438957695b
(No such file or directory)[0m
[31m at java.io.FileOutputStream.open(Native Method)[0m
[31m at java.io.FileOutputStream.<init>(FileOutputStream.java:221)[0m
[31m at java.io.FileOutputStream.<init>(FileOutputStream.java:171)[0m
[31m at
org.apache.spark.shuffle.IndexShuffleBlockResolver.writeIndexFileAndCommit(IndexShuffleBlockResolver.scala:142)[0m
[31m at
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:128)[0m
[31m at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)[0m
[31m at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)[0m
[31m at org.apache.spark.scheduler.Task.run(Task.scala:89)[0m
[31m at
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)[0m
[31m at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)[0m
[31m at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)[0m
[31m at java.lang.Thread.run(Thread.java:745)[0m
[31m[0m
[31mDriver stacktrace:[0m
[31m at
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)[0m
[31m at
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)[0m
[31m at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)[0m
[31m at scala.Option.foreach(Option.scala:236)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)[0m
[31m ...[0m
[31m Cause: java.io.FileNotFoundException:
/tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/2f/shuffle_47_2_0.index.2f3483a4-17ab-42d2-9be9-46438957695b
(No such file or directory)[0m
[31m at java.io.FileOutputStream.open(Native Method)[0m
[31m at java.io.FileOutputStream.<init>(FileOutputStream.java:221)[0m
[31m at java.io.FileOutputStream.<init>(FileOutputStream.java:171)[0m
[31m at
org.apache.spark.shuffle.IndexShuffleBlockResolver.writeIndexFileAndCommit(IndexShuffleBlockResolver.scala:142)[0m
[31m at
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:128)[0m
[31m at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)[0m
[31m at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)[0m
[31m at org.apache.spark.scheduler.Task.run(Task.scala:89)[0m
[31m at
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)[0m
[31m at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)[0m
[31m ...[0m
collected A =
{
0 => {0:1.0,1:2.0,2:3.0}
1 => {}
2 => {}
3 => {0:3.0,1:4.0,2:5.0}
}
[31m- B = A + 1.0 missing rows *** FAILED ***[0m
[31m org.apache.spark.SparkException: Job aborted due to stage failure: Task
0 in stage 124.0 failed 1 times, most recent failure: Lost task 0.0 in stage
124.0 (TID 327, localhost): java.io.FileNotFoundException:
/tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/2f/shuffle_49_0_0.index.cea4ee67-ee4a-48c8-a29e-6a841c6ed1c8
(No such file or directory)[0m
[31m at java.io.FileOutputStream.open(Native Method)[0m
[31m at java.io.FileOutputStream.<init>(FileOutputStream.java:221)[0m
[31m at java.io.FileOutputStream.<init>(FileOutputStream.java:171)[0m
[31m at
org.apache.spark.shuffle.IndexShuffleBlockResolver.writeIndexFileAndCommit(IndexShuffleBlockResolver.scala:142)[0m
[31m at
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:128)[0m
[31m at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)[0m
[31m at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)[0m
[31m at org.apache.spark.scheduler.Task.run(Task.scala:89)[0m
[31m at
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)[0m
[31m at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)[0m
[31m at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)[0m
[31m at java.lang.Thread.run(Thread.java:745)[0m
[31m[0m
[31mDriver stacktrace:[0m
[31m at
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)[0m
[31m at
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)[0m
[31m at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)[0m
[31m at scala.Option.foreach(Option.scala:236)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)[0m
[31m ...[0m
[31m Cause: java.io.FileNotFoundException:
/tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/2f/shuffle_49_0_0.index.cea4ee67-ee4a-48c8-a29e-6a841c6ed1c8
(No such file or directory)[0m
[31m at java.io.FileOutputStream.open(Native Method)[0m
[31m at java.io.FileOutputStream.<init>(FileOutputStream.java:221)[0m
[31m at java.io.FileOutputStream.<init>(FileOutputStream.java:171)[0m
[31m at
org.apache.spark.shuffle.IndexShuffleBlockResolver.writeIndexFileAndCommit(IndexShuffleBlockResolver.scala:142)[0m
[31m at
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:128)[0m
[31m at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)[0m
[31m at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)[0m
[31m at org.apache.spark.scheduler.Task.run(Task.scala:89)[0m
[31m at
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)[0m
[31m at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)[0m
[31m ...[0m
[INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver
[WARN] Unable to create class GPUMMul: attempting OpenMP version
[INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver
org.apache.mahout.viennacl.openmp.OMPMMul$
[INFO] Unable to create class OMPMMul: falling back to java version
in-core mul ms: 1498
[31m- A'B, bigger *** FAILED ***[0m
[31m org.apache.spark.SparkException: Job aborted due to stage failure: Task
0 in stage 127.0 failed 1 times, most recent failure: Lost task 0.0 in stage
127.0 (TID 329, localhost): java.io.FileNotFoundException:
/tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/06/temp_shuffle_27b7da17-0df2-48ae-8277-43460317ed34
(No such file or directory)[0m
[31m at java.io.FileOutputStream.open(Native Method)[0m
[31m at java.io.FileOutputStream.<init>(FileOutputStream.java:221)[0m
[31m at
org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:88)[0m
[31m at
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:140)[0m
[31m at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)[0m
[31m at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)[0m
[31m at org.apache.spark.scheduler.Task.run(Task.scala:89)[0m
[31m at
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)[0m
[31m at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)[0m
[31m at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)[0m
[31m at java.lang.Thread.run(Thread.java:745)[0m
[31m[0m
[31mDriver stacktrace:[0m
[31m at
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)[0m
[31m at
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)[0m
[31m at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)[0m
[31m at scala.Option.foreach(Option.scala:236)[0m
[31m at
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)[0m
[31m ...[0m
[31m Cause: java.io.FileNotFoundException:
/tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/06/temp_shuffle_27b7da17-0df2-48ae-8277-43460317ed34
(No such file or directory)[0m
[31m at java.io.FileOutputStream.open(Native Method)[0m
[31m at java.io.FileOutputStream.<init>(FileOutputStream.java:221)[0m
[31m at
org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:88)[0m
[31m at
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:140)[0m
[31m at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)[0m
[31m at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)[0m
[31m at org.apache.spark.scheduler.Task.run(Task.scala:89)[0m
[31m at
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)[0m
[31m at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)[0m
[31m at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)[0m
[31m ...[0m
[INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver
[WARN] Unable to create class GPUMMul: attempting OpenMP version
[INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver
org.apache.mahout.viennacl.openmp.OMPMMul$
[INFO] Unable to create class OMPMMul: falling back to java version
[INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver
[WARN] Unable to create class GPUMMul: attempting OpenMP version
[INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver
org.apache.mahout.viennacl.openmp.OMPMMul$
[INFO] Unable to create class OMPMMul: falling back to java version
[INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver
[WARN] Unable to create class GPUMMul: attempting OpenMP version
[INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver
org.apache.mahout.viennacl.openmp.OMPMMul$
[INFO] Unable to create class OMPMMul: falling back to java version
[INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver
[WARN] Unable to create class GPUMMul: attempting OpenMP version
[INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver
org.apache.mahout.viennacl.openmp.OMPMMul$
[INFO] Unable to create class OMPMMul: falling back to java version
[INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver
[WARN] Unable to create class GPUMMul: attempting OpenMP version
[INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver
org.apache.mahout.viennacl.openmp.OMPMMul$
[INFO] Unable to create class OMPMMul: falling back to java version
[32m- C = At %*% B , zippable[0m
[32mTextDelimitedReaderWriterSuite:[0m
[32m- indexedDatasetDFSRead should read sparse matrix file with null rows[0m
[32mPreprocessorSuite:[0m
OpMapBlock(org.apache.mahout.sparkbindings.drm.CheckpointedDrmSpark@70b9d2ee,<function1>,8,-1,true)
{1:3.0,2:5.0,3:6.0}
[32m- asfactor test[0m
{0:2.0,1:5.0,2:-4.0}
{0:0.8164965809277263,1:3.2659863237109037,2:8.286535263104035}
[32m- standard scaler test[0m
[32m- mean center test[0m
[32mTFIDFSparkTestSuite:[0m
[32m- TF test[0m
[32m- TFIDF test[0m
[32m- MLlib TFIDF test[0m
[36mRun completed in 1 minute, 39 seconds.[0m
[36mTotal number of tests run: 128[0m
[36mSuites: completed 18, aborted 0[0m
[36mTests: succeeded 119, failed 9, canceled 0, ignored 1, pending 0[0m
[31m*** 9 TESTS FAILED ***[0m
[INFO]
[INFO] ------------------------------------------------------------------------
[INFO] Skipping Apache Mahout
[INFO] This project has been banned from the build due to previous failures.
[INFO] ------------------------------------------------------------------------
[INFO]
[INFO] ------------------------------------------------------------------------
[INFO] Skipping Mahout Build Tools
[INFO] This project has been banned from the build due to previous failures.
[INFO] ------------------------------------------------------------------------
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Mahout Build Tools ................................. SUCCESS [ 3.068 s]
[INFO] Apache Mahout ...................................... SUCCESS [ 0.176 s]
[INFO] Mahout Math ........................................ SUCCESS [01:13 min]
[INFO] Mahout HDFS ........................................ SUCCESS [ 5.306 s]
[INFO] Mahout Map-Reduce .................................. SUCCESS [12:12 min]
[INFO] Mahout Integration ................................. SUCCESS [ 45.598 s]
[INFO] Mahout Examples .................................... SUCCESS [ 25.099 s]
[INFO] Mahout Math Scala bindings ......................... SUCCESS [05:10 min]
[INFO] Mahout Spark bindings .............................. FAILURE [02:34 min]
[INFO] Mahout Flink bindings .............................. SKIPPED
[INFO] Mahout Release Package ............................. SKIPPED
[INFO] Mahout H2O backend ................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 22:32 min
[INFO] Finished at: 2017-02-26T00:36:46+00:00
[INFO] Final Memory: 65M/606M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.scalatest:scalatest-maven-plugin:1.0:test
(test) on project mahout-spark_2.10: There are test failures -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please
read the following articles:
[ERROR] [Help 1]
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :mahout-spark_2.10
Build step 'Invoke top-level Maven targets' marked build as failure
[PMD] Skipping publisher since build result is FAILURE
[TASKS] Skipping publisher since build result is FAILURE
Archiving artifacts
Compressed 153.92 MB of artifacts by 89.6% relative to #3418
Recording test results
Publishing Javadoc
[JIRA] Updating issue MAHOUT-1912