See 
<https://builds.apache.org/job/Mahout-Quality/3421/display/redirect?page=changes>

Changes:

[apalumbo] MAHOUT-1912: MAHOUT-1912: CLI driver tests not working with 
vienniacl.

------------------------------------------
[...truncated 434.20 KB...]

Driver stacktrace:
  at 
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)
  at 
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
  at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
  at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
  at scala.Option.foreach(Option.scala:236)
  at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)
  ...
  Cause: java.io.FileNotFoundException: 
/tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/0e/shuffle_45_1_0.index.686b1f33-07ce-4c6a-8675-03f4bf699f82
 (No such file or directory)
  at java.io.FileOutputStream.open(Native Method)
  at java.io.FileOutputStream.<init>(FileOutputStream.java:221)
  at java.io.FileOutputStream.<init>(FileOutputStream.java:171)
  at 
org.apache.spark.shuffle.IndexShuffleBlockResolver.writeIndexFileAndCommit(IndexShuffleBlockResolver.scala:142)
  at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:128)
  at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  at org.apache.spark.scheduler.Task.run(Task.scala:89)
  at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
  at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
  ...
- C = cbind(A, B) with missing rows *** FAILED ***
  org.apache.spark.SparkException: Job aborted due to stage failure: Task 
2 in stage 117.0 failed 1 times, most recent failure: Lost task 2.0 in stage 
117.0 (TID 285, localhost): java.io.FileNotFoundException: 
/tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/2f/shuffle_47_2_0.index.2f3483a4-17ab-42d2-9be9-46438957695b
 (No such file or directory)
   at java.io.FileOutputStream.open(Native Method)
   at java.io.FileOutputStream.<init>(FileOutputStream.java:221)
   at java.io.FileOutputStream.<init>(FileOutputStream.java:171)
   at 
org.apache.spark.shuffle.IndexShuffleBlockResolver.writeIndexFileAndCommit(IndexShuffleBlockResolver.scala:142)
   at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:128)
   at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
   at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
   at org.apache.spark.scheduler.Task.run(Task.scala:89)
   at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
   at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:745)

Driver stacktrace:
  at 
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)
  at 
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
  at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
  at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
  at scala.Option.foreach(Option.scala:236)
  at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)
  ...
  Cause: java.io.FileNotFoundException: 
/tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/2f/shuffle_47_2_0.index.2f3483a4-17ab-42d2-9be9-46438957695b
 (No such file or directory)
  at java.io.FileOutputStream.open(Native Method)
  at java.io.FileOutputStream.<init>(FileOutputStream.java:221)
  at java.io.FileOutputStream.<init>(FileOutputStream.java:171)
  at 
org.apache.spark.shuffle.IndexShuffleBlockResolver.writeIndexFileAndCommit(IndexShuffleBlockResolver.scala:142)
  at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:128)
  at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  at org.apache.spark.scheduler.Task.run(Task.scala:89)
  at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
  at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
  ...
collected A = 
{
 0 =>   {0:1.0,1:2.0,2:3.0}
 1 =>   {}
 2 =>   {}
 3 =>   {0:3.0,1:4.0,2:5.0}
}
- B = A + 1.0 missing rows *** FAILED ***
  org.apache.spark.SparkException: Job aborted due to stage failure: Task 
0 in stage 124.0 failed 1 times, most recent failure: Lost task 0.0 in stage 
124.0 (TID 327, localhost): java.io.FileNotFoundException: 
/tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/2f/shuffle_49_0_0.index.cea4ee67-ee4a-48c8-a29e-6a841c6ed1c8
 (No such file or directory)
   at java.io.FileOutputStream.open(Native Method)
   at java.io.FileOutputStream.<init>(FileOutputStream.java:221)
   at java.io.FileOutputStream.<init>(FileOutputStream.java:171)
   at 
org.apache.spark.shuffle.IndexShuffleBlockResolver.writeIndexFileAndCommit(IndexShuffleBlockResolver.scala:142)
   at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:128)
   at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
   at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
   at org.apache.spark.scheduler.Task.run(Task.scala:89)
   at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
   at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:745)

Driver stacktrace:
  at 
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)
  at 
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
  at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
  at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
  at scala.Option.foreach(Option.scala:236)
  at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)
  ...
  Cause: java.io.FileNotFoundException: 
/tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/2f/shuffle_49_0_0.index.cea4ee67-ee4a-48c8-a29e-6a841c6ed1c8
 (No such file or directory)
  at java.io.FileOutputStream.open(Native Method)
  at java.io.FileOutputStream.<init>(FileOutputStream.java:221)
  at java.io.FileOutputStream.<init>(FileOutputStream.java:171)
  at 
org.apache.spark.shuffle.IndexShuffleBlockResolver.writeIndexFileAndCommit(IndexShuffleBlockResolver.scala:142)
  at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:128)
  at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  at org.apache.spark.scheduler.Task.run(Task.scala:89)
  at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
  at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
  ...
[INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver
[WARN] Unable to create class GPUMMul: attempting OpenMP version
[INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver
org.apache.mahout.viennacl.openmp.OMPMMul$
[INFO] Unable to create class OMPMMul: falling back to java version
in-core mul ms: 1498
- A'B, bigger *** FAILED ***
  org.apache.spark.SparkException: Job aborted due to stage failure: Task 
0 in stage 127.0 failed 1 times, most recent failure: Lost task 0.0 in stage 
127.0 (TID 329, localhost): java.io.FileNotFoundException: 
/tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/06/temp_shuffle_27b7da17-0df2-48ae-8277-43460317ed34
 (No such file or directory)
   at java.io.FileOutputStream.open(Native Method)
   at java.io.FileOutputStream.<init>(FileOutputStream.java:221)
   at 
org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:88)
   at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:140)
   at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
   at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
   at org.apache.spark.scheduler.Task.run(Task.scala:89)
   at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
   at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:745)

Driver stacktrace:
  at 
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)
  at 
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
  at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
  at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
  at scala.Option.foreach(Option.scala:236)
  at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)
  ...
  Cause: java.io.FileNotFoundException: 
/tmp/blockmgr-074cfbcd-0dda-4581-af88-284104f3529d/06/temp_shuffle_27b7da17-0df2-48ae-8277-43460317ed34
 (No such file or directory)
  at java.io.FileOutputStream.open(Native Method)
  at java.io.FileOutputStream.<init>(FileOutputStream.java:221)
  at 
org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:88)
  at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:140)
  at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  at org.apache.spark.scheduler.Task.run(Task.scala:89)
  at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
  at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
  at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
  ...
[INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver
[WARN] Unable to create class GPUMMul: attempting OpenMP version
[INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver
org.apache.mahout.viennacl.openmp.OMPMMul$
[INFO] Unable to create class OMPMMul: falling back to java version
[INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver
[WARN] Unable to create class GPUMMul: attempting OpenMP version
[INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver
org.apache.mahout.viennacl.openmp.OMPMMul$
[INFO] Unable to create class OMPMMul: falling back to java version
[INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver
[WARN] Unable to create class GPUMMul: attempting OpenMP version
[INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver
org.apache.mahout.viennacl.openmp.OMPMMul$
[INFO] Unable to create class OMPMMul: falling back to java version
[INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver
[WARN] Unable to create class GPUMMul: attempting OpenMP version
[INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver
org.apache.mahout.viennacl.openmp.OMPMMul$
[INFO] Unable to create class OMPMMul: falling back to java version
[INFO] Creating org.apache.mahout.viennacl.opencl.GPUMMul solver
[WARN] Unable to create class GPUMMul: attempting OpenMP version
[INFO] Creating org.apache.mahout.viennacl.openmp.OMPMMul solver
org.apache.mahout.viennacl.openmp.OMPMMul$
[INFO] Unable to create class OMPMMul: falling back to java version
- C = At %*% B , zippable
TextDelimitedReaderWriterSuite:
- indexedDatasetDFSRead should read sparse matrix file with null rows
PreprocessorSuite:
OpMapBlock(org.apache.mahout.sparkbindings.drm.CheckpointedDrmSpark@70b9d2ee,<function1>,8,-1,true)
{1:3.0,2:5.0,3:6.0}
- asfactor test
{0:2.0,1:5.0,2:-4.0}
{0:0.8164965809277263,1:3.2659863237109037,2:8.286535263104035}
- standard scaler test
- mean center test
TFIDFSparkTestSuite:
- TF test
- TFIDF test
- MLlib TFIDF test
Run completed in 1 minute, 39 seconds.
Total number of tests run: 128
Suites: completed 18, aborted 0
Tests: succeeded 119, failed 9, canceled 0, ignored 1, pending 0
*** 9 TESTS FAILED ***
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Skipping Apache Mahout
[INFO] This project has been banned from the build due to previous failures.
[INFO] ------------------------------------------------------------------------
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Skipping Mahout Build Tools
[INFO] This project has been banned from the build due to previous failures.
[INFO] ------------------------------------------------------------------------
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Mahout Build Tools ................................. SUCCESS [  3.068 s]
[INFO] Apache Mahout ...................................... SUCCESS [  0.176 s]
[INFO] Mahout Math ........................................ SUCCESS [01:13 min]
[INFO] Mahout HDFS ........................................ SUCCESS [  5.306 s]
[INFO] Mahout Map-Reduce .................................. SUCCESS [12:12 min]
[INFO] Mahout Integration ................................. SUCCESS [ 45.598 s]
[INFO] Mahout Examples .................................... SUCCESS [ 25.099 s]
[INFO] Mahout Math Scala bindings ......................... SUCCESS [05:10 min]
[INFO] Mahout Spark bindings .............................. FAILURE [02:34 min]
[INFO] Mahout Flink bindings .............................. SKIPPED
[INFO] Mahout Release Package ............................. SKIPPED
[INFO] Mahout H2O backend ................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 22:32 min
[INFO] Finished at: 2017-02-26T00:36:46+00:00
[INFO] Final Memory: 65M/606M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.scalatest:scalatest-maven-plugin:1.0:test 
(test) on project mahout-spark_2.10: There are test failures -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :mahout-spark_2.10
Build step 'Invoke top-level Maven targets' marked build as failure
[PMD] Skipping publisher since build result is FAILURE
[TASKS] Skipping publisher since build result is FAILURE
Archiving artifacts
Compressed 153.92 MB of artifacts by 89.6% relative to #3418
Recording test results
Publishing Javadoc
[JIRA] Updating issue MAHOUT-1912

Reply via email to