[ https://issues.apache.org/jira/browse/SPARK-5585?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Patrick Wendell updated SPARK-5585: ----------------------------------- Labels: flaky-test (was: ) > Flaky test: Python regression > ----------------------------- > > Key: SPARK-5585 > URL: https://issues.apache.org/jira/browse/SPARK-5585 > Project: Spark > Issue Type: Bug > Components: MLlib > Affects Versions: 1.3.0 > Reporter: Patrick Wendell > Assignee: Davies Liu > Priority: Critical > Labels: flaky-test > > Hey [~davies] any chance you can take a look at this? The master build is > having random python failures fairly often. Not quite sure what is going on: > {code} > 0inputs+128outputs (0major+13320minor)pagefaults 0swaps > Run mllib tests ... > Running test: pyspark/mllib/classification.py > tput: No value for $TERM and no -T specified > Spark assembly has been built with Hive, including Datanucleus jars on > classpath > 0.43user 0.12system 0:14.85elapsed 3%CPU (0avgtext+0avgdata 94272maxresident)k > 0inputs+280outputs (0major+12627minor)pagefaults 0swaps > Running test: pyspark/mllib/clustering.py > tput: No value for $TERM and no -T specified > Spark assembly has been built with Hive, including Datanucleus jars on > classpath > 0.35user 0.11system 0:12.63elapsed 3%CPU (0avgtext+0avgdata 93568maxresident)k > 0inputs+88outputs (0major+12532minor)pagefaults 0swaps > Running test: pyspark/mllib/feature.py > tput: No value for $TERM and no -T specified > Spark assembly has been built with Hive, including Datanucleus jars on > classpath > 0.28user 0.08system 0:05.73elapsed 6%CPU (0avgtext+0avgdata 93424maxresident)k > 0inputs+32outputs (0major+12548minor)pagefaults 0swaps > Running test: pyspark/mllib/linalg.py > 0.16user 0.05system 0:00.22elapsed 98%CPU (0avgtext+0avgdata > 89888maxresident)k > 0inputs+0outputs (0major+8099minor)pagefaults 0swaps > Running test: pyspark/mllib/rand.py > tput: No value for $TERM and no -T specified > Spark assembly has been built with Hive, including Datanucleus jars on > classpath > 0.25user 0.08system 0:05.42elapsed 6%CPU (0avgtext+0avgdata 87872maxresident)k > 0inputs+0outputs (0major+11849minor)pagefaults 0swaps > Running test: pyspark/mllib/recommendation.py > tput: No value for $TERM and no -T specified > Spark assembly has been built with Hive, including Datanucleus jars on > classpath > 0.32user 0.09system 0:11.42elapsed 3%CPU (0avgtext+0avgdata 94256maxresident)k > 0inputs+32outputs (0major+11797minor)pagefaults 0swaps > Running test: pyspark/mllib/regression.py > tput: No value for $TERM and no -T specified > Spark assembly has been built with Hive, including Datanucleus jars on > classpath > 0.53user 0.17system 0:23.53elapsed 3%CPU (0avgtext+0avgdata 99600maxresident)k > 0inputs+48outputs (0major+12402minor)pagefaults 0swaps > Running test: pyspark/mllib/stat/_statistics.py > tput: No value for $TERM and no -T specified > Spark assembly has been built with Hive, including Datanucleus jars on > classpath > 0.29user 0.09system 0:08.03elapsed 4%CPU (0avgtext+0avgdata 92656maxresident)k > 0inputs+48outputs (0major+12508minor)pagefaults 0swaps > Running test: pyspark/mllib/tree.py > tput: No value for $TERM and no -T specified > Spark assembly has been built with Hive, including Datanucleus jars on > classpath > 0.57user 0.16system 0:25.30elapsed 2%CPU (0avgtext+0avgdata 94400maxresident)k > 0inputs+144outputs (0major+12600minor)pagefaults 0swaps > Running test: pyspark/mllib/util.py > tput: No value for $TERM and no -T specified > Spark assembly has been built with Hive, including Datanucleus jars on > classpath > 0.20user 0.06system 0:08.08elapsed 3%CPU (0avgtext+0avgdata 92768maxresident)k > 0inputs+56outputs (0major+12474minor)pagefaults 0swaps > Running test: pyspark/mllib/tests.py > tput: No value for $TERM and no -T specified > Spark assembly has been built with Hive, including Datanucleus jars on > classpath > .........F/usr/lib64/python2.6/site-packages/numpy/core/fromnumeric.py:2499: > VisibleDeprecationWarning: `rank` is deprecated; use the `ndim` attribute or > function instead. To find the rank of a matrix see `numpy.linalg.matrix_rank`. > VisibleDeprecationWarning) > ./usr/lib64/python2.6/site-packages/numpy/core/fromnumeric.py:2499: > VisibleDeprecationWarning: `rank` is deprecated; use the `ndim` attribute or > function instead. To find the rank of a matrix see `numpy.linalg.matrix_rank`. > VisibleDeprecationWarning) > /usr/lib64/python2.6/site-packages/numpy/core/fromnumeric.py:2499: > VisibleDeprecationWarning: `rank` is deprecated; use the `ndim` attribute or > function instead. To find the rank of a matrix see `numpy.linalg.matrix_rank`. > VisibleDeprecationWarning) > /usr/lib64/python2.6/site-packages/numpy/core/fromnumeric.py:2499: > VisibleDeprecationWarning: `rank` is deprecated; use the `ndim` attribute or > function instead. To find the rank of a matrix see `numpy.linalg.matrix_rank`. > VisibleDeprecationWarning) > /usr/lib64/python2.6/site-packages/numpy/core/fromnumeric.py:2499: > VisibleDeprecationWarning: `rank` is deprecated; use the `ndim` attribute or > function instead. To find the rank of a matrix see `numpy.linalg.matrix_rank`. > VisibleDeprecationWarning) > ./usr/lib64/python2.6/site-packages/numpy/lib/utils.py:95: > DeprecationWarning: `dot` is deprecated! > warnings.warn(depdoc, DeprecationWarning) > ............. > ====================================================================== > FAIL: test_regression (__main__.ListTests) > ---------------------------------------------------------------------- > Traceback (most recent call last): > File "pyspark/mllib/tests.py", line 289, in test_regression > self.assertTrue(rf_model.predict(features[0]) <= 0) > AssertionError: False is not true > ---------------------------------------------------------------------- > Ran 25 tests in 53.798s > {code} > https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-Master-SBT/AMPLAB_JENKINS_BUILD_PROFILE=hadoop1.0,label=centos/1496/console -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org