(copy-pasta of previous message) another project hosted on our jenkins (e-mission) needs anaconda scipy upgraded from 0.15.1 to 0.17.0. this will also upgrade a few other libs, which i've included at the end of this email.
i've spoken w/josh @ databricks and we don't believe that this will impact the spark builds at all. if this causes serious breakage, i will roll everything back to pre-update. i have created a JIRA issue to look in to creating conda environments for spark builds, something that we should have done long ago: https://issues.apache.org/jira/browse/SPARK-14905 builds will be paused: ~7am PDT anaconda package updates: ~8am jenkins quiet time ends: ~9am at the latest i do not expect the downtime to last very long, and will update this thread w/updates as they come. here's what will be updated under anaconda: The following NEW packages will be INSTALLED: libgfortran: 3.0-0 mkl: 11.3.1-0 wheel: 0.29.0-py27_0 The following packages will be UPDATED: conda: 3.10.1-py27_0 --> 4.0.5-py27_0 conda-env: 2.1.4-py27_0 --> 2.4.5-py27_0 numpy: 1.9.2-py27_0 --> 1.11.0-py27_0 openssl: 1.0.1k-1 --> 1.0.2g-0 pip: 6.1.1-py27_0 --> 8.1.1-py27_1 python: 2.7.9-2 --> 2.7.11-0 pyyaml: 3.11-py27_0 --> 3.11-py27_1 requests: 2.6.0-py27_0 --> 2.9.1-py27_0 scipy: 0.15.1-np19py27_0 --> 0.17.0-np111py27_2 setuptools: 15.0-py27_0 --> 20.7.0-py27_0 sqlite: 3.8.4.1-1 --> 3.9.2-0 yaml: 0.1.4-0 --> 0.1.6-0 --------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org For additional commands, e-mail: dev-h...@spark.apache.org