+1 I watched Gabor working on this and this is a very comprehensive, which also includes testing in downstreamers (HBase and Hive). So very good work.
Thanks! On Wed, Apr 3, 2019 at 3:41 AM Steve Loughran <ste...@cloudera.com.invalid> wrote: > I am taking silence as happiness here. > > +1 to the patch > > On Tue, Apr 2, 2019 at 9:54 AM Steve Loughran <ste...@cloudera.com> wrote: > > > > > I know that the number of guava updates we could call painless is 0, but > > we need to do this. > > > > The last time we successfully updated Guava was 2012: h > > ttps://issues.apache.org/jira/browse/HDFS-3187 > > That was the java 6 era > > > > The last unsuccessful attempt, April 2017: > > https://issues.apache.org/jira/browse/HADOOP-14386 > > > > Let's try again and this time if there are problems say: sorry, but its > > time to move on. > > > > I think we should only worry about branch-3.2+ for now, though the other > > branches could be lined up for those changes needed to ensure that > > everything builds if you explicitly set the version (e.g findbugs > changes. > > Then we can worry about 3.1.x line, which is the 3.x branch most widely > > picked up to date. > > > > I want to avoid branch-2 entirely, though as Gabor notes, I want to move > > us on to java 8 builds there so that people can do a branch-2 build if > they > > need to. > > > > *Is everyone happy with the proposed patch*: > > https://github.com/apache/hadoop/pull/674 > > > > -Steve > > > > > > On Mon, Apr 1, 2019 at 8:35 PM Gabor Bota <gabor.b...@cloudera.com > .invalid> > > wrote: > > > >> Hi devs, > >> > >> I'm working on the guava version from 11.0.2 to 27.0-jre in > >> hadoop-project. > >> We need to do the upgrade because of CVE-2018-10237 > >> <https://nvd.nist.gov/vuln/detail/CVE-2018-10237>. > >> > >> I've created an issue (HADOOP-15960 > >> <https://issues.apache.org/jira/browse/HADOOP-15960>) to track progress > >> and > >> created subtasks for hadoop branches 3.0, 3.1, 3.2 and trunk. The first > >> update should be done in the trunk, and then it can be backported to > lower > >> version branches. Backporting to 2.x is not feasible right now, because > of > >> Guava 20 is the last Java 7 compatible version[1], and we have Java 7 > >> compatibility on version 2 branches - but we are planning to update ( > >> HADOOP-16219 <https://issues.apache.org/jira/browse/HADOOP-16219>). > >> > >> For the new deprecations after the update, I've created another issue ( > >> HADOOP-16222 <https://issues.apache.org/jira/browse/HADOOP-16222>). > Those > >> can be fixed after the update is committed. > >> > >> Unit and integration testing in hadoop trunk > >> There were modifications in the test in the following modules so > >> precommit tests were running on jenkins: > >> > >> - hadoop-common-project > >> - hadoop-hdfs-project > >> - hadoop-mapreduce-project > >> - hadoop-yarn-project > >> > >> There was one failure but after re-running the test locally it was > >> successful, so not related to the change. > >> > >> Because of 5 hour test time limit for jenkins precommit build, I had to > >> run > >> tests on hadoop-tools manually and the tests were successful. You can > find > >> test results for trunk under HADOOP-16210 > >> <https://issues.apache.org/jira/browse/HADOOP-16210>. > >> > >> Integration testing with other components > >> I've done testing with HBase master on hadoop branch-3.0 with guava 27, > >> and > >> the tests were running fine. Thanks to Peter Somogyi for help. > >> We are planning to do some testing with Peter Vary on Hive with > branch-3.1 > >> this week. > >> > >> Thanks, > >> Gabor > >> > >> [1] > >> > >> > https://groups.google.com/forum/#!msg/guava-discuss/ZRmDJnAq9T0/-HExv44eCAAJ > >> > > >