I am taking silence as happiness here.

+1 to the patch

On Tue, Apr 2, 2019 at 9:54 AM Steve Loughran <ste...@cloudera.com> wrote:

>
> I know that the number of guava updates we could call painless is 0, but
> we need to do this.
>
> The last time we successfully updated Guava was 2012: h
> ttps://issues.apache.org/jira/browse/HDFS-3187
> That was the java 6 era
>
> The last unsuccessful attempt, April 2017:
> https://issues.apache.org/jira/browse/HADOOP-14386
>
> Let's try again and this time if there are problems say: sorry, but its
> time to move on.
>
> I think we should only worry about branch-3.2+ for now, though the other
> branches could be lined up for those changes needed to ensure that
> everything builds if you explicitly set the version (e.g findbugs changes.
> Then we can worry about 3.1.x line, which is the 3.x branch most widely
> picked up to date.
>
> I want to avoid branch-2 entirely, though as Gabor notes, I want to move
> us on to java 8 builds there so that people can do a branch-2 build if they
> need to.
>
> *Is everyone happy with the proposed patch*:
> https://github.com/apache/hadoop/pull/674
>
> -Steve
>
>
> On Mon, Apr 1, 2019 at 8:35 PM Gabor Bota <gabor.b...@cloudera.com.invalid>
> wrote:
>
>> Hi devs,
>>
>> I'm working on the guava version from 11.0.2 to 27.0-jre in
>> hadoop-project.
>> We need to do the upgrade because of CVE-2018-10237
>> <https://nvd.nist.gov/vuln/detail/CVE-2018-10237>.
>>
>> I've created an issue (HADOOP-15960
>> <https://issues.apache.org/jira/browse/HADOOP-15960>) to track progress
>> and
>> created subtasks for hadoop branches 3.0, 3.1, 3.2 and trunk. The first
>> update should be done in the trunk, and then it can be backported to lower
>> version branches. Backporting to 2.x is not feasible right now, because of
>> Guava 20 is the last Java 7 compatible version[1], and we have Java 7
>> compatibility on version 2 branches - but we are planning to update (
>> HADOOP-16219 <https://issues.apache.org/jira/browse/HADOOP-16219>).
>>
>> For the new deprecations after the update, I've created another issue (
>> HADOOP-16222 <https://issues.apache.org/jira/browse/HADOOP-16222>). Those
>> can be fixed after the update is committed.
>>
>> Unit and integration testing in hadoop trunk
>> There were modifications in the test in the following modules so
>> precommit tests were running on jenkins:
>>
>>    - hadoop-common-project
>>    - hadoop-hdfs-project
>>    - hadoop-mapreduce-project
>>    - hadoop-yarn-project
>>
>> There was one failure but after re-running the test locally it was
>> successful, so not related to the change.
>>
>> Because of 5 hour test time limit for jenkins precommit build, I had to
>> run
>> tests on hadoop-tools manually and the tests were successful. You can find
>> test results for trunk under HADOOP-16210
>> <https://issues.apache.org/jira/browse/HADOOP-16210>.
>>
>> Integration testing with other components
>> I've done testing with HBase master on hadoop branch-3.0 with guava 27,
>> and
>> the tests were running fine. Thanks to Peter Somogyi for help.
>> We are planning to do some testing with Peter Vary on Hive with branch-3.1
>> this week.
>>
>> Thanks,
>> Gabor
>>
>> [1]
>>
>> https://groups.google.com/forum/#!msg/guava-discuss/ZRmDJnAq9T0/-HExv44eCAAJ
>>
>

Reply via email to