Gehel has uploaded a new change for review. ( https://gerrit.wikimedia.org/r/382014 )
Change subject: Fix spotbugs warnings on ...-plugin module ...................................................................... Fix spotbugs warnings on ...-plugin module Change-Id: Ibacfe655b04a5c7a98d8ef89e0648d27e5886985 --- M experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/ExperimentalHighlighter.java M experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/FieldWrapper.java 2 files changed, 24 insertions(+), 25 deletions(-) git pull ssh://gerrit.wikimedia.org:29418/search/highlighter refs/changes/14/382014/1 diff --git a/experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/ExperimentalHighlighter.java b/experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/ExperimentalHighlighter.java index d48fee2..e98e705 100644 --- a/experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/ExperimentalHighlighter.java +++ b/experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/ExperimentalHighlighter.java @@ -1,5 +1,8 @@ package org.wikimedia.highlighter.experimental.elasticsearch; +import static java.lang.Boolean.FALSE; +import static java.lang.Boolean.TRUE; + import java.io.IOException; import java.util.ArrayList; import java.util.Collections; @@ -161,7 +164,7 @@ } segmenter = new DelayedSegmenter(defaultField); List<Snippet> snippets = buildChooser().choose(segmenter, buildHitEnum(), numberOfSnippets); - if (snippets.size() != 0) { + if (!snippets.isEmpty()) { cache.lastMatched = true; return new HighlightField(context.fieldName, formatSnippets(snippets)); } @@ -249,8 +252,8 @@ if (weigher != null) { return; } - boolean phraseAsTerms = getOption("phrase_as_terms", false); - boolean removeHighFrequencyTermsFromCommonTerms = getOption("remove_high_freq_terms_from_common_terms", true); + boolean phraseAsTerms = getOption("phrase_as_terms", FALSE); + boolean removeHighFrequencyTermsFromCommonTerms = getOption("remove_high_freq_terms_from_common_terms", TRUE); int maxExpandedTerms = getOption("max_expanded_terms", 1024); // TODO simplify QueryCacheKey key = new QueryCacheKey(context.query, maxExpandedTerms, phraseAsTerms, @@ -280,7 +283,7 @@ // analyzers that make overlaps. e = new OverlapMergingHitEnumWrapper(e); - if (getOption(OPTION_RETURN_DEBUG_GRAPH, false)) { + if (getOption(OPTION_RETURN_DEBUG_GRAPH, FALSE)) { e = new GraphvizHitEnum(e); } return e; @@ -322,7 +325,7 @@ List<HitEnum> hitEnums = new ArrayList<>(); List<String> fieldValues = defaultField.getFieldValues(); - if (fieldValues.size() == 0) { + if (fieldValues.isEmpty()) { return hitEnums; } @@ -451,7 +454,7 @@ } extraFields.add(wrapper); } - if (hitEnums.size() == 0) { + if (hitEnums.isEmpty()) { return Collections.emptyList(); } return hitEnums; @@ -459,7 +462,7 @@ private SnippetChooser buildChooser() { HitBuilder hitBuilder = Snippet.DEFAULT_HIT_BUILDER; - if (getOption(OPTION_RETURN_DEBUG_GRAPH, false)) { + if (getOption(OPTION_RETURN_DEBUG_GRAPH, FALSE)) { hitBuilder = GraphvizHit.GRAPHVIZ_HIT_BUILDER; } if (context.field.fieldOptions().scoreOrdered()) { @@ -516,11 +519,11 @@ private Text[] formatSnippets(List<Snippet> snippets) throws IOException { final SnippetFormatter formatter; - if (getOption("return_offsets", false)) { + if (getOption("return_offsets", FALSE)) { formatter = new OffsetSnippetFormatter(); - } else if (getOption(OPTION_RETURN_DEBUG_GRAPH, false)) { + } else if (getOption(OPTION_RETURN_DEBUG_GRAPH, FALSE)) { formatter = new GraphvizSnippetFormatter(defaultField.buildSourceExtracter()); - } else if (getOption(OPTION_RETURN_SNIPPETS_WITH_OFFSET, false)) { + } else if (getOption(OPTION_RETURN_SNIPPETS_WITH_OFFSET, FALSE)) { formatter = new OffsetAugmenterSnippetFormatter( new SnippetFormatter.Default( defaultField.buildSourceExtracter(), diff --git a/experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/FieldWrapper.java b/experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/FieldWrapper.java index d6975d3..a4711b3 100644 --- a/experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/FieldWrapper.java +++ b/experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/FieldWrapper.java @@ -1,11 +1,14 @@ package org.wikimedia.highlighter.experimental.elasticsearch; +import static java.util.stream.Collectors.toList; + import java.io.IOException; import java.util.ArrayList; import java.util.Iterator; import java.util.List; import java.util.Map; import java.util.TreeMap; +import java.util.stream.Collectors; import org.apache.lucene.analysis.Analyzer; import org.apache.lucene.analysis.TokenStream; @@ -105,10 +108,7 @@ if (values == null) { List<Object> objs = HighlightUtils.loadFieldValues(context.field, context.mapper, context.context, context.hitContext); - values = new ArrayList<String>(objs.size()); - for (Object obj : objs) { - values.add(obj.toString()); - } + values = objs.stream().map(Object::toString).collect(toList()); } return values; } @@ -186,9 +186,7 @@ // Filter 0 weight hits which pop out from the TokenStreamHitEnum, // phrase match misses. - e = new WeightFilteredHitEnumWrapper(e, 0f); - - return e; + return new WeightFilteredHitEnumWrapper(e, 0f); } } // TODO move this up so we don't have to redo it per matched_field @@ -211,9 +209,7 @@ // Filter 0 weight hits which pop out from the TokenStreamHitEnum, // phrase match misses, and boost_before being used as a filter. - e = new WeightFilteredHitEnumWrapper(e, 0f); - - return e; + return new WeightFilteredHitEnumWrapper(e, 0f); } private HitEnum buildHitEnumForSource() throws IOException { @@ -262,13 +258,13 @@ private HitEnum buildPostingsHitEnum() throws IOException { return PostingsHitEnum.fromPostings(context.hitContext.reader(), context.hitContext.docId(), context.mapper.fieldType().name(), - weigher.acceptableTerms(), getQueryWeigher(false), getCorpusWeigher(false), weigher); + weigher.acceptableTerms(), getQueryWeigher(), getCorpusWeigher(false), weigher); } private HitEnum buildTermVectorsHitEnum() throws IOException { return PostingsHitEnum.fromTermVectors(context.hitContext.reader(), context.hitContext.docId(), context.mapper.fieldType().name(), - weigher.acceptableTerms(), getQueryWeigher(false), getCorpusWeigher(false), weigher); + weigher.acceptableTerms(), getQueryWeigher(), getCorpusWeigher(false), weigher); } private HitEnum buildTokenStreamHitEnum() throws IOException { @@ -314,7 +310,7 @@ } } - private HitEnum buildTokenStreamHitEnum(Analyzer analyzer, String source) throws IOException { + private HitEnum buildTokenStreamHitEnum(Analyzer analyzer, String source) { TokenStream tokenStream; try { tokenStream = analyzer.tokenStream(context.fieldName, source); @@ -325,10 +321,10 @@ "If analyzing to find hits each matched field must have a unique analyzer.", e); } this.tokenStream = tokenStream; - return new TokenStreamHitEnum(tokenStream, getQueryWeigher(true), getCorpusWeigher(true), weigher); + return new TokenStreamHitEnum(tokenStream, getQueryWeigher(), getCorpusWeigher(true), weigher); } - private TermWeigher<BytesRef> getQueryWeigher(boolean mightWeighTermsMultipleTimes) { + private TermWeigher<BytesRef> getQueryWeigher() { return weigher; } -- To view, visit https://gerrit.wikimedia.org/r/382014 To unsubscribe, visit https://gerrit.wikimedia.org/r/settings Gerrit-MessageType: newchange Gerrit-Change-Id: Ibacfe655b04a5c7a98d8ef89e0648d27e5886985 Gerrit-PatchSet: 1 Gerrit-Project: search/highlighter Gerrit-Branch: master Gerrit-Owner: Gehel <guillaume.leder...@wikimedia.org> _______________________________________________ MediaWiki-commits mailing list MediaWiki-commits@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/mediawiki-commits