Gehel has uploaded a new change for review. ( 
https://gerrit.wikimedia.org/r/382014 )

Change subject: Fix spotbugs warnings on ...-plugin module
......................................................................

Fix spotbugs warnings on ...-plugin module

Change-Id: Ibacfe655b04a5c7a98d8ef89e0648d27e5886985
---
M 
experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/ExperimentalHighlighter.java
M 
experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/FieldWrapper.java
2 files changed, 24 insertions(+), 25 deletions(-)


  git pull ssh://gerrit.wikimedia.org:29418/search/highlighter 
refs/changes/14/382014/1

diff --git 
a/experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/ExperimentalHighlighter.java
 
b/experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/ExperimentalHighlighter.java
index d48fee2..e98e705 100644
--- 
a/experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/ExperimentalHighlighter.java
+++ 
b/experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/ExperimentalHighlighter.java
@@ -1,5 +1,8 @@
 package org.wikimedia.highlighter.experimental.elasticsearch;
 
+import static java.lang.Boolean.FALSE;
+import static java.lang.Boolean.TRUE;
+
 import java.io.IOException;
 import java.util.ArrayList;
 import java.util.Collections;
@@ -161,7 +164,7 @@
             }
             segmenter = new DelayedSegmenter(defaultField);
             List<Snippet> snippets = buildChooser().choose(segmenter, 
buildHitEnum(), numberOfSnippets);
-            if (snippets.size() != 0) {
+            if (!snippets.isEmpty()) {
                 cache.lastMatched = true;
                 return new HighlightField(context.fieldName, 
formatSnippets(snippets));
             }
@@ -249,8 +252,8 @@
             if (weigher != null) {
                 return;
             }
-            boolean phraseAsTerms = getOption("phrase_as_terms", false);
-            boolean removeHighFrequencyTermsFromCommonTerms = 
getOption("remove_high_freq_terms_from_common_terms", true);
+            boolean phraseAsTerms = getOption("phrase_as_terms", FALSE);
+            boolean removeHighFrequencyTermsFromCommonTerms = 
getOption("remove_high_freq_terms_from_common_terms", TRUE);
             int maxExpandedTerms = getOption("max_expanded_terms", 1024);
             // TODO simplify
             QueryCacheKey key = new QueryCacheKey(context.query, 
maxExpandedTerms, phraseAsTerms,
@@ -280,7 +283,7 @@
             // analyzers that make overlaps.
             e = new OverlapMergingHitEnumWrapper(e);
 
-            if (getOption(OPTION_RETURN_DEBUG_GRAPH, false)) {
+            if (getOption(OPTION_RETURN_DEBUG_GRAPH, FALSE)) {
                 e = new GraphvizHitEnum(e);
             }
             return e;
@@ -322,7 +325,7 @@
 
             List<HitEnum> hitEnums = new ArrayList<>();
             List<String> fieldValues = defaultField.getFieldValues();
-            if (fieldValues.size() == 0) {
+            if (fieldValues.isEmpty()) {
                 return hitEnums;
             }
 
@@ -451,7 +454,7 @@
                 }
                 extraFields.add(wrapper);
             }
-            if (hitEnums.size() == 0) {
+            if (hitEnums.isEmpty()) {
                 return Collections.emptyList();
             }
             return hitEnums;
@@ -459,7 +462,7 @@
 
         private SnippetChooser buildChooser() {
             HitBuilder hitBuilder = Snippet.DEFAULT_HIT_BUILDER;
-            if (getOption(OPTION_RETURN_DEBUG_GRAPH, false)) {
+            if (getOption(OPTION_RETURN_DEBUG_GRAPH, FALSE)) {
                 hitBuilder = GraphvizHit.GRAPHVIZ_HIT_BUILDER;
             }
             if (context.field.fieldOptions().scoreOrdered()) {
@@ -516,11 +519,11 @@
 
         private Text[] formatSnippets(List<Snippet> snippets) throws 
IOException {
             final SnippetFormatter formatter;
-            if (getOption("return_offsets", false)) {
+            if (getOption("return_offsets", FALSE)) {
                 formatter = new OffsetSnippetFormatter();
-            } else if (getOption(OPTION_RETURN_DEBUG_GRAPH, false)) {
+            } else if (getOption(OPTION_RETURN_DEBUG_GRAPH, FALSE)) {
                 formatter = new 
GraphvizSnippetFormatter(defaultField.buildSourceExtracter());
-            } else if (getOption(OPTION_RETURN_SNIPPETS_WITH_OFFSET, false)) {
+            } else if (getOption(OPTION_RETURN_SNIPPETS_WITH_OFFSET, FALSE)) {
                 formatter = new OffsetAugmenterSnippetFormatter(
                         new SnippetFormatter.Default(
                                 defaultField.buildSourceExtracter(),
diff --git 
a/experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/FieldWrapper.java
 
b/experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/FieldWrapper.java
index d6975d3..a4711b3 100644
--- 
a/experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/FieldWrapper.java
+++ 
b/experimental-highlighter-elasticsearch-plugin/src/main/java/org/wikimedia/highlighter/experimental/elasticsearch/FieldWrapper.java
@@ -1,11 +1,14 @@
 package org.wikimedia.highlighter.experimental.elasticsearch;
 
+import static java.util.stream.Collectors.toList;
+
 import java.io.IOException;
 import java.util.ArrayList;
 import java.util.Iterator;
 import java.util.List;
 import java.util.Map;
 import java.util.TreeMap;
+import java.util.stream.Collectors;
 
 import org.apache.lucene.analysis.Analyzer;
 import org.apache.lucene.analysis.TokenStream;
@@ -105,10 +108,7 @@
         if (values == null) {
             List<Object> objs = HighlightUtils.loadFieldValues(context.field, 
context.mapper,
                     context.context, context.hitContext);
-            values = new ArrayList<String>(objs.size());
-            for (Object obj : objs) {
-                values.add(obj.toString());
-            }
+            values = objs.stream().map(Object::toString).collect(toList());
         }
         return values;
     }
@@ -186,9 +186,7 @@
 
                 // Filter 0 weight hits which pop out from the 
TokenStreamHitEnum,
                 // phrase match misses.
-                e = new WeightFilteredHitEnumWrapper(e, 0f);
-
-                return e;
+                return new WeightFilteredHitEnumWrapper(e, 0f);
             }
         }
         // TODO move this up so we don't have to redo it per matched_field
@@ -211,9 +209,7 @@
 
         // Filter 0 weight hits which pop out from the TokenStreamHitEnum,
         // phrase match misses, and boost_before being used as a filter.
-        e = new WeightFilteredHitEnumWrapper(e, 0f);
-
-        return e;
+        return new WeightFilteredHitEnumWrapper(e, 0f);
     }
 
     private HitEnum buildHitEnumForSource() throws IOException {
@@ -262,13 +258,13 @@
     private HitEnum buildPostingsHitEnum() throws IOException {
         return PostingsHitEnum.fromPostings(context.hitContext.reader(),
                 context.hitContext.docId(), context.mapper.fieldType().name(),
-                weigher.acceptableTerms(), getQueryWeigher(false), 
getCorpusWeigher(false), weigher);
+                weigher.acceptableTerms(), getQueryWeigher(), 
getCorpusWeigher(false), weigher);
     }
 
     private HitEnum buildTermVectorsHitEnum() throws IOException {
         return PostingsHitEnum.fromTermVectors(context.hitContext.reader(),
                 context.hitContext.docId(), context.mapper.fieldType().name(),
-                weigher.acceptableTerms(), getQueryWeigher(false), 
getCorpusWeigher(false), weigher);
+                weigher.acceptableTerms(), getQueryWeigher(), 
getCorpusWeigher(false), weigher);
     }
 
     private HitEnum buildTokenStreamHitEnum() throws IOException {
@@ -314,7 +310,7 @@
         }
     }
 
-    private HitEnum buildTokenStreamHitEnum(Analyzer analyzer, String source) 
throws IOException {
+    private HitEnum buildTokenStreamHitEnum(Analyzer analyzer, String source) {
         TokenStream tokenStream;
         try {
             tokenStream = analyzer.tokenStream(context.fieldName, source);
@@ -325,10 +321,10 @@
                     "If analyzing to find hits each matched field must have a 
unique analyzer.", e);
         }
         this.tokenStream = tokenStream;
-        return new TokenStreamHitEnum(tokenStream, getQueryWeigher(true), 
getCorpusWeigher(true), weigher);
+        return new TokenStreamHitEnum(tokenStream, getQueryWeigher(), 
getCorpusWeigher(true), weigher);
     }
 
-    private TermWeigher<BytesRef> getQueryWeigher(boolean 
mightWeighTermsMultipleTimes) {
+    private TermWeigher<BytesRef> getQueryWeigher() {
         return weigher;
     }
 

-- 
To view, visit https://gerrit.wikimedia.org/r/382014
To unsubscribe, visit https://gerrit.wikimedia.org/r/settings

Gerrit-MessageType: newchange
Gerrit-Change-Id: Ibacfe655b04a5c7a98d8ef89e0648d27e5886985
Gerrit-PatchSet: 1
Gerrit-Project: search/highlighter
Gerrit-Branch: master
Gerrit-Owner: Gehel <guillaume.leder...@wikimedia.org>

_______________________________________________
MediaWiki-commits mailing list
MediaWiki-commits@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/mediawiki-commits

Reply via email to