http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/schema/data/partitioner/DataPartitioner.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/schema/data/partitioner/DataPartitioner.java b/src/main/java/org/apache/pirk/schema/data/partitioner/DataPartitioner.java index 1bbecca..cd1e632 100644 --- a/src/main/java/org/apache/pirk/schema/data/partitioner/DataPartitioner.java +++ b/src/main/java/org/apache/pirk/schema/data/partitioner/DataPartitioner.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.schema.data.partitioner; import java.io.Serializable; @@ -35,32 +35,32 @@ public interface DataPartitioner extends Serializable * <p> * If the Object does not have/need a specific type identifier, use null */ - public ArrayList<BigInteger> toPartitions(Object object, String type) throws Exception; + ArrayList<BigInteger> toPartitions(Object object, String type) throws Exception; /** * Method to reconstruct an Object given an ArrayList of its BigInteger partition elements and its type identifier * <p> * If the Object does not have/need a specific type identifier, use null */ - public Object fromPartitions(ArrayList<BigInteger> parts, int partsIndex, String type) throws Exception; + Object fromPartitions(ArrayList<BigInteger> parts, int partsIndex, String type) throws Exception; /** * Method to return the number of bits of an object with the given type */ - public int getBits(String type) throws Exception; + int getBits(String type) throws Exception; /** * Create partitions for an array of the same type of elements - used when a data value field is an array and we wish to encode these into the return value */ - public ArrayList<BigInteger> arrayToPartitions(List<?> elementList, String type) throws Exception; + ArrayList<BigInteger> arrayToPartitions(List<?> elementList, String type) throws Exception; /** * Method to get an empty set of partitions by data type - used for padding return array values */ - public ArrayList<BigInteger> getPaddedPartitions(String type) throws Exception; + ArrayList<BigInteger> getPaddedPartitions(String type) throws Exception; /** * Method to get the number of partitions of the data object given the type */ - public int getNumPartitions(String type) throws Exception; + int getNumPartitions(String type) throws Exception; }
http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/schema/data/partitioner/IPDataPartitioner.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/schema/data/partitioner/IPDataPartitioner.java b/src/main/java/org/apache/pirk/schema/data/partitioner/IPDataPartitioner.java index 25e49a4..494aba0 100644 --- a/src/main/java/org/apache/pirk/schema/data/partitioner/IPDataPartitioner.java +++ b/src/main/java/org/apache/pirk/schema/data/partitioner/IPDataPartitioner.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.schema.data.partitioner; import java.math.BigInteger; @@ -36,7 +36,7 @@ public class IPDataPartitioner implements DataPartitioner @Override public ArrayList<BigInteger> toPartitions(Object object, String type) throws Exception { - ArrayList<BigInteger> parts = new ArrayList<BigInteger>(); + ArrayList<BigInteger> parts = new ArrayList<>(); String[] octets = ((String) object).split("\\."); for (String oct : octets) @@ -50,7 +50,7 @@ public class IPDataPartitioner implements DataPartitioner @Override public Object fromPartitions(ArrayList<BigInteger> parts, int partsIndex, String type) throws Exception { - Object element = null; + Object element; element = parts.get(partsIndex).toString() + "." + parts.get(partsIndex + 1).toString() + "." + parts.get(partsIndex + 2).toString() + "." + parts.get(partsIndex + 3).toString(); @@ -67,7 +67,7 @@ public class IPDataPartitioner implements DataPartitioner @Override public ArrayList<BigInteger> getPaddedPartitions(String type) throws Exception { - ArrayList<BigInteger> parts = new ArrayList<BigInteger>(); + ArrayList<BigInteger> parts = new ArrayList<>(); for (int i = 0; i < 4; ++i) { @@ -82,7 +82,7 @@ public class IPDataPartitioner implements DataPartitioner @Override public ArrayList<BigInteger> arrayToPartitions(List<?> elementList, String type) throws Exception { - ArrayList<BigInteger> parts = new ArrayList<BigInteger>(); + ArrayList<BigInteger> parts = new ArrayList<>(); int numArrayElementsToReturn = Integer.parseInt(SystemConfiguration.getProperty("pir.numReturnArrayElements", "1")); for (int i = 0; i < numArrayElementsToReturn; ++i) http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/schema/data/partitioner/ISO8601DatePartitioner.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/schema/data/partitioner/ISO8601DatePartitioner.java b/src/main/java/org/apache/pirk/schema/data/partitioner/ISO8601DatePartitioner.java index 4bd29b6..715bf15 100644 --- a/src/main/java/org/apache/pirk/schema/data/partitioner/ISO8601DatePartitioner.java +++ b/src/main/java/org/apache/pirk/schema/data/partitioner/ISO8601DatePartitioner.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.schema.data.partitioner; import java.math.BigInteger; @@ -33,7 +33,7 @@ public class ISO8601DatePartitioner implements DataPartitioner { private static final long serialVersionUID = 1L; - PrimitiveTypePartitioner ptp = null; + private PrimitiveTypePartitioner ptp = null; public ISO8601DatePartitioner() { http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/schema/data/partitioner/PrimitiveTypePartitioner.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/schema/data/partitioner/PrimitiveTypePartitioner.java b/src/main/java/org/apache/pirk/schema/data/partitioner/PrimitiveTypePartitioner.java index 1c3f42d..3688e81 100644 --- a/src/main/java/org/apache/pirk/schema/data/partitioner/PrimitiveTypePartitioner.java +++ b/src/main/java/org/apache/pirk/schema/data/partitioner/PrimitiveTypePartitioner.java @@ -18,16 +18,16 @@ */ package org.apache.pirk.schema.data.partitioner; -import org.apache.http.util.ByteArrayBuffer; -import org.apache.log4j.Logger; -import org.apache.pirk.utils.LogUtils; -import org.apache.pirk.utils.SystemConfiguration; - import java.math.BigInteger; import java.nio.ByteBuffer; import java.util.ArrayList; import java.util.List; +import org.apache.http.util.ByteArrayBuffer; +import org.apache.pirk.utils.SystemConfiguration; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + /** * Class for partitioning objects with primitive Java types * @@ -36,7 +36,7 @@ public class PrimitiveTypePartitioner implements DataPartitioner { private static final long serialVersionUID = 1L; - private Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(PrimitiveTypePartitioner.class); public static final String BYTE = "byte"; public static final String SHORT = "short"; http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/schema/query/LoadQuerySchemas.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/schema/query/LoadQuerySchemas.java b/src/main/java/org/apache/pirk/schema/query/LoadQuerySchemas.java index 213e983..fa00756 100644 --- a/src/main/java/org/apache/pirk/schema/query/LoadQuerySchemas.java +++ b/src/main/java/org/apache/pirk/schema/query/LoadQuerySchemas.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.schema.query; import java.io.File; @@ -29,12 +29,12 @@ import javax.xml.parsers.DocumentBuilderFactory; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; -import org.apache.log4j.Logger; import org.apache.pirk.schema.data.DataSchema; import org.apache.pirk.schema.data.LoadDataSchemas; import org.apache.pirk.schema.data.partitioner.DataPartitioner; -import org.apache.pirk.utils.LogUtils; import org.apache.pirk.utils.SystemConfiguration; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.w3c.dom.Document; import org.w3c.dom.Element; import org.w3c.dom.Node; @@ -66,15 +66,15 @@ import org.w3c.dom.NodeList; */ public class LoadQuerySchemas { - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(LoadQuerySchemas.class); - public static HashMap<String,QuerySchema> schemaMap; + private static HashMap<String,QuerySchema> schemaMap; static { logger.info("Loading query schemas: "); - schemaMap = new HashMap<String,QuerySchema>(); + schemaMap = new HashMap<>(); try { initialize(); @@ -124,13 +124,13 @@ public class LoadQuerySchemas private static QuerySchema loadQuerySchemaFile(String schemaFile, boolean hdfs, FileSystem fs) throws Exception { - QuerySchema querySchema = null; + QuerySchema querySchema; DocumentBuilderFactory dbFactory = DocumentBuilderFactory.newInstance(); DocumentBuilder dBuilder = dbFactory.newDocumentBuilder(); // Read in and parse the schema file - Document doc = null; + Document doc; if (hdfs) { Path filePath = new Path(schemaFile); @@ -176,7 +176,7 @@ public class LoadQuerySchemas } Element elements = (Element) elementsList.item(0); - TreeSet<String> elementNames = new TreeSet<String>(); + TreeSet<String> elementNames = new TreeSet<>(); int dataElementSize = 0; NodeList nList = elements.getElementsByTagName("name"); for (int i = 0; i < nList.getLength(); i++) @@ -193,7 +193,7 @@ public class LoadQuerySchemas // Compute the number of bits for this element logger.info("name = " + name); logger.info("partitionerName = " + dataSchema.getPartitionerName(name)); - if (((DataPartitioner) dataSchema.getPartitionerForElement(name)) == null) + if ((dataSchema.getPartitionerForElement(name)) == null) { logger.info("partitioner is null"); } @@ -218,7 +218,7 @@ public class LoadQuerySchemas } // Extract the filterNames, if they exist - HashSet<String> filterNamesSet = new HashSet<String>(); + HashSet<String> filterNamesSet = new HashSet<>(); if (doc.getElementsByTagName("filterNames").item(0) != null) { NodeList filterNamesList = doc.getElementsByTagName("filterNames"); @@ -256,7 +256,7 @@ public class LoadQuerySchemas */ private static String extractValue(Document doc, String valueName) throws Exception { - String value = null; + String value; NodeList itemList = doc.getElementsByTagName(valueName); if (itemList.getLength() > 1) http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/schema/query/QuerySchema.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/schema/query/QuerySchema.java b/src/main/java/org/apache/pirk/schema/query/QuerySchema.java index b3ea515..7610b52 100644 --- a/src/main/java/org/apache/pirk/schema/query/QuerySchema.java +++ b/src/main/java/org/apache/pirk/schema/query/QuerySchema.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.schema.query; import java.io.Serializable; @@ -35,24 +35,24 @@ public class QuerySchema implements Serializable { private static final long serialVersionUID = 1L; - public static final String NO_FILTER = "noFilter"; + static final String NO_FILTER = "noFilter"; - String schemaName = null; + private String schemaName = null; - String dataSchemaName = null; // name of the DataSchema for this query schema + private String dataSchemaName = null; // name of the DataSchema for this query schema - TreeSet<String> elementNames = null; // names of elements in the data schema to + private TreeSet<String> elementNames = null; // names of elements in the data schema to // include in the response, order matters for packing/unpacking - String filter = null; // name of filter class to use in data filtering + private String filter = null; // name of filter class to use in data filtering - Object filterObj = null; // instance of the filter + private Object filterObj = null; // instance of the filter - HashSet<String> filterElementNames = null; // set of element names to apply filtering in pre-processing + private HashSet<String> filterElementNames = null; // set of element names to apply filtering in pre-processing - String selectorName = null; // name of element in the dataSchema to be used as the selector + private String selectorName = null; // name of element in the dataSchema to be used as the selector - int dataElementSize = 0; // total number of bits to be returned for each data element hit + private int dataElementSize = 0; // total number of bits to be returned for each data element hit public QuerySchema(String schemaNameInput, String dataSchemaNameInput, TreeSet<String> elementNamesInput, String selectorNameInput, int dataElementSizeInput, HashSet<String> filterElementNamesInput, String filterIn) throws Exception http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/schema/query/filter/DataFilter.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/schema/query/filter/DataFilter.java b/src/main/java/org/apache/pirk/schema/query/filter/DataFilter.java index 17f387d..35c64e0 100644 --- a/src/main/java/org/apache/pirk/schema/query/filter/DataFilter.java +++ b/src/main/java/org/apache/pirk/schema/query/filter/DataFilter.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.schema.query.filter; import java.io.Serializable; http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/schema/query/filter/FilterFactory.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/schema/query/filter/FilterFactory.java b/src/main/java/org/apache/pirk/schema/query/filter/FilterFactory.java index 8c86235..946f803 100644 --- a/src/main/java/org/apache/pirk/schema/query/filter/FilterFactory.java +++ b/src/main/java/org/apache/pirk/schema/query/filter/FilterFactory.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.schema.query.filter; import java.io.BufferedReader; @@ -44,12 +44,12 @@ public class FilterFactory FileSystem fs = FileSystem.get(new Configuration()); // Grab the stopList - HashSet<String> stopList = new HashSet<String>(); + HashSet<String> stopList = new HashSet<>(); String stopListFile = SystemConfiguration.getProperty("pir.stopListFile", "none"); if (!stopListFile.equals("none")) { - BufferedReader br = null; + BufferedReader br; if (fs.exists(new Path(stopListFile))) { br = new BufferedReader(new InputStreamReader(fs.open(new Path(stopListFile)))); @@ -60,7 +60,7 @@ public class FilterFactory br = new BufferedReader(fr); } - String qLine = null; + String qLine; while ((qLine = br.readLine()) != null) { stopList.add(qLine); http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/schema/query/filter/StopListFilter.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/schema/query/filter/StopListFilter.java b/src/main/java/org/apache/pirk/schema/query/filter/StopListFilter.java index 1f458c1..6c58d2c 100644 --- a/src/main/java/org/apache/pirk/schema/query/filter/StopListFilter.java +++ b/src/main/java/org/apache/pirk/schema/query/filter/StopListFilter.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.schema.query.filter; import java.util.Arrays; @@ -24,11 +24,11 @@ import java.util.List; import org.apache.hadoop.io.ArrayWritable; import org.apache.hadoop.io.MapWritable; -import org.apache.log4j.Logger; import org.apache.pirk.schema.data.DataSchema; -import org.apache.pirk.utils.LogUtils; import org.apache.pirk.utils.StopListUtils; import org.elasticsearch.hadoop.mr.WritableArrayWritable; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * Filter class to filter data elements based upon a stoplist applied to specified field elements @@ -37,10 +37,10 @@ public class StopListFilter implements DataFilter { private static final long serialVersionUID = 1L; - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(StopListFilter.class); - HashSet<String> filterSet = null; - HashSet<String> stopList = null; + private HashSet<String> filterSet = null; + private HashSet<String> stopList = null; public StopListFilter(HashSet<String> filterSetIn, HashSet<String> stopListIn) { http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/schema/response/QueryResponseJSON.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/schema/response/QueryResponseJSON.java b/src/main/java/org/apache/pirk/schema/response/QueryResponseJSON.java index 5b3c108..5f7e79c 100644 --- a/src/main/java/org/apache/pirk/schema/response/QueryResponseJSON.java +++ b/src/main/java/org/apache/pirk/schema/response/QueryResponseJSON.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.schema.response; import java.io.Serializable; @@ -25,15 +25,15 @@ import java.util.HashSet; import java.util.Set; import org.apache.hadoop.io.Text; -import org.apache.log4j.Logger; import org.apache.pirk.query.wideskies.QueryInfo; import org.apache.pirk.schema.data.DataSchema; import org.apache.pirk.schema.data.LoadDataSchemas; import org.apache.pirk.schema.query.LoadQuerySchemas; import org.apache.pirk.schema.query.QuerySchema; -import org.apache.pirk.utils.LogUtils; import org.json.simple.JSONObject; import org.json.simple.JSONValue; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * JSON helper class for query results @@ -44,13 +44,13 @@ public class QueryResponseJSON implements Serializable { private static final long serialVersionUID = 1L; - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(QueryResponseJSON.class); - JSONObject jsonObj = null; + private JSONObject jsonObj = null; - DataSchema dSchema = null; + private DataSchema dSchema = null; - QueryInfo queryInfo = null; + private QueryInfo queryInfo = null; public static final String EVENT_TYPE = "event_type"; // notification type the matched the record public static final Text EVENT_TYPE_TEXT = new Text(EVENT_TYPE); @@ -131,7 +131,7 @@ public class QueryResponseJSON implements Serializable HashSet<String> schemaListRep = dSchema.getListRep(); for (String key : schemaListRep) { - jsonObj.put(key, new ArrayList<Object>()); + jsonObj.put(key, new ArrayList<>()); } } @@ -152,10 +152,10 @@ public class QueryResponseJSON implements Serializable { if (!(val instanceof ArrayList)) { - ArrayList<Object> list = null; + ArrayList<Object> list; if (!jsonObj.containsKey(key)) { - list = new ArrayList<Object>(); + list = new ArrayList<>(); jsonObj.put(key, list); } list = (ArrayList<Object>) jsonObj.get(key); http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/test/distributed/DistributedTestCLI.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/test/distributed/DistributedTestCLI.java b/src/main/java/org/apache/pirk/test/distributed/DistributedTestCLI.java index 977fa5e..1535e1f 100644 --- a/src/main/java/org/apache/pirk/test/distributed/DistributedTestCLI.java +++ b/src/main/java/org/apache/pirk/test/distributed/DistributedTestCLI.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.test.distributed; import org.apache.commons.cli.CommandLine; @@ -24,18 +24,18 @@ import org.apache.commons.cli.GnuParser; import org.apache.commons.cli.HelpFormatter; import org.apache.commons.cli.Option; import org.apache.commons.cli.Options; -import org.apache.log4j.Logger; -import org.apache.pirk.utils.LogUtils; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * A container for Apache's Command Line Interface that contains custom functionality for the MapReduce functional tests. */ public class DistributedTestCLI { - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(DistributedTestCLI.class); - CommandLine commandLine = null; - Options cliOptions = null; + private CommandLine commandLine = null; + private Options cliOptions = null; /** * Create and parse allowable options http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/test/distributed/DistributedTestDriver.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/test/distributed/DistributedTestDriver.java b/src/main/java/org/apache/pirk/test/distributed/DistributedTestDriver.java index f567bab..d822e81 100755 --- a/src/main/java/org/apache/pirk/test/distributed/DistributedTestDriver.java +++ b/src/main/java/org/apache/pirk/test/distributed/DistributedTestDriver.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,22 +15,22 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.test.distributed; import java.util.ArrayList; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; -import org.apache.log4j.Logger; import org.apache.pirk.schema.data.LoadDataSchemas; import org.apache.pirk.schema.query.LoadQuerySchemas; import org.apache.pirk.schema.query.filter.StopListFilter; import org.apache.pirk.test.distributed.testsuite.DistTestSuite; import org.apache.pirk.test.utils.Inputs; -import org.apache.pirk.utils.LogUtils; import org.apache.pirk.utils.SystemConfiguration; import org.json.simple.JSONObject; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * Driver class to run the suite of functional tests for MR and Spark PIR @@ -38,7 +38,7 @@ import org.json.simple.JSONObject; */ public class DistributedTestDriver { - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(DistributedTestDriver.class); // Input public static final String JSON_PIR_INPUT_FILE_PROPERTY = "test.pir.inputJSONFile"; http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/test/distributed/testsuite/DistTestSuite.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/test/distributed/testsuite/DistTestSuite.java b/src/main/java/org/apache/pirk/test/distributed/testsuite/DistTestSuite.java index dec00f9..37cb43c 100644 --- a/src/main/java/org/apache/pirk/test/distributed/testsuite/DistTestSuite.java +++ b/src/main/java/org/apache/pirk/test/distributed/testsuite/DistTestSuite.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.test.distributed.testsuite; import java.io.File; @@ -24,7 +24,6 @@ import java.util.ArrayList; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.util.ToolRunner; -import org.apache.log4j.Logger; import org.apache.pirk.encryption.Paillier; import org.apache.pirk.inputformat.hadoop.InputFormatConst; import org.apache.pirk.inputformat.hadoop.json.JSONInputFormatBase; @@ -41,10 +40,11 @@ import org.apache.pirk.test.distributed.DistributedTestDriver; import org.apache.pirk.test.utils.BaseTests; import org.apache.pirk.test.utils.Inputs; import org.apache.pirk.test.utils.TestUtils; -import org.apache.pirk.utils.LogUtils; import org.apache.pirk.utils.SystemConfiguration; import org.apache.spark.launcher.SparkLauncher; import org.json.simple.JSONObject; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * Distributed test class for PIR @@ -52,7 +52,7 @@ import org.json.simple.JSONObject; */ public class DistTestSuite { - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(DistTestSuite.class); // This method also tests all non-query specific configuration options/properties // for the MapReduce version of PIR @@ -299,7 +299,7 @@ public class DistTestSuite SystemConfiguration.setProperty("pir.numReduceTasks", "1"); SystemConfiguration.setProperty("pir.stopListFile", SystemConfiguration.getProperty(DistributedTestDriver.PIR_STOPLIST_FILE)); - ArrayList<QueryResponseJSON> results = null; + ArrayList<QueryResponseJSON> results; // Create the temp result file File fileFinalResults = File.createTempFile("finalResultsFile", ".txt"); @@ -361,7 +361,7 @@ public class DistTestSuite // Build args String inputFormat = SystemConfiguration.getProperty("pir.dataInputFormat"); logger.info("inputFormat = " + inputFormat); - ArrayList<String> args = new ArrayList<String>(); + ArrayList<String> args = new ArrayList<>(); args.add("-" + ResponderCLI.PLATFORM + "=spark"); args.add("-" + ResponderCLI.DATAINPUTFORMAT + "=" + inputFormat); args.add("-" + ResponderCLI.QUERYINPUT + "=" + SystemConfiguration.getProperty("pir.queryInput")); http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/test/utils/BaseTests.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/test/utils/BaseTests.java b/src/main/java/org/apache/pirk/test/utils/BaseTests.java index 621b64e..51497d6 100644 --- a/src/main/java/org/apache/pirk/test/utils/BaseTests.java +++ b/src/main/java/org/apache/pirk/test/utils/BaseTests.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,41 +15,40 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.test.utils; -import static org.junit.Assert.fail; - import java.util.ArrayList; import java.util.Arrays; import java.util.HashSet; import java.util.Set; import org.apache.hadoop.fs.FileSystem; -import org.apache.log4j.Logger; import org.apache.pirk.query.wideskies.QueryUtils; import org.apache.pirk.schema.response.QueryResponseJSON; import org.apache.pirk.test.distributed.testsuite.DistTestSuite; -import org.apache.pirk.utils.LogUtils; import org.apache.pirk.utils.StringUtils; import org.apache.pirk.utils.SystemConfiguration; import org.json.simple.JSONArray; import org.json.simple.JSONObject; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; +import static org.junit.Assert.fail; /** * Class to hold the base functional distributed tests */ public class BaseTests { - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(BaseTests.class); public static double queryNum = 1.0; public static int dataPartitionBitSize = 8; // Selectors for domain and IP queries, queryNum is the first entry for file generation - public static ArrayList<String> selectorsDomain = new ArrayList<String>(Arrays.asList("s.t.u.net", "d.e.com", "r.r.r.r", "a.b.c.com", "something.else", + private static ArrayList<String> selectorsDomain = new ArrayList<>(Arrays.asList("s.t.u.net", "d.e.com", "r.r.r.r", "a.b.c.com", "something.else", "x.y.net")); - public static ArrayList<String> selectorsIP = new ArrayList<String>(Arrays.asList("55.55.55.55", "5.6.7.8", "10.20.30.40", "13.14.15.16", "21.22.23.24")); + private static ArrayList<String> selectorsIP = new ArrayList<>(Arrays.asList("55.55.55.55", "5.6.7.8", "10.20.30.40", "13.14.15.16", "21.22.23.24")); // Encryption variables -- Paillier mechanisms are tested in the Paillier test code, so these are fixed... public static int hashBitSize = 12; @@ -75,7 +74,7 @@ public class BaseTests logger.info("Running testDNSHostnameQuery(): "); int numExpectedResults = 6; - ArrayList<QueryResponseJSON> results = null; + ArrayList<QueryResponseJSON> results; if (isDistributed) { results = DistTestSuite.performQuery(Inputs.DNS_HOSTNAME_QUERY, selectorsDomain, fs, isSpark, numThreads); @@ -100,12 +99,12 @@ public class BaseTests } // Check that each qname appears once in the result set - HashSet<String> correctQnames = new HashSet<String>(); + HashSet<String> correctQnames = new HashSet<>(); correctQnames.add("a.b.c.com"); correctQnames.add("d.e.com"); correctQnames.add("something.else"); - HashSet<String> resultQnames = new HashSet<String>(); + HashSet<String> resultQnames = new HashSet<>(); for (QueryResponseJSON qrJSON : results) { resultQnames.add((String) qrJSON.getValue(Inputs.QNAME)); @@ -138,7 +137,7 @@ public class BaseTests removeTailElements = 3; } - ArrayList<QueryResponseJSON> correctResults = new ArrayList<QueryResponseJSON>(); + ArrayList<QueryResponseJSON> correctResults = new ArrayList<>(); int i = 0; while (i < (dataElements.size() - removeTailElements)) { @@ -197,7 +196,7 @@ public class BaseTests { logger.info("Running testDNSIPQuery(): "); - ArrayList<QueryResponseJSON> results = null; + ArrayList<QueryResponseJSON> results; if (isDistributed) { results = DistTestSuite.performQuery(Inputs.DNS_IP_QUERY, selectorsIP, fs, isSpark, numThreads); @@ -218,7 +217,7 @@ public class BaseTests } printResultList(results); - ArrayList<QueryResponseJSON> correctResults = new ArrayList<QueryResponseJSON>(); + ArrayList<QueryResponseJSON> correctResults = new ArrayList<>(); int i = 0; while (i < (dataElements.size() - 3)) // last three data elements not hit - one on stoplist, two don't match selectors { @@ -270,7 +269,7 @@ public class BaseTests { logger.info("Running testDNSNXDOMAINQuery(): "); - ArrayList<QueryResponseJSON> results = null; + ArrayList<QueryResponseJSON> results; if (isDistributed) { results = DistTestSuite.performQuery(Inputs.DNS_NXDOMAIN_QUERY, selectorsDomain, fs, isSpark, numThreads); @@ -286,7 +285,7 @@ public class BaseTests fail("results.size() = " + results.size() + " -- must equal 1"); } - ArrayList<QueryResponseJSON> correctResults = new ArrayList<QueryResponseJSON>(); + ArrayList<QueryResponseJSON> correctResults = new ArrayList<>(); int i = 0; while (i < dataElements.size()) { @@ -332,7 +331,7 @@ public class BaseTests { logger.info("Running testSRCIPQuery(): "); - ArrayList<QueryResponseJSON> results = null; + ArrayList<QueryResponseJSON> results; int removeTailElements = 0; int numExpectedResults = 1; if (isDistributed) @@ -352,7 +351,7 @@ public class BaseTests fail("results.size() = " + results.size() + " -- must equal " + numExpectedResults); } - ArrayList<QueryResponseJSON> correctResults = new ArrayList<QueryResponseJSON>(); + ArrayList<QueryResponseJSON> correctResults = new ArrayList<>(); int i = 0; while (i < (dataElements.size() - removeTailElements)) { @@ -404,7 +403,7 @@ public class BaseTests { logger.info("Running testSRCIPQueryNoFilter(): "); - ArrayList<QueryResponseJSON> results = null; + ArrayList<QueryResponseJSON> results; int numExpectedResults = 3; if (isDistributed) { @@ -421,7 +420,7 @@ public class BaseTests fail("results.size() = " + results.size() + " -- must equal " + numExpectedResults); } - ArrayList<QueryResponseJSON> correctResults = new ArrayList<QueryResponseJSON>(); + ArrayList<QueryResponseJSON> correctResults = new ArrayList<>(); int i = 0; while (i < dataElements.size()) { @@ -471,9 +470,9 @@ public class BaseTests // Method to convert a ArrayList<String> into the correct (padded) returned ArrayList private static ArrayList<String> parseArray(JSONObject dataMap, String fieldName, boolean isIP) { - ArrayList<String> retArray = new ArrayList<String>(); + ArrayList<String> retArray = new ArrayList<>(); - ArrayList<String> values = null; + ArrayList<String> values; if (dataMap.get(fieldName) instanceof ArrayList) { values = (ArrayList<String>) dataMap.get(fieldName); @@ -506,7 +505,7 @@ public class BaseTests // Method to convert a ArrayList<Short> into the correct (padded) returned ArrayList private static ArrayList<Short> parseShortArray(JSONObject dataMap, String fieldName) { - ArrayList<Short> retArray = new ArrayList<Short>(); + ArrayList<Short> retArray = new ArrayList<>(); ArrayList<Short> values = (ArrayList<Short>) dataMap.get(fieldName); @@ -529,7 +528,7 @@ public class BaseTests // Method to convert the String field value to the correct returned substring private static String parseString(JSONObject dataMap, String fieldName) { - String ret = null; + String ret; String element = (String) dataMap.get(fieldName); int numParts = Integer.parseInt(SystemConfiguration.getProperty("pir.stringBits")) / dataPartitionBitSize; @@ -604,7 +603,7 @@ public class BaseTests // Method to pull the elements of a list (either an ArrayList or JSONArray) into a HashSet private static HashSet<String> getSetFromList(Object list) { - HashSet<String> set = new HashSet<String>(); + HashSet<String> set = new HashSet<>(); if (list instanceof ArrayList) { http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/test/utils/Inputs.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/test/utils/Inputs.java b/src/main/java/org/apache/pirk/test/utils/Inputs.java index 1b362ef..5070672 100644 --- a/src/main/java/org/apache/pirk/test/utils/Inputs.java +++ b/src/main/java/org/apache/pirk/test/utils/Inputs.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.test.utils; import java.io.File; @@ -23,6 +23,7 @@ import java.io.IOException; import java.io.OutputStreamWriter; import java.util.ArrayList; import java.util.Arrays; +import java.util.Collections; import java.util.List; import javax.xml.parsers.DocumentBuilder; @@ -34,7 +35,6 @@ import javax.xml.transform.stream.StreamResult; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; -import org.apache.log4j.Logger; import org.apache.pirk.schema.data.LoadDataSchemas; import org.apache.pirk.schema.data.partitioner.IPDataPartitioner; import org.apache.pirk.schema.data.partitioner.ISO8601DatePartitioner; @@ -42,9 +42,10 @@ import org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner; import org.apache.pirk.schema.query.LoadQuerySchemas; import org.apache.pirk.test.distributed.DistributedTestDriver; import org.apache.pirk.utils.HDFS; -import org.apache.pirk.utils.LogUtils; import org.apache.pirk.utils.SystemConfiguration; import org.json.simple.JSONObject; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.w3c.dom.Document; import org.w3c.dom.Element; @@ -54,7 +55,7 @@ import org.w3c.dom.Element; */ public class Inputs { - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(Inputs.class); // Test data schema fields public static final String DATE = "date"; @@ -126,16 +127,16 @@ public class Inputs @SuppressWarnings("unchecked") public static ArrayList<JSONObject> createJSONDataElements() { - ArrayList<JSONObject> dataElementsJSON = new ArrayList<JSONObject>(); + ArrayList<JSONObject> dataElementsJSON = new ArrayList<>(); JSONObject jsonObj1 = new JSONObject(); jsonObj1.put(DATE, "2016-02-20T23:29:05.000Z"); jsonObj1.put(QNAME, "a.b.c.com"); // hits on domain selector jsonObj1.put(SRCIP, "55.55.55.55"); // hits on IP selector jsonObj1.put(DSTIP, "1.2.3.6"); - jsonObj1.put(QTYPE, new ArrayList<Short>(Arrays.asList((short) 1))); + jsonObj1.put(QTYPE, new ArrayList<>(Collections.singletonList((short) 1))); jsonObj1.put(RCODE, 0); - jsonObj1.put(IPS, new ArrayList<String>(Arrays.asList("10.20.30.40", "10.20.30.60"))); + jsonObj1.put(IPS, new ArrayList<>(Arrays.asList("10.20.30.40", "10.20.30.60"))); dataElementsJSON.add(jsonObj1); @@ -144,9 +145,9 @@ public class Inputs jsonObj2.put(QNAME, "d.e.com"); jsonObj2.put(SRCIP, "127.128.129.130"); jsonObj2.put(DSTIP, "1.2.3.4"); - jsonObj2.put(QTYPE, new ArrayList<Short>(Arrays.asList((short) 1))); + jsonObj2.put(QTYPE, new ArrayList<>(Collections.singletonList((short) 1))); jsonObj2.put(RCODE, 0); - jsonObj2.put(IPS, new ArrayList<String>(Arrays.asList("5.6.7.8"))); + jsonObj2.put(IPS, new ArrayList<>(Collections.singletonList("5.6.7.8"))); dataElementsJSON.add(jsonObj2); @@ -155,9 +156,9 @@ public class Inputs jsonObj3.put(QNAME, "d.e.com"); jsonObj3.put(SRCIP, "131.132.133.134"); jsonObj3.put(DSTIP, "9.10.11.12"); - jsonObj3.put(QTYPE, new ArrayList<Short>(Arrays.asList((short) 1))); + jsonObj3.put(QTYPE, new ArrayList<>(Collections.singletonList((short) 1))); jsonObj3.put(RCODE, 0); - jsonObj3.put(IPS, new ArrayList<String>(Arrays.asList("13.14.15.16"))); + jsonObj3.put(IPS, new ArrayList<>(Collections.singletonList("13.14.15.16"))); dataElementsJSON.add(jsonObj3); @@ -166,9 +167,9 @@ public class Inputs jsonObj4.put(QNAME, "d.e.com"); jsonObj4.put(SRCIP, "135.136.137.138"); jsonObj4.put(DSTIP, "17.18.19.20"); - jsonObj4.put(QTYPE, new ArrayList<Short>(Arrays.asList((short) 1))); + jsonObj4.put(QTYPE, new ArrayList<>(Collections.singletonList((short) 1))); jsonObj4.put(RCODE, 3); - jsonObj4.put(IPS, new ArrayList<String>(Arrays.asList("21.22.23.24"))); + jsonObj4.put(IPS, new ArrayList<>(Collections.singletonList("21.22.23.24"))); dataElementsJSON.add(jsonObj4); @@ -177,9 +178,9 @@ public class Inputs jsonObj5.put(QNAME, "d.e.com"); jsonObj5.put(SRCIP, "139.140.141.142"); jsonObj5.put(DSTIP, "25.26.27.28"); - jsonObj5.put(QTYPE, new ArrayList<Short>(Arrays.asList((short) 1))); + jsonObj5.put(QTYPE, new ArrayList<>(Collections.singletonList((short) 1))); jsonObj5.put(RCODE, 0); - jsonObj5.put(IPS, new ArrayList<String>(Arrays.asList("5.6.7.8"))); + jsonObj5.put(IPS, new ArrayList<>(Collections.singletonList("5.6.7.8"))); dataElementsJSON.add(jsonObj5); @@ -188,9 +189,9 @@ public class Inputs jsonObj6.put(QNAME, "d.e.com"); jsonObj6.put(SRCIP, "143.144.145.146"); jsonObj6.put(DSTIP, "33.34.35.36"); - jsonObj6.put(QTYPE, new ArrayList<Short>(Arrays.asList((short) 1))); + jsonObj6.put(QTYPE, new ArrayList<>(Collections.singletonList((short) 1))); jsonObj6.put(RCODE, 0); - jsonObj6.put(IPS, new ArrayList<String>(Arrays.asList("5.6.7.8"))); + jsonObj6.put(IPS, new ArrayList<>(Collections.singletonList("5.6.7.8"))); dataElementsJSON.add(jsonObj6); @@ -199,9 +200,9 @@ public class Inputs jsonObj7.put(QNAME, "something.else"); jsonObj7.put(SRCIP, "1.1.1.1"); jsonObj7.put(DSTIP, "2.2.2.2"); - jsonObj7.put(QTYPE, new ArrayList<Short>(Arrays.asList((short) 1))); + jsonObj7.put(QTYPE, new ArrayList<>(Collections.singletonList((short) 1))); jsonObj7.put(RCODE, 0); - jsonObj7.put(IPS, new ArrayList<String>(Arrays.asList("3.3.3.3"))); + jsonObj7.put(IPS, new ArrayList<>(Collections.singletonList("3.3.3.3"))); dataElementsJSON.add(jsonObj7); @@ -212,9 +213,9 @@ public class Inputs jsonObj8.put(QNAME, "something.else2"); jsonObj8.put(SRCIP, "5.6.7.8"); jsonObj8.put(DSTIP, "2.2.2.22"); - jsonObj8.put(QTYPE, new ArrayList<Short>(Arrays.asList((short) 1))); + jsonObj8.put(QTYPE, new ArrayList<>(Collections.singletonList((short) 1))); jsonObj8.put(RCODE, 0); - jsonObj8.put(IPS, new ArrayList<String>(Arrays.asList("3.3.3.132"))); + jsonObj8.put(IPS, new ArrayList<>(Collections.singletonList("3.3.3.132"))); dataElementsJSON.add(jsonObj8); @@ -224,9 +225,9 @@ public class Inputs jsonObj9.put(QNAME, "something.else.on.stoplist"); jsonObj9.put(SRCIP, "55.55.55.55"); jsonObj9.put(DSTIP, "2.2.2.232"); - jsonObj9.put(QTYPE, new ArrayList<Short>(Arrays.asList((short) 1))); + jsonObj9.put(QTYPE, new ArrayList<>(Collections.singletonList((short) 1))); jsonObj9.put(RCODE, 0); - jsonObj9.put(IPS, new ArrayList<String>(Arrays.asList("3.3.3.132"))); + jsonObj9.put(IPS, new ArrayList<>(Collections.singletonList("3.3.3.132"))); dataElementsJSON.add(jsonObj9); @@ -239,16 +240,16 @@ public class Inputs @SuppressWarnings("unchecked") public static ArrayList<JSONObject> getRcode3JSONDataElements() { - ArrayList<JSONObject> dataElementsJSON = new ArrayList<JSONObject>(); + ArrayList<JSONObject> dataElementsJSON = new ArrayList<>(); JSONObject jsonObj4 = new JSONObject(); jsonObj4.put(DATE, "2016-02-20T23:29:08.000Z"); jsonObj4.put(QNAME, "d.e.com"); jsonObj4.put(SRCIP, "135.136.137.138"); jsonObj4.put(DSTIP, "17.18.19.20"); - jsonObj4.put(QTYPE, new ArrayList<Short>(Arrays.asList((short) 1))); + jsonObj4.put(QTYPE, new ArrayList<>(Collections.singletonList((short) 1))); jsonObj4.put(RCODE, 3); - jsonObj4.put(IPS, new ArrayList<String>(Arrays.asList("21.22.23.24"))); + jsonObj4.put(IPS, new ArrayList<>(Collections.singletonList("21.22.23.24"))); dataElementsJSON.add(jsonObj4); @@ -393,9 +394,9 @@ public class Inputs { logger.info("PIR stopList file being created"); - String tmpFileName = null; + String tmpFileName; - ArrayList<String> elements = new ArrayList<String>(); + ArrayList<String> elements = new ArrayList<>(); elements.add("something.else.on.stoplist"); elements.add("3.3.3.132"); @@ -441,7 +442,7 @@ public class Inputs // Create and load the query schemas // DNS_HOSTNAME_QUERY List<String> dnsHostnameQueryElements = Arrays.asList(DATE, SRCIP, DSTIP, QTYPE, RCODE, IPS); - List<String> dnsHostnameQueryFilterElements = Arrays.asList(QNAME); + List<String> dnsHostnameQueryFilterElements = Collections.singletonList(QNAME); TestUtils.createQuerySchema(DNS_HOSTNAME_QUERY_FILE, DNS_HOSTNAME_QUERY, TEST_DATA_SCHEMA_NAME, QNAME, dnsHostnameQueryElements, dnsHostnameQueryFilterElements, filter); @@ -453,7 +454,7 @@ public class Inputs // DNS_IP_QUERY List<String> dnsIPQueryElements = Arrays.asList(SRCIP, DSTIP, IPS); - List<String> dnsIPQueryFilterElements = Arrays.asList(QNAME); + List<String> dnsIPQueryFilterElements = Collections.singletonList(QNAME); TestUtils.createQuerySchema(DNS_IP_QUERY_FILE, DNS_IP_QUERY, TEST_DATA_SCHEMA_NAME, IPS, dnsIPQueryElements, dnsIPQueryFilterElements, filter); if (hdfs) @@ -464,7 +465,7 @@ public class Inputs // DNS_NXDOMAIN_QUERY List<String> dnsNXQueryElements = Arrays.asList(QNAME, SRCIP, DSTIP); - List<String> dnsNXQueryFilterElements = Arrays.asList(QNAME); + List<String> dnsNXQueryFilterElements = Collections.singletonList(QNAME); TestUtils .createQuerySchema(DNS_NXDOMAIN_QUERY_FILE, DNS_NXDOMAIN_QUERY, TEST_DATA_SCHEMA_NAME, QNAME, dnsNXQueryElements, dnsNXQueryFilterElements, filter); http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/test/utils/StandaloneQuery.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/test/utils/StandaloneQuery.java b/src/main/java/org/apache/pirk/test/utils/StandaloneQuery.java index 027e042..aeda7dc 100644 --- a/src/main/java/org/apache/pirk/test/utils/StandaloneQuery.java +++ b/src/main/java/org/apache/pirk/test/utils/StandaloneQuery.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,17 +15,14 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.test.utils; -import static org.junit.Assert.fail; - import java.io.File; import java.io.IOException; import java.util.ArrayList; import java.util.HashMap; -import org.apache.log4j.Logger; import org.apache.pirk.encryption.Paillier; import org.apache.pirk.querier.wideskies.Querier; import org.apache.pirk.querier.wideskies.QuerierConst; @@ -37,25 +34,23 @@ import org.apache.pirk.query.wideskies.QueryUtils; import org.apache.pirk.responder.wideskies.standalone.Responder; import org.apache.pirk.response.wideskies.Response; import org.apache.pirk.schema.response.QueryResponseJSON; -import org.apache.pirk.utils.LogUtils; import org.apache.pirk.utils.PIRException; import org.apache.pirk.utils.SystemConfiguration; import org.json.simple.JSONObject; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; +import static org.junit.Assert.fail; public class StandaloneQuery { - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(StandaloneQuery.class); static String queryFileDomain = "qfDomain"; static String queryFileIP = "qfIP"; - static String querySideOuputFilePrefix = "querySideOut"; // the file pre-fix for the query side output files - static String finalResultsFile = "finalResultFile"; // file to hold the final results String testDataSchemaName = "testDataSchema"; String testQuerySchemaName = "testQuerySchema"; - static String responseFile = "encryptedResponse"; // the PIR response file from the responder - // Base method to perform the query public static ArrayList<QueryResponseJSON> performStandaloneQuery(ArrayList<JSONObject> dataElements, String queryType, ArrayList<String> selectors, int numThreads, boolean testFalsePositive) throws IOException, InterruptedException, PIRException @@ -65,9 +60,12 @@ public class StandaloneQuery ArrayList<QueryResponseJSON> results = null; // Create the necessary files + String querySideOuputFilePrefix = "querySideOut"; File fileQuerier = File.createTempFile(querySideOuputFilePrefix + "-" + QuerierConst.QUERIER_FILETAG, ".txt"); File fileQuery = File.createTempFile(querySideOuputFilePrefix + "-" + QuerierConst.QUERY_FILETAG, ".txt"); + String responseFile = "encryptedResponse"; File fileResponse = File.createTempFile(responseFile, ".txt"); + String finalResultsFile = "finalResultFile"; File fileFinalResults = File.createTempFile(finalResultsFile, ".txt"); logger.info("fileQuerier = " + fileQuerier.getAbsolutePath() + " fileQuery = " + fileQuery.getAbsolutePath() + " responseFile = " http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/test/utils/TestUtils.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/test/utils/TestUtils.java b/src/main/java/org/apache/pirk/test/utils/TestUtils.java index 11348cb..57af5c4 100644 --- a/src/main/java/org/apache/pirk/test/utils/TestUtils.java +++ b/src/main/java/org/apache/pirk/test/utils/TestUtils.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.test.utils; import java.io.BufferedReader; @@ -38,10 +38,10 @@ import javax.xml.transform.stream.StreamResult; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; -import org.apache.log4j.Logger; import org.apache.pirk.schema.response.QueryResponseJSON; -import org.apache.pirk.utils.LogUtils; import org.apache.pirk.utils.SystemConfiguration; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.w3c.dom.Document; import org.w3c.dom.Element; @@ -51,7 +51,7 @@ import org.w3c.dom.Element; */ public class TestUtils { - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(TestUtils.class); /** * Method to delete an ES index @@ -81,7 +81,7 @@ public class TestUtils // Read the output from the command logger.info("Standard output of the command:\n"); - String s = null; + String s; while ((s = stdInput.readLine()) != null) { logger.info(s); @@ -142,7 +142,7 @@ public class TestUtils logger.info("createQuerySchema: querySchemaName = " + querySchemaName); // Create a temporary file for the test schema, set in the properties - String fileName = null; + String fileName; File file = null; OutputStreamWriter osw = null; if (hdfs) @@ -237,7 +237,7 @@ public class TestUtils TransformerFactory transformerFactory = TransformerFactory.newInstance(); Transformer transformer = transformerFactory.newTransformer(); DOMSource source = new DOMSource(doc); - StreamResult result = null; + StreamResult result; if (hdfs) { result = new StreamResult(osw); @@ -269,7 +269,7 @@ public class TestUtils */ public static ArrayList<QueryResponseJSON> readResultsFile(File file) { - ArrayList<QueryResponseJSON> results = new ArrayList<QueryResponseJSON>(); + ArrayList<QueryResponseJSON> results = new ArrayList<>(); try { FileReader fr = new FileReader(file); @@ -295,7 +295,7 @@ public class TestUtils */ public static String writeToTmpFile(ArrayList<String> list, String fileName, String suffix) throws IOException { - String filename = null; + String filename; File file = File.createTempFile(fileName, suffix); file.deleteOnExit(); http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/utils/CSVOutputUtils.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/utils/CSVOutputUtils.java b/src/main/java/org/apache/pirk/utils/CSVOutputUtils.java index ebb1f21..311fe87 100644 --- a/src/main/java/org/apache/pirk/utils/CSVOutputUtils.java +++ b/src/main/java/org/apache/pirk/utils/CSVOutputUtils.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,11 +15,12 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.utils; import org.apache.hadoop.io.Text; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * Utils class for CSV value output @@ -31,7 +32,7 @@ public class CSVOutputUtils { public static String EMPTYFIELD = ""; - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(CSVOutputUtils.class); public static Text setCSVOutput(String domain, String ip, String timestamp) { http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/utils/EpochDateParser.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/utils/EpochDateParser.java b/src/main/java/org/apache/pirk/utils/EpochDateParser.java index 388726f..870128c 100644 --- a/src/main/java/org/apache/pirk/utils/EpochDateParser.java +++ b/src/main/java/org/apache/pirk/utils/EpochDateParser.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,10 +15,11 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.utils; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * Class to parse dates in Epoch date format @@ -26,7 +27,7 @@ import org.apache.log4j.Logger; public class EpochDateParser { - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(EpochDateParser.class); public static boolean isEpochDateFormat(String date) { http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/utils/FileConst.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/utils/FileConst.java b/src/main/java/org/apache/pirk/utils/FileConst.java index 83e43d8..382b97d 100644 --- a/src/main/java/org/apache/pirk/utils/FileConst.java +++ b/src/main/java/org/apache/pirk/utils/FileConst.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.utils; /** http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/utils/FileIOUtils.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/utils/FileIOUtils.java b/src/main/java/org/apache/pirk/utils/FileIOUtils.java index 52bde2e..17b97ec 100644 --- a/src/main/java/org/apache/pirk/utils/FileIOUtils.java +++ b/src/main/java/org/apache/pirk/utils/FileIOUtils.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.utils; import java.io.BufferedReader; @@ -29,14 +29,15 @@ import java.util.AbstractCollection; import java.util.ArrayList; import java.util.HashSet; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * Class holding basic fileIO utils */ public class FileIOUtils { - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(FileIOUtils.class); public interface Callable<V> { http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/utils/HDFS.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/utils/HDFS.java b/src/main/java/org/apache/pirk/utils/HDFS.java index 8f3992e..32e0c05 100644 --- a/src/main/java/org/apache/pirk/utils/HDFS.java +++ b/src/main/java/org/apache/pirk/utils/HDFS.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.utils; import java.io.BufferedReader; @@ -33,15 +33,16 @@ import java.util.Map; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; -import org.apache.log4j.Logger; import org.json.simple.JSONObject; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * Class for basic HDFS fileIO utils */ public class HDFS { - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(HDFS.class); public static void writeFile(Collection<String> elements, FileSystem fs, String path, boolean deleteOnExit) { @@ -190,7 +191,7 @@ public class HDFS public static HashSet<String> readFileHashSet(FileSystem fs, Path path) { - HashSet<String> rv = new HashSet<String>(); + HashSet<String> rv = new HashSet<>(); try { InputStreamReader isr = new InputStreamReader(fs.open(path)); http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/utils/ISO8601DateParser.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/utils/ISO8601DateParser.java b/src/main/java/org/apache/pirk/utils/ISO8601DateParser.java index 0826322..675fb21 100755 --- a/src/main/java/org/apache/pirk/utils/ISO8601DateParser.java +++ b/src/main/java/org/apache/pirk/utils/ISO8601DateParser.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.utils; import java.text.ParseException; @@ -23,7 +23,8 @@ import java.text.SimpleDateFormat; import java.util.Date; import java.util.TimeZone; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * Class to parse a date in ISO86091 format @@ -31,7 +32,7 @@ import org.apache.log4j.Logger; */ public class ISO8601DateParser { - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(ISO8601DateParser.class); static { @@ -51,7 +52,7 @@ public class ISO8601DateParser try { return format.parse(date).getTime() + ""; - } catch (Exception e) + } catch (Exception ignore) {} return null; http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/utils/KeyedHash.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/utils/KeyedHash.java b/src/main/java/org/apache/pirk/utils/KeyedHash.java index 481108e..579b196 100644 --- a/src/main/java/org/apache/pirk/utils/KeyedHash.java +++ b/src/main/java/org/apache/pirk/utils/KeyedHash.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,13 +15,14 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.utils; import java.security.MessageDigest; import java.security.NoSuchAlgorithmException; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * Class for the PIR keyed hash @@ -31,7 +32,7 @@ import org.apache.log4j.Logger; */ public class KeyedHash { - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(KeyedHash.class); /** * Hash method that uses the java String hashCode() @@ -58,9 +59,9 @@ public class KeyedHash */ public static int hash(String key, int bitSize, String input, String hashType) { - int bitLimitedHash = 0; + int bitLimitedHash; - MessageDigest md = null; + MessageDigest md; try { md = MessageDigest.getInstance(hashType); http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/utils/LogUtils.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/utils/LogUtils.java b/src/main/java/org/apache/pirk/utils/LogUtils.java deleted file mode 100755 index d10d43b..0000000 --- a/src/main/java/org/apache/pirk/utils/LogUtils.java +++ /dev/null @@ -1,119 +0,0 @@ -/******************************************************************************* - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, - * software distributed under the License is distributed on an - * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY - * KIND, either express or implied. See the License for the - * specific language governing permissions and limitations - * under the License. - *******************************************************************************/ -package org.apache.pirk.utils; - -import java.io.InputStream; -import java.util.Properties; - -import org.apache.log4j.Logger; -import org.apache.log4j.PropertyConfigurator; - -/** - * Class for basic logging utils - */ -public class LogUtils -{ - public static Logger logger; - - static - { - initialize(); - } - - public static void initialize() - { - Properties props = new Properties(); - try - { - String log4jFilename = SystemConfiguration.getProperty("log4jPropertiesFile"); - if (log4jFilename == null) - { - System.err.println("log4jPropertiesFile property not found during LogUtils initialization."); - } - else - { - InputStream stream = SystemConfiguration.class.getClassLoader().getResourceAsStream(log4jFilename); - if (stream != null) - { - System.out.println("Loading log4j properties file: '" + log4jFilename + "'"); - props.load(stream); - PropertyConfigurator.configure(props); - stream.close(); - } - else - { - System.err.println("log4j properties file not found: '" + log4jFilename + "'"); - } - } - } catch (Exception e) - { - System.err.println("Exception occured configuring the log4j system: " + e.toString()); - e.printStackTrace(); - } - } - - /** - * Should be called at the from from each class using log4j. Example: static private Logger logger = LogUtils.getLoggerForThisClass(); - * - * @return - */ - public static Logger getLoggerForThisClass() - { - // We use the third stack element; second is this method, first is - // .getStackTrace() - StackTraceElement myCaller = Thread.currentThread().getStackTrace()[2]; - return Logger.getLogger(myCaller.getClassName()); - } - - /** - * Returns the name of the class calling this method. - * - */ - public static String getNameForThisClass() - { - // We use the third stack element; second is this method, first is - // .getStackTrace() - StackTraceElement myCaller = Thread.currentThread().getStackTrace()[2]; - return myCaller.getClassName(); - } - - public static String entering() - { - StackTraceElement myCaller = Thread.currentThread().getStackTrace()[2]; - String methodName = myCaller.getMethodName(); - return entering(methodName); - } - - public static String exiting() - { - StackTraceElement myCaller = Thread.currentThread().getStackTrace()[2]; - String methodName = myCaller.getMethodName(); - return exiting(methodName); - } - - public static String entering(String methodName) - { - return String.format("Entering %s", methodName); - } - - public static String exiting(String methodName) - { - return String.format("Exiting %s", methodName); - } -} http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/utils/PIRException.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/utils/PIRException.java b/src/main/java/org/apache/pirk/utils/PIRException.java index 9010c0f..7ec9c8b 100644 --- a/src/main/java/org/apache/pirk/utils/PIRException.java +++ b/src/main/java/org/apache/pirk/utils/PIRException.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.utils; /** http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/utils/QueryParserUtils.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/utils/QueryParserUtils.java b/src/main/java/org/apache/pirk/utils/QueryParserUtils.java index 5bc9794..afef6c9 100644 --- a/src/main/java/org/apache/pirk/utils/QueryParserUtils.java +++ b/src/main/java/org/apache/pirk/utils/QueryParserUtils.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,21 +15,23 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.utils; import java.text.ParseException; import java.util.ArrayList; +import java.util.Arrays; import java.util.Map; import java.util.regex.Pattern; import org.apache.hadoop.io.MapWritable; import org.apache.hadoop.io.Text; -import org.apache.log4j.Logger; +import org.slf4j.Logger; import org.apache.pirk.inputformat.hadoop.TextArrayWritable; import org.apache.pirk.schema.data.DataSchema; import org.apache.pirk.schema.data.partitioner.IPDataPartitioner; import org.elasticsearch.hadoop.mr.WritableArrayWritable; +import org.slf4j.LoggerFactory; /** * Class used for URI query parsing @@ -41,7 +43,7 @@ import org.elasticsearch.hadoop.mr.WritableArrayWritable; */ public class QueryParserUtils { - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(QueryParserUtils.class); /** * Given a URI query string, checks to see if the given document satisfies the query @@ -73,7 +75,7 @@ public class QueryParserUtils String[] queryTokens = uriQuery.split("\\+(?=AND)|\\+(?=OR)|\\+(?=[a-z])"); // booleans of the form +AND+, +OR+, don't split on +T0+ int index = 0; - String item = null; + String item; while (index < queryTokens.length) { boolean ignoreCase = false; @@ -227,7 +229,7 @@ public class QueryParserUtils { if (queryTokens[index].equals("AND")) // Do nothing and keep going { - if (satisfiesQuery == false) + if (!satisfiesQuery) { break; } @@ -247,7 +249,7 @@ public class QueryParserUtils satisfiesQuery = true; // reset so that we pick up matches for the next term } } - else if (satisfiesQuery == false) + else if (!satisfiesQuery) { logger.debug("Does not satisfy the query and no boolean ops next..."); break; @@ -274,7 +276,7 @@ public class QueryParserUtils String[] queryTokens = uriQuery.split("\\+(?=AND)|\\+(?=OR)|\\+(?=[a-z])"); // booleans of the form +AND+, +OR+, don't split on +T0+ int index = 0; - String item = null; + String item; while (index < queryTokens.length) { item = queryTokens[index]; @@ -374,12 +376,12 @@ public class QueryParserUtils logger.debug("itemTokens[1] = " + itemTokens[1] + " contains wildcard"); if (!Pattern.matches(wildcardToRegex(itemTokens[1]), (String) value)) { - logger.debug("stringValue = " + (String) value + " did not satisfy itemTokens[1] = " + itemTokens[1]); + logger.debug("stringValue = " + value + " did not satisfy itemTokens[1] = " + itemTokens[1]); satisfiesQuery = false; } - logger.debug("stringValue = " + (String) value + " did satisfy itemTokens[1] = " + itemTokens[1]); + logger.debug("stringValue = " + value + " did satisfy itemTokens[1] = " + itemTokens[1]); } - else if (!((String) value).equals(itemTokens[1])) // Single value match + else if (!(value).equals(itemTokens[1])) // Single value match { logger.debug("We do not have a single value match: stringValue " + (String) value + " != itemTokens[1] = " + itemTokens[1]); satisfiesQuery = false; @@ -427,7 +429,7 @@ public class QueryParserUtils { if (queryTokens[index].equals("AND")) // Do nothing and keep going { - if (satisfiesQuery == false) + if (!satisfiesQuery) { break; } @@ -447,7 +449,7 @@ public class QueryParserUtils satisfiesQuery = true; // reset so that we pick up matches for the next term } } - else if (satisfiesQuery == false) + else if (!satisfiesQuery) { logger.debug("Does not satisfy the query and no boolean ops next..."); break; @@ -474,7 +476,7 @@ public class QueryParserUtils String[] queryTokens = uriQuery.split("\\+(?=AND)|\\+(?=OR)|\\+(?=[a-z])"); // booleans of the form +AND+, +OR+, don't split on +T0+ int index = 0; - String item = null; + String item; while (index < queryTokens.length) { boolean ignoreCase = false; @@ -628,7 +630,7 @@ public class QueryParserUtils { if (queryTokens[index].equals("AND")) // Do nothing and keep going { - if (satisfiesQuery == false) + if (!satisfiesQuery) { break; } @@ -648,7 +650,7 @@ public class QueryParserUtils satisfiesQuery = true; // reset so that we pick up matches for the next term } } - else if (satisfiesQuery == false) + else if (!satisfiesQuery) { logger.debug("Does not satisfy the query and no boolean ops next..."); break; @@ -690,7 +692,7 @@ public class QueryParserUtils String[] ranges = query.split("\\+TO\\+"); logger.info("ranges[0] = " + ranges[0] + " ranges[1] = " + ranges[1]); - if ((inclusive == false) && (value.equals(ranges[0]) || value.equals(ranges[1]))) + if ((!inclusive) && (value.equals(ranges[0]) || value.equals(ranges[1]))) { logger.debug("inclusive = false and either value.equals(ranges[0]) or value.equals(ranges[1])"); matches = false; @@ -769,10 +771,10 @@ public class QueryParserUtils toMilli = Integer.parseInt(toDateArr[1]); } catch (Exception e) { - logger.info(e.getStackTrace().toString()); + logger.info(Arrays.toString(e.getStackTrace())); } - if ((inclusive == false) && (fromDate == valueDate || toDate == valueDate)) + if ((!inclusive) && (fromDate == valueDate || toDate == valueDate)) { logger.debug("(inclusive == false) && (fromDate == valueDate || toDate == valueDate))"); matches = false; @@ -800,7 +802,7 @@ public class QueryParserUtils { e.printStackTrace(); } - if ((inclusive == false) && (lower == valueDate || upper == valueDate)) + if ((!inclusive) && (lower == valueDate || upper == valueDate)) { logger.debug("(inclusive == false) && (lower == valueDate || upper == valueDate))"); matches = false; @@ -830,7 +832,7 @@ public class QueryParserUtils int valueInt = Integer.parseInt(value); logger.debug("valueInt = " + valueInt + " lower = " + lower + " upper = " + upper); - if ((inclusive == false) && (lower == valueInt || upper == valueInt)) + if ((!inclusive) && (lower == valueInt || upper == valueInt)) { logger.debug("(inclusive == false) && (lower == valueInt || upper == valueInt))"); matches = false; @@ -855,7 +857,7 @@ public class QueryParserUtils */ public static String wildcardToRegex(String wildcard) { - StringBuffer s = new StringBuffer(wildcard.length()); + StringBuilder s = new StringBuilder(wildcard.length()); for (int i = 0, is = wildcard.length(); i < is; i++) { char c = wildcard.charAt(i); http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/utils/StopListUtils.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/utils/StopListUtils.java b/src/main/java/org/apache/pirk/utils/StopListUtils.java index 2019fe1..285179e 100644 --- a/src/main/java/org/apache/pirk/utils/StopListUtils.java +++ b/src/main/java/org/apache/pirk/utils/StopListUtils.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.utils; import java.util.HashSet; http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/ef8d1c1a/src/main/java/org/apache/pirk/utils/StringUtils.java ---------------------------------------------------------------------- diff --git a/src/main/java/org/apache/pirk/utils/StringUtils.java b/src/main/java/org/apache/pirk/utils/StringUtils.java index 9ac0279..be8ea09 100755 --- a/src/main/java/org/apache/pirk/utils/StringUtils.java +++ b/src/main/java/org/apache/pirk/utils/StringUtils.java @@ -1,4 +1,4 @@ -/******************************************************************************* +/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information @@ -15,7 +15,7 @@ * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. - *******************************************************************************/ + */ package org.apache.pirk.utils; import java.util.ArrayList; @@ -27,12 +27,13 @@ import org.apache.hadoop.io.ArrayWritable; import org.apache.hadoop.io.MapWritable; import org.apache.hadoop.io.Text; import org.apache.hadoop.io.Writable; -import org.apache.log4j.Logger; +import org.slf4j.Logger; import org.apache.pirk.schema.data.DataSchema; import org.elasticsearch.hadoop.mr.WritableArrayWritable; import org.json.simple.JSONObject; import org.json.simple.parser.JSONParser; import org.json.simple.parser.ParseException; +import org.slf4j.LoggerFactory; /** * Pirk-specific string utilities @@ -40,7 +41,7 @@ import org.json.simple.parser.ParseException; */ public class StringUtils { - private static Logger logger = LogUtils.getLoggerForThisClass(); + private static final Logger logger = LoggerFactory.getLogger(StringUtils.class); /** * Method to convert a MapWritable into a JSON string @@ -176,7 +177,7 @@ public class StringUtils */ public static Map<String,Object> jsonStringToMap(String jsonString, DataSchema dataSchema) { - Map<String,Object> value = new HashMap<String,Object>(); + Map<String,Object> value = new HashMap<>(); JSONParser jsonParser = new JSONParser(); try @@ -255,7 +256,7 @@ public class StringUtils modString = modString.replaceAll("\"", ""); String[] elements = modString.split("\\s*,\\s*"); - return new ArrayList<String>(Arrays.asList(elements)); + return new ArrayList<>(Arrays.asList(elements)); } /** @@ -266,8 +267,6 @@ public class StringUtils String modString = jsonString.replaceFirst("\\[", ""); modString = modString.replaceFirst("\\]", ""); modString = modString.replaceAll("\"", ""); - String[] elements = modString.split("\\s*,\\s*"); - - return elements; + return modString.split("\\s*,\\s*"); } }