sdimbsn commented on a change in pull request #6235: Druid with Oak supporting 
also plain mode  v04
URL: https://github.com/apache/incubator-druid/pull/6235#discussion_r213229161
 
 

 ##########
 File path: 
processing/src/main/java/io/druid/segment/incremental/OakIncrementalIndex.java
 ##########
 @@ -0,0 +1,569 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *   http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing,
+ * software distributed under the License is distributed on an
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ * KIND, either express or implied.  See the License for the
+ * specific language governing permissions and limitations
+ * under the License.
+ */
+
+package io.druid.segment.incremental;
+
+import com.google.common.collect.Iterators;
+import com.google.common.collect.Maps;
+import io.druid.data.input.InputRow;
+import io.druid.data.input.MapBasedRow;
+import io.druid.data.input.Row;
+import io.druid.java.util.common.StringUtils;
+import io.druid.java.util.common.logger.Logger;
+import io.druid.java.util.common.parsers.ParseException;
+import io.druid.query.aggregation.AggregatorFactory;
+import io.druid.query.aggregation.BufferAggregator;
+import io.druid.query.aggregation.PostAggregator;
+import io.druid.segment.ColumnValueSelector;
+
+import java.nio.ByteBuffer;
+import java.util.Iterator;
+import java.util.Map;
+import java.util.List;
+import java.util.concurrent.atomic.AtomicInteger;
+import java.util.function.Function;
+
+import io.druid.segment.DimensionHandler;
+import io.druid.segment.DimensionIndexer;
+import io.druid.segment.column.ColumnCapabilitiesImpl;
+import io.druid.segment.column.ValueType;
+import com.oath.oak.OakMapBuilder;
+import com.oath.oak.OakMap;
+import com.oath.oak.CloseableIterator;
+import com.oath.oak.OakTransformView;
+import sun.misc.VM;
+
+import javax.annotation.Nullable;
+
+
+/**
+ */
+public class OakIncrementalIndex extends 
InternalDataIncrementalIndex<BufferAggregator>
+{
+
+  private static final Logger log = new Logger(OakIncrementalIndex.class);
+
+  // When serializing an object from IncrementalIndexRow.dims, we use:
+  // 1. 4 bytes for representing its type (Double, Float, Long or String)
+  // 2. 8 bytes for saving its value or the array position and length (in the 
case of String)
+  static final Integer ALLOC_PER_DIM = 12;
+  static final Integer NO_DIM = -1;
+  static final Integer TIME_STAMP_INDEX = 0;
+  static final Integer DIMS_LENGTH_INDEX = TIME_STAMP_INDEX + Long.BYTES;
+  static final Integer ROW_INDEX_INDEX = DIMS_LENGTH_INDEX + Integer.BYTES;
+  static final Integer DIMS_INDEX = ROW_INDEX_INDEX + Integer.BYTES;
+  // Serialization and deserialization offsets
+  static final Integer VALUE_TYPE_OFFSET = 0;
+  static final Integer DATA_OFFSET = VALUE_TYPE_OFFSET + Integer.BYTES;
+  static final Integer ARRAY_INDEX_OFFSET = VALUE_TYPE_OFFSET + Integer.BYTES;
+  static final Integer ARRAY_LENGTH_OFFSET = ARRAY_INDEX_OFFSET + 
Integer.BYTES;
+
+  OakMap<IncrementalIndexRow, Row> oak;
+
+  private OffheapAggsManager aggsManager;
+  private AtomicInteger versionCounter; // Only used in Plain mode
+
+  Map<String, String> env = System.getenv();
+
+  OakIncrementalIndex(
+          IncrementalIndexSchema incrementalIndexSchema,
+          boolean deserializeComplexMetrics,
+          boolean reportParseExceptions,
+          boolean concurrentEventAdd,
+          int maxRowCount
+  )
+  {
+    super(incrementalIndexSchema, reportParseExceptions, maxRowCount);
+
+    this.aggsManager = new OffheapAggsManager(incrementalIndexSchema, 
deserializeComplexMetrics,
+            reportParseExceptions, concurrentEventAdd, rowSupplier,
+            columnCapabilities, null, this);
+
+    this.versionCounter = new AtomicInteger(0);
+    IncrementalIndexRow minIncrementalIndexRow = getMinIncrementalIndexRow();
+
+    long maxDirectMemory = VM.maxDirectMemory();
+    int memoryCapacity = Integer.MAX_VALUE;
+    if (maxDirectMemory < memoryCapacity) {
+      memoryCapacity = (int) maxDirectMemory;
+    }
+
+    OakMapBuilder builder = new OakMapBuilder()
+            .setKeySerializer(new OakKeySerializer(dimensionDescsList))
+            .setValueSerializer(new OakValueSerializer(dimensionDescsList, 
aggsManager, reportParseExceptions, in))
+            .setMinKey(minIncrementalIndexRow)
+            .setComparator(new OakKeysComparator(dimensionDescsList, 
this.isRollup()))
+            .setMemoryCapacity(memoryCapacity);
+
+    if (env != null) {
+      String chunkMaxItems = env.get("chunkMaxItems");
+      if (chunkMaxItems != null) {
+        builder = builder.setChunkMaxItems(Integer.getInteger(chunkMaxItems));
+      }
+      String chunkBytesPerItem = env.get("chunkBytesPerItem");
+      if (chunkMaxItems != null) {
+        builder = 
builder.setChunkBytesPerItem(Integer.getInteger(chunkBytesPerItem));
+      }
+    }
+
+    oak = builder.build();
+  }
+
+  @Override
+  public Iterable<Row> iterableWithPostAggregations(List<PostAggregator> 
postAggs, boolean descending)
+  {
+    Function<Map.Entry<ByteBuffer, ByteBuffer>, Row> transformer = new 
Function<Map.Entry<ByteBuffer, ByteBuffer>, Row>()
+    {
+
+      @Override
+      public Row apply(Map.Entry<ByteBuffer, ByteBuffer> entry)
+      {
+        ByteBuffer serializedKey = entry.getKey();
+        ByteBuffer serializedValue = entry.getValue();
+        long timeStamp = OakIncrementalIndex.getTimestamp(serializedKey);
+        int dimsLength = OakIncrementalIndex.getDimsLength(serializedKey);
+        Map<String, Object> theVals = Maps.newLinkedHashMap();
+        for (int i = 0; i < dimsLength; ++i) {
+          Object dim = OakIncrementalIndex.getDimValue(serializedKey, i);
+          DimensionDesc dimensionDesc = dimensionDescsList.get(i);
+          if (dimensionDesc == null) {
+            continue;
+          }
+          String dimensionName = dimensionDesc.getName();
+          DimensionHandler handler = dimensionDesc.getHandler();
+          if (dim == null || handler.getLengthOfEncodedKeyComponent(dim) == 0) 
{
+            theVals.put(dimensionName, null);
+            continue;
+          }
+          final DimensionIndexer indexer = dimensionDesc.getIndexer();
+          Object rowVals = 
indexer.convertUnsortedEncodedKeyComponentToActualArrayOrList(dim, 
DimensionIndexer.LIST);
+          theVals.put(dimensionName, rowVals);
+        }
+
+        BufferAggregator[] aggs = aggsManager.getAggs();
+        for (int i = 0; i < aggs.length; ++i) {
+          theVals.put(aggsManager.metrics[i].getName(), 
aggs[i].get(serializedValue, aggsManager.aggOffsetInBuffer[i]));
+        }
+
+        return new MapBasedRow(timeStamp, theVals);
+      }
+    };
+
+    OakMap tmpOakMap = descending ? oak.descendingMap() : oak;
+    OakTransformView transformView = 
tmpOakMap.createTransformView(transformer);
+    CloseableIterator<Row> valuesIterator = transformView.valuesIterator();
+    return new Iterable<Row>()
+    {
+      @Override
+      public Iterator<Row> iterator()
+      {
+        return Iterators.transform(
+            valuesIterator,
+            row -> row
+        );
+      }
+    };
+  }
+
+  @Override
+  public Iterable<IncrementalIndexRow> persistIterable()
+  {
+    return keySet();
+  }
+
+  @Override
+  public void close() {}
+
+  @Override
+  protected long getMinTimeMillis()
+  {
+    return oak.getMinKey().getTimestamp();
+  }
+
+  @Override
+  protected long getMaxTimeMillis()
+  {
+    return oak.getMaxKey().getTimestamp();
+  }
+
+  @Override
+  protected BufferAggregator[] getAggsForRow(IncrementalIndexRow 
incrementalIndexRow)
+  {
+    return getAggs();
+  }
+
+  @Override
+  protected Object getAggVal(IncrementalIndexRow incrementalIndexRow, int 
aggIndex)
+  {
+    Function<Map.Entry<ByteBuffer, ByteBuffer>, Object> transformer = new 
Function<Map.Entry<ByteBuffer, ByteBuffer>, Object>() {
+      @Override
+      public Object apply(Map.Entry<ByteBuffer, ByteBuffer> entry)
+      {
+        ByteBuffer serializedValue = entry.getValue();
+        BufferAggregator agg = getAggs()[aggIndex];
+        return agg.get(serializedValue, serializedValue.position() + 
aggsManager.aggOffsetInBuffer[aggIndex]);
+      }
+    };
+
+    OakTransformView<IncrementalIndexRow, Object> transformView = 
(OakTransformView<IncrementalIndexRow, Object>) 
oak.createTransformView(transformer);
+    return transformView.get(incrementalIndexRow);
+  }
+
+  @Override
+  protected float getMetricFloatValue(IncrementalIndexRow incrementalIndexRow, 
int aggIndex)
+  {
+    Function<Map.Entry<ByteBuffer, ByteBuffer>, Float> transformer = new 
Function<Map.Entry<ByteBuffer, ByteBuffer>, Float>() {
+      @Override
+      public Float apply(Map.Entry<ByteBuffer, ByteBuffer> entry)
+      {
+        ByteBuffer serializedValue = entry.getValue();
+        BufferAggregator agg = getAggs()[aggIndex];
+        return agg.getFloat(serializedValue, serializedValue.position() + 
aggsManager.aggOffsetInBuffer[aggIndex]);
+      }
+    };
+
+    OakTransformView<IncrementalIndexRow, Float> transformView = 
(OakTransformView<IncrementalIndexRow, Float>) 
oak.createTransformView(transformer);
+    return transformView.get(incrementalIndexRow);
+  }
+
+  @Override
+  protected long getMetricLongValue(IncrementalIndexRow incrementalIndexRow, 
int aggIndex)
+  {
+    Function<Map.Entry<ByteBuffer, ByteBuffer>, Long> transformer = new 
Function<Map.Entry<ByteBuffer, ByteBuffer>, Long>() {
+      @Override
+      public Long apply(Map.Entry<ByteBuffer, ByteBuffer> entry)
+      {
+        ByteBuffer serializedValue = entry.getValue();
+        BufferAggregator agg = getAggs()[aggIndex];
+        return agg.getLong(serializedValue, serializedValue.position() + 
aggsManager.aggOffsetInBuffer[aggIndex]);
+      }
+    };
+
+    OakTransformView<IncrementalIndexRow, Long> transformView = 
(OakTransformView<IncrementalIndexRow, Long>) 
oak.createTransformView(transformer);
+    return transformView.get(incrementalIndexRow);
+  }
+
+  @Override
+  protected Object getMetricObjectValue(IncrementalIndexRow 
incrementalIndexRow, int aggIndex)
+  {
+    Function<Map.Entry<ByteBuffer, ByteBuffer>, Object> transformer = new 
Function<Map.Entry<ByteBuffer, ByteBuffer>, Object>() {
+      @Override
+      public Object apply(Map.Entry<ByteBuffer, ByteBuffer> entry)
+      {
+        ByteBuffer serializedValue = entry.getValue();
+        BufferAggregator agg = getAggs()[aggIndex];
+        return agg.get(serializedValue, serializedValue.position() + 
aggsManager.aggOffsetInBuffer[aggIndex]);
+      }
+    };
+
+    OakTransformView<IncrementalIndexRow, Object> transformView = 
(OakTransformView<IncrementalIndexRow, Object>) 
oak.createTransformView(transformer);
+    return transformView.get(incrementalIndexRow);
+  }
+
+  @Override
+  protected double getMetricDoubleValue(IncrementalIndexRow 
incrementalIndexRow, int aggIndex)
+  {
+    Function<Map.Entry<ByteBuffer, ByteBuffer>, Double> transformer = new 
Function<Map.Entry<ByteBuffer, ByteBuffer>, Double>() {
+      @Override
+      public Double apply(Map.Entry<ByteBuffer, ByteBuffer> entry)
+      {
+        ByteBuffer serializedValue = entry.getValue();
+        BufferAggregator agg = getAggs()[aggIndex];
+        return agg.getDouble(serializedValue, serializedValue.position() + 
aggsManager.aggOffsetInBuffer[aggIndex]);
+      }
+    };
+
+    OakTransformView<IncrementalIndexRow, Double> transformView = 
(OakTransformView<IncrementalIndexRow, Double>) 
oak.createTransformView(transformer);
+    return transformView.get(incrementalIndexRow);
+  }
+
+  @Override
+  protected boolean isNull(IncrementalIndexRow incrementalIndexRow, int 
aggIndex)
+  {
+    Function<Map.Entry<ByteBuffer, ByteBuffer>, Boolean> transformer = new 
Function<Map.Entry<ByteBuffer, ByteBuffer>, Boolean>() {
+      @Override
+      public Boolean apply(Map.Entry<ByteBuffer, ByteBuffer> entry)
+      {
+        ByteBuffer serializedValue = entry.getValue();
+        BufferAggregator agg = getAggs()[aggIndex];
+        return agg.isNull(serializedValue, serializedValue.position() + 
aggsManager.aggOffsetInBuffer[aggIndex]);
+      }
+    };
+
+    OakTransformView<IncrementalIndexRow, Boolean> transformView = 
(OakTransformView<IncrementalIndexRow, Boolean>) 
oak.createTransformView(transformer);
+    return transformView.get(incrementalIndexRow);
+  }
+
+  @Override
+  public Iterable<IncrementalIndexRow> timeRangeIterable(
+          boolean descending, long timeStart, long timeEnd)
+  {
+    if (timeStart > timeEnd) {
+      return null;
+    }
+
+    IncrementalIndexRow from = new IncrementalIndexRow(timeStart, null, 
dimensionDescsList, IncrementalIndexRow.EMPTY_ROW_INDEX);
+    IncrementalIndexRow to = new IncrementalIndexRow(timeEnd + 1, null, 
dimensionDescsList, IncrementalIndexRow.EMPTY_ROW_INDEX);
+    OakMap subMap = oak.subMap(from, true, to, false);
+    if (descending == true) {
+      subMap = subMap.descendingMap();
+    }
+    CloseableIterator<IncrementalIndexRow> keysIterator = 
subMap.keysIterator();
 
 Review comment:
   No `close()` call for keysIter here. May cause GC block.

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscr...@druid.apache.org
For additional commands, e-mail: commits-h...@druid.apache.org

Reply via email to