sdimbsn commented on a change in pull request #6235: Druid with Oak supporting also plain mode v04 URL: https://github.com/apache/incubator-druid/pull/6235#discussion_r213229161
########## File path: processing/src/main/java/io/druid/segment/incremental/OakIncrementalIndex.java ########## @@ -0,0 +1,569 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, + * software distributed under the License is distributed on an + * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY + * KIND, either express or implied. See the License for the + * specific language governing permissions and limitations + * under the License. + */ + +package io.druid.segment.incremental; + +import com.google.common.collect.Iterators; +import com.google.common.collect.Maps; +import io.druid.data.input.InputRow; +import io.druid.data.input.MapBasedRow; +import io.druid.data.input.Row; +import io.druid.java.util.common.StringUtils; +import io.druid.java.util.common.logger.Logger; +import io.druid.java.util.common.parsers.ParseException; +import io.druid.query.aggregation.AggregatorFactory; +import io.druid.query.aggregation.BufferAggregator; +import io.druid.query.aggregation.PostAggregator; +import io.druid.segment.ColumnValueSelector; + +import java.nio.ByteBuffer; +import java.util.Iterator; +import java.util.Map; +import java.util.List; +import java.util.concurrent.atomic.AtomicInteger; +import java.util.function.Function; + +import io.druid.segment.DimensionHandler; +import io.druid.segment.DimensionIndexer; +import io.druid.segment.column.ColumnCapabilitiesImpl; +import io.druid.segment.column.ValueType; +import com.oath.oak.OakMapBuilder; +import com.oath.oak.OakMap; +import com.oath.oak.CloseableIterator; +import com.oath.oak.OakTransformView; +import sun.misc.VM; + +import javax.annotation.Nullable; + + +/** + */ +public class OakIncrementalIndex extends InternalDataIncrementalIndex<BufferAggregator> +{ + + private static final Logger log = new Logger(OakIncrementalIndex.class); + + // When serializing an object from IncrementalIndexRow.dims, we use: + // 1. 4 bytes for representing its type (Double, Float, Long or String) + // 2. 8 bytes for saving its value or the array position and length (in the case of String) + static final Integer ALLOC_PER_DIM = 12; + static final Integer NO_DIM = -1; + static final Integer TIME_STAMP_INDEX = 0; + static final Integer DIMS_LENGTH_INDEX = TIME_STAMP_INDEX + Long.BYTES; + static final Integer ROW_INDEX_INDEX = DIMS_LENGTH_INDEX + Integer.BYTES; + static final Integer DIMS_INDEX = ROW_INDEX_INDEX + Integer.BYTES; + // Serialization and deserialization offsets + static final Integer VALUE_TYPE_OFFSET = 0; + static final Integer DATA_OFFSET = VALUE_TYPE_OFFSET + Integer.BYTES; + static final Integer ARRAY_INDEX_OFFSET = VALUE_TYPE_OFFSET + Integer.BYTES; + static final Integer ARRAY_LENGTH_OFFSET = ARRAY_INDEX_OFFSET + Integer.BYTES; + + OakMap<IncrementalIndexRow, Row> oak; + + private OffheapAggsManager aggsManager; + private AtomicInteger versionCounter; // Only used in Plain mode + + Map<String, String> env = System.getenv(); + + OakIncrementalIndex( + IncrementalIndexSchema incrementalIndexSchema, + boolean deserializeComplexMetrics, + boolean reportParseExceptions, + boolean concurrentEventAdd, + int maxRowCount + ) + { + super(incrementalIndexSchema, reportParseExceptions, maxRowCount); + + this.aggsManager = new OffheapAggsManager(incrementalIndexSchema, deserializeComplexMetrics, + reportParseExceptions, concurrentEventAdd, rowSupplier, + columnCapabilities, null, this); + + this.versionCounter = new AtomicInteger(0); + IncrementalIndexRow minIncrementalIndexRow = getMinIncrementalIndexRow(); + + long maxDirectMemory = VM.maxDirectMemory(); + int memoryCapacity = Integer.MAX_VALUE; + if (maxDirectMemory < memoryCapacity) { + memoryCapacity = (int) maxDirectMemory; + } + + OakMapBuilder builder = new OakMapBuilder() + .setKeySerializer(new OakKeySerializer(dimensionDescsList)) + .setValueSerializer(new OakValueSerializer(dimensionDescsList, aggsManager, reportParseExceptions, in)) + .setMinKey(minIncrementalIndexRow) + .setComparator(new OakKeysComparator(dimensionDescsList, this.isRollup())) + .setMemoryCapacity(memoryCapacity); + + if (env != null) { + String chunkMaxItems = env.get("chunkMaxItems"); + if (chunkMaxItems != null) { + builder = builder.setChunkMaxItems(Integer.getInteger(chunkMaxItems)); + } + String chunkBytesPerItem = env.get("chunkBytesPerItem"); + if (chunkMaxItems != null) { + builder = builder.setChunkBytesPerItem(Integer.getInteger(chunkBytesPerItem)); + } + } + + oak = builder.build(); + } + + @Override + public Iterable<Row> iterableWithPostAggregations(List<PostAggregator> postAggs, boolean descending) + { + Function<Map.Entry<ByteBuffer, ByteBuffer>, Row> transformer = new Function<Map.Entry<ByteBuffer, ByteBuffer>, Row>() + { + + @Override + public Row apply(Map.Entry<ByteBuffer, ByteBuffer> entry) + { + ByteBuffer serializedKey = entry.getKey(); + ByteBuffer serializedValue = entry.getValue(); + long timeStamp = OakIncrementalIndex.getTimestamp(serializedKey); + int dimsLength = OakIncrementalIndex.getDimsLength(serializedKey); + Map<String, Object> theVals = Maps.newLinkedHashMap(); + for (int i = 0; i < dimsLength; ++i) { + Object dim = OakIncrementalIndex.getDimValue(serializedKey, i); + DimensionDesc dimensionDesc = dimensionDescsList.get(i); + if (dimensionDesc == null) { + continue; + } + String dimensionName = dimensionDesc.getName(); + DimensionHandler handler = dimensionDesc.getHandler(); + if (dim == null || handler.getLengthOfEncodedKeyComponent(dim) == 0) { + theVals.put(dimensionName, null); + continue; + } + final DimensionIndexer indexer = dimensionDesc.getIndexer(); + Object rowVals = indexer.convertUnsortedEncodedKeyComponentToActualArrayOrList(dim, DimensionIndexer.LIST); + theVals.put(dimensionName, rowVals); + } + + BufferAggregator[] aggs = aggsManager.getAggs(); + for (int i = 0; i < aggs.length; ++i) { + theVals.put(aggsManager.metrics[i].getName(), aggs[i].get(serializedValue, aggsManager.aggOffsetInBuffer[i])); + } + + return new MapBasedRow(timeStamp, theVals); + } + }; + + OakMap tmpOakMap = descending ? oak.descendingMap() : oak; + OakTransformView transformView = tmpOakMap.createTransformView(transformer); + CloseableIterator<Row> valuesIterator = transformView.valuesIterator(); + return new Iterable<Row>() + { + @Override + public Iterator<Row> iterator() + { + return Iterators.transform( + valuesIterator, + row -> row + ); + } + }; + } + + @Override + public Iterable<IncrementalIndexRow> persistIterable() + { + return keySet(); + } + + @Override + public void close() {} + + @Override + protected long getMinTimeMillis() + { + return oak.getMinKey().getTimestamp(); + } + + @Override + protected long getMaxTimeMillis() + { + return oak.getMaxKey().getTimestamp(); + } + + @Override + protected BufferAggregator[] getAggsForRow(IncrementalIndexRow incrementalIndexRow) + { + return getAggs(); + } + + @Override + protected Object getAggVal(IncrementalIndexRow incrementalIndexRow, int aggIndex) + { + Function<Map.Entry<ByteBuffer, ByteBuffer>, Object> transformer = new Function<Map.Entry<ByteBuffer, ByteBuffer>, Object>() { + @Override + public Object apply(Map.Entry<ByteBuffer, ByteBuffer> entry) + { + ByteBuffer serializedValue = entry.getValue(); + BufferAggregator agg = getAggs()[aggIndex]; + return agg.get(serializedValue, serializedValue.position() + aggsManager.aggOffsetInBuffer[aggIndex]); + } + }; + + OakTransformView<IncrementalIndexRow, Object> transformView = (OakTransformView<IncrementalIndexRow, Object>) oak.createTransformView(transformer); + return transformView.get(incrementalIndexRow); + } + + @Override + protected float getMetricFloatValue(IncrementalIndexRow incrementalIndexRow, int aggIndex) + { + Function<Map.Entry<ByteBuffer, ByteBuffer>, Float> transformer = new Function<Map.Entry<ByteBuffer, ByteBuffer>, Float>() { + @Override + public Float apply(Map.Entry<ByteBuffer, ByteBuffer> entry) + { + ByteBuffer serializedValue = entry.getValue(); + BufferAggregator agg = getAggs()[aggIndex]; + return agg.getFloat(serializedValue, serializedValue.position() + aggsManager.aggOffsetInBuffer[aggIndex]); + } + }; + + OakTransformView<IncrementalIndexRow, Float> transformView = (OakTransformView<IncrementalIndexRow, Float>) oak.createTransformView(transformer); + return transformView.get(incrementalIndexRow); + } + + @Override + protected long getMetricLongValue(IncrementalIndexRow incrementalIndexRow, int aggIndex) + { + Function<Map.Entry<ByteBuffer, ByteBuffer>, Long> transformer = new Function<Map.Entry<ByteBuffer, ByteBuffer>, Long>() { + @Override + public Long apply(Map.Entry<ByteBuffer, ByteBuffer> entry) + { + ByteBuffer serializedValue = entry.getValue(); + BufferAggregator agg = getAggs()[aggIndex]; + return agg.getLong(serializedValue, serializedValue.position() + aggsManager.aggOffsetInBuffer[aggIndex]); + } + }; + + OakTransformView<IncrementalIndexRow, Long> transformView = (OakTransformView<IncrementalIndexRow, Long>) oak.createTransformView(transformer); + return transformView.get(incrementalIndexRow); + } + + @Override + protected Object getMetricObjectValue(IncrementalIndexRow incrementalIndexRow, int aggIndex) + { + Function<Map.Entry<ByteBuffer, ByteBuffer>, Object> transformer = new Function<Map.Entry<ByteBuffer, ByteBuffer>, Object>() { + @Override + public Object apply(Map.Entry<ByteBuffer, ByteBuffer> entry) + { + ByteBuffer serializedValue = entry.getValue(); + BufferAggregator agg = getAggs()[aggIndex]; + return agg.get(serializedValue, serializedValue.position() + aggsManager.aggOffsetInBuffer[aggIndex]); + } + }; + + OakTransformView<IncrementalIndexRow, Object> transformView = (OakTransformView<IncrementalIndexRow, Object>) oak.createTransformView(transformer); + return transformView.get(incrementalIndexRow); + } + + @Override + protected double getMetricDoubleValue(IncrementalIndexRow incrementalIndexRow, int aggIndex) + { + Function<Map.Entry<ByteBuffer, ByteBuffer>, Double> transformer = new Function<Map.Entry<ByteBuffer, ByteBuffer>, Double>() { + @Override + public Double apply(Map.Entry<ByteBuffer, ByteBuffer> entry) + { + ByteBuffer serializedValue = entry.getValue(); + BufferAggregator agg = getAggs()[aggIndex]; + return agg.getDouble(serializedValue, serializedValue.position() + aggsManager.aggOffsetInBuffer[aggIndex]); + } + }; + + OakTransformView<IncrementalIndexRow, Double> transformView = (OakTransformView<IncrementalIndexRow, Double>) oak.createTransformView(transformer); + return transformView.get(incrementalIndexRow); + } + + @Override + protected boolean isNull(IncrementalIndexRow incrementalIndexRow, int aggIndex) + { + Function<Map.Entry<ByteBuffer, ByteBuffer>, Boolean> transformer = new Function<Map.Entry<ByteBuffer, ByteBuffer>, Boolean>() { + @Override + public Boolean apply(Map.Entry<ByteBuffer, ByteBuffer> entry) + { + ByteBuffer serializedValue = entry.getValue(); + BufferAggregator agg = getAggs()[aggIndex]; + return agg.isNull(serializedValue, serializedValue.position() + aggsManager.aggOffsetInBuffer[aggIndex]); + } + }; + + OakTransformView<IncrementalIndexRow, Boolean> transformView = (OakTransformView<IncrementalIndexRow, Boolean>) oak.createTransformView(transformer); + return transformView.get(incrementalIndexRow); + } + + @Override + public Iterable<IncrementalIndexRow> timeRangeIterable( + boolean descending, long timeStart, long timeEnd) + { + if (timeStart > timeEnd) { + return null; + } + + IncrementalIndexRow from = new IncrementalIndexRow(timeStart, null, dimensionDescsList, IncrementalIndexRow.EMPTY_ROW_INDEX); + IncrementalIndexRow to = new IncrementalIndexRow(timeEnd + 1, null, dimensionDescsList, IncrementalIndexRow.EMPTY_ROW_INDEX); + OakMap subMap = oak.subMap(from, true, to, false); + if (descending == true) { + subMap = subMap.descendingMap(); + } + CloseableIterator<IncrementalIndexRow> keysIterator = subMap.keysIterator(); Review comment: No `close()` call for keysIter here. May cause GC block. ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@druid.apache.org For additional commands, e-mail: commits-h...@druid.apache.org