[ 
https://issues.apache.org/jira/browse/DRILL-5432?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16065708#comment-16065708
 ] 

ASF GitHub Bot commented on DRILL-5432:
---------------------------------------

Github user paul-rogers commented on a diff in the pull request:

    https://github.com/apache/drill/pull/831#discussion_r124427466
  
    --- Diff: 
exec/java-exec/src/main/java/org/apache/drill/exec/store/pcap/PcapRecordReader.java
 ---
    @@ -0,0 +1,295 @@
    +/*
    + * Licensed to the Apache Software Foundation (ASF) under one or more
    + * contributor license agreements.  See the NOTICE file distributed with
    + * this work for additional information regarding copyright ownership.
    + * The ASF licenses this file to you under the Apache License, Version 2.0
    + * (the "License"); you may not use this file except in compliance with
    + * the License.  You may obtain a copy of the License at
    + * <p>
    + * http://www.apache.org/licenses/LICENSE-2.0
    + * <p>
    + * Unless required by applicable law or agreed to in writing, software
    + * distributed under the License is distributed on an "AS IS" BASIS,
    + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
    + * See the License for the specific language governing permissions and
    + * limitations under the License.
    + */
    +package org.apache.drill.exec.store.pcap;
    +
    +import com.google.common.collect.ImmutableList;
    +import com.google.common.collect.ImmutableMap;
    +import org.apache.drill.common.exceptions.ExecutionSetupException;
    +import org.apache.drill.common.expression.SchemaPath;
    +import org.apache.drill.common.types.TypeProtos;
    +import org.apache.drill.common.types.TypeProtos.MajorType;
    +import org.apache.drill.common.types.TypeProtos.MinorType;
    +import org.apache.drill.common.types.Types;
    +import org.apache.drill.exec.exception.SchemaChangeException;
    +import org.apache.drill.exec.expr.TypeHelper;
    +import org.apache.drill.exec.ops.OperatorContext;
    +import org.apache.drill.exec.physical.impl.OutputMutator;
    +import org.apache.drill.exec.record.MaterializedField;
    +import org.apache.drill.exec.store.AbstractRecordReader;
    +import org.apache.drill.exec.store.pcap.decoder.Packet;
    +import org.apache.drill.exec.store.pcap.decoder.PacketDecoder;
    +import org.apache.drill.exec.store.pcap.dto.ColumnDto;
    +import org.apache.drill.exec.store.pcap.schema.PcapTypes;
    +import org.apache.drill.exec.store.pcap.schema.Schema;
    +import org.apache.drill.exec.vector.NullableBigIntVector;
    +import org.apache.drill.exec.vector.NullableIntVector;
    +import org.apache.drill.exec.vector.NullableTimeStampVector;
    +import org.apache.drill.exec.vector.NullableVarCharVector;
    +import org.apache.drill.exec.vector.ValueVector;
    +
    +import java.io.FileInputStream;
    +import java.io.IOException;
    +import java.io.InputStream;
    +import java.nio.ByteBuffer;
    +import java.util.List;
    +import java.util.Map;
    +
    +import static java.nio.charset.StandardCharsets.UTF_8;
    +import static org.apache.drill.exec.store.pcap.Utils.parseBytesToASCII;
    +
    +public class PcapRecordReader extends AbstractRecordReader {
    +
    +  private OutputMutator output;
    +
    +  private final PacketDecoder decoder;
    +  private ImmutableList<ProjectedColumnInfo> projectedCols;
    +
    +  private byte[] buffer = new byte[100000];
    +  private int offset = 0;
    +  private InputStream in;
    +  private int validBytes;
    +
    +  private static final Map<PcapTypes, MinorType> TYPES;
    +
    +  private static class ProjectedColumnInfo {
    +    ValueVector vv;
    +    ColumnDto pcapColumn;
    +  }
    +
    +  static {
    +    TYPES = ImmutableMap.<PcapTypes, TypeProtos.MinorType>builder()
    +        .put(PcapTypes.STRING, MinorType.VARCHAR)
    +        .put(PcapTypes.INTEGER, MinorType.INT)
    +        .put(PcapTypes.LONG, MinorType.BIGINT)
    +        .put(PcapTypes.TIMESTAMP, MinorType.TIMESTAMP)
    +        .build();
    +  }
    +
    +  public PcapRecordReader(final String inputPath,
    +                          final List<SchemaPath> projectedColumns) {
    +    try {
    +      this.in = new FileInputStream(inputPath);
    +      this.decoder = getPacketDecoder();
    +      validBytes = in.read(buffer);
    +    } catch (IOException e) {
    +      throw new RuntimeException("File " + inputPath + " not Found");
    +    }
    +    setColumns(projectedColumns);
    +  }
    +
    +  @Override
    +  public void setup(final OperatorContext context, final OutputMutator 
output) throws ExecutionSetupException {
    +    this.output = output;
    +  }
    +
    +  @Override
    +  public int next() {
    +    projectedCols = getProjectedColsIfItNull();
    +    try {
    +      return parsePcapFilesAndPutItToTable();
    +    } catch (IOException io) {
    +      throw new RuntimeException("Trouble with reading packets in file!");
    +    }
    +  }
    +
    +  @Override
    +  public void close() throws Exception {
    +  }
    +
    +  private PacketDecoder getPacketDecoder() {
    +    try {
    +      return new PacketDecoder(in);
    +    } catch (IOException io) {
    +      throw new RuntimeException("File Not Found or some I/O issue");
    +    }
    +  }
    +
    +  private ImmutableList<ProjectedColumnInfo> getProjectedColsIfItNull() {
    +    return projectedCols != null ? projectedCols : initCols(new Schema());
    +  }
    +
    +  private ImmutableList<ProjectedColumnInfo> initCols(final Schema schema) 
{
    +    ImmutableList.Builder<ProjectedColumnInfo> pciBuilder = 
ImmutableList.builder();
    +    ColumnDto column;
    +
    +    for (int i = 0; i < schema.getNumberOfColumns(); i++) {
    +      column = schema.getColumnByIndex(i);
    +
    +      final String name = column.getColumnName();
    +      final PcapTypes type = column.getColumnType();
    +      TypeProtos.MinorType minorType = TYPES.get(type);
    +
    +      ProjectedColumnInfo pci = getProjectedColumnInfo(column, name, 
minorType);
    +      pciBuilder.add(pci);
    +    }
    +    return pciBuilder.build();
    +  }
    +
    +  private ProjectedColumnInfo getProjectedColumnInfo(final ColumnDto 
column,
    --- End diff --
    
    I stand corrected. Couldn't find that functionality when reviewing; should 
have downloaded the code and tried it.


> Want a memory format for PCAP files
> -----------------------------------
>
>                 Key: DRILL-5432
>                 URL: https://issues.apache.org/jira/browse/DRILL-5432
>             Project: Apache Drill
>          Issue Type: New Feature
>            Reporter: Ted Dunning
>
> PCAP files [1] are the de facto standard for storing network capture data. In 
> security and protocol applications, it is very common to want to extract 
> particular packets from a capture for further analysis.
> At a first level, it is desirable to query and filter by source and 
> destination IP and port or by protocol. Beyond that, however, it would be 
> very useful to be able to group packets by TCP session and eventually to look 
> at packet contents. For now, however, the most critical requirement is that 
> we should be able to scan captures at very high speed.
> I previously wrote a (kind of working) proof of concept for a PCAP decoder 
> that did lazy deserialization and could traverse hundreds of MB of PCAP data 
> per second per core. This compares to roughly 2-3 MB/s for widely available 
> Apache-compatible open source PCAP decoders.
> This JIRA covers the integration and extension of that proof of concept as a 
> Drill file format.
> Initial work is available at https://github.com/mapr-demos/drill-pcap-format
> [1] https://en.wikipedia.org/wiki/Pcap



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

Reply via email to