[ https://issues.apache.org/jira/browse/DRILL-5459?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15999113#comment-15999113 ]
ASF GitHub Bot commented on DRILL-5459: --------------------------------------- Github user paul-rogers commented on a diff in the pull request: https://github.com/apache/drill/pull/823#discussion_r115103967 --- Diff: exec/java-exec/src/test/java/org/apache/drill/exec/physical/unit/TestMiniPlan.java --- @@ -0,0 +1,200 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * <p/> + * http://www.apache.org/licenses/LICENSE-2.0 + * <p/> + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.drill.exec.physical.unit; + +import com.google.common.collect.Lists; +import org.apache.drill.common.types.TypeProtos; +import org.apache.drill.common.util.FileUtils; +import org.apache.drill.exec.physical.config.Filter; +import org.apache.drill.exec.physical.config.UnionAll; +import org.apache.drill.exec.record.BatchSchema; +import org.apache.drill.exec.record.RecordBatch; +import org.apache.drill.exec.store.dfs.DrillFileSystem; +import org.apache.drill.test.rowSet.SchemaBuilder; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.fs.FileSystem; +import org.junit.BeforeClass; +import org.junit.Ignore; +import org.junit.Test; + +import java.util.Collections; +import java.util.List; + +public class TestMiniPlan extends MiniPlanUnitTestBase { + + protected static DrillFileSystem fs; + + @BeforeClass + public static void initFS() throws Exception { + Configuration conf = new Configuration(); + conf.set(FileSystem.FS_DEFAULT_NAME_KEY, "local"); + fs = new DrillFileSystem(conf); + } + + @Test + @Ignore("A bug in JsonRecordReader handling empty file") + public void testEmptyInput() throws Exception { + String emptyFile = FileUtils.getResourceAsFile("/project/pushdown/empty.json").toURI().toString(); + + RecordBatch scanBatch = new JsonScanBuilder() + .fileSystem(fs) + .inputPaths(Lists.newArrayList(emptyFile)) + .build(); + + new MiniPlanTestBuilder() + .root(scanBatch) + .expectZeroBatch(true) + .go(); + } + + @Test + public void testSimpleParquetScan() throws Exception { + String file = FileUtils.getResourceAsFile("/tpchmulti/region/01.parquet").toURI().toString(); + + RecordBatch scanBatch = new ParquetScanBuilder() + .fileSystem(fs) + .columnsToRead("R_REGIONKEY") + .inputPaths(Lists.newArrayList(file)) + .build(); + + BatchSchema expectedSchema = new SchemaBuilder() + .add("R_REGIONKEY", TypeProtos.MinorType.BIGINT) + .build(); + + new MiniPlanTestBuilder() + .root(scanBatch) + .expectedSchema(expectedSchema) + .baselineValues(0L) + .baselineValues(1L) + .go(); + } + + @Test + public void testSimpleJson() throws Exception { + List<String> jsonBatches = Lists.newArrayList( + "{\"a\":100}" + ); + + RecordBatch scanBatch = new JsonScanBuilder() + .jsonBatches(jsonBatches) + .build(); + + BatchSchema expectedSchema = new SchemaBuilder() + .addNullable("a", TypeProtos.MinorType.BIGINT) + .build(); + + new MiniPlanTestBuilder() + .root(scanBatch) + .expectedSchema(expectedSchema) + .baselineValues(100L) + .go(); + } + + @Test + public void testUnionFilter() throws Exception { + List<String> leftJsonBatches = Lists.newArrayList( + "[{\"a\": 5, \"b\" : 1 }]", + "[{\"a\": 5, \"b\" : 5},{\"a\": 3, \"b\" : 8}]", + "[{\"a\": 40, \"b\" : 3},{\"a\": 13, \"b\" : 100}]"); + + List<String> rightJsonBatches = Lists.newArrayList( + "[{\"a\": 5, \"b\" : 10 }]", + "[{\"a\": 50, \"b\" : 100}]"); + + RecordBatch batch = new PopBuilder() + .physicalOperator(new UnionAll(Collections.EMPTY_LIST)) // Children list is provided through RecordBatch + .addInputAsChild() + .physicalOperator(new Filter(null, parseExpr("a=5"), 1.0f)) + .addJsonScanAsChild() + .jsonBatches(leftJsonBatches) + .columnsToRead("a", "b") + .buildAddAsInput() + .buildAddAsInput() + .addInputAsChild() + .physicalOperator(new Filter(null, parseExpr("a=50"), 1.0f)) + .addJsonScanAsChild() + .jsonBatches(rightJsonBatches) + .columnsToRead("a", "b") + .buildAddAsInput() + .buildAddAsInput() + .build(); + + BatchSchema expectedSchema = new SchemaBuilder() + .addNullable("a", TypeProtos.MinorType.BIGINT) + .addNullable("b", TypeProtos.MinorType.BIGINT) + .withSVMode(BatchSchema.SelectionVectorMode.NONE) + .build(); + + new MiniPlanTestBuilder() + .root(batch) + .expectedSchema(expectedSchema) + .baselineValues(5l, 1l) + .baselineValues(5l, 5l) + .baselineValues(50l, 100l) + .go(); + } + + @Test + @Ignore ("A bug in UnionAll handling empty inputs from both sides") --- End diff -- Again, file a JIRA and put ticket number here? > Extend physical operator test framework to test mini plans consisting of > multiple operators > ------------------------------------------------------------------------------------------- > > Key: DRILL-5459 > URL: https://issues.apache.org/jira/browse/DRILL-5459 > Project: Apache Drill > Issue Type: Improvement > Components: Tools, Build & Test > Reporter: Jinfeng Ni > Assignee: Jinfeng Ni > > DRILL-4437 introduced a unit test framework to test a non-scan physical > operator. A JSON reader is implicitly used to specify the inputs to the > physical operator under test. > There are needs to extend such unit test framework for two scenarios. > 1. We need a way to test scan operator with different record readers. Drill > supports a variety of data source, and it's important to make sure every > record reader work properly according to the protocol defined. > 2. We need a way to test a so-called mini-plan (aka plan fragment) consisting > of multiple non-scan operators. > For the 2nd need, an alternative is to leverage SQL statement and query > planner. However, such approach has a direct dependency on query planner; 1) > any planner change may impact the testcase and lead to a different plan, 2) > it's not always easy job to force the planner to get a desired plan fragment > for testing. > In particular, it would be good to have a relatively easy way to specify a > mini-plan with a couple of targeted physical operators. > This JIRA is created to track the work to extend the unit test framework in > DRILL-4437. > > Related work: DRILL-5318 introduced a sub-operator test fixture, which mainly > targeted to test at sub-operator level. The framework in DRILL-4437 and the > extension would focus on operator level, or multiple operator levels, where > execution would go through RecordBatch's API call. > Same as DRILL-4437, we are going to use mockit to mock required objects such > fragment context, operator context etc. -- This message was sent by Atlassian JIRA (v6.3.15#6346)