Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1642#discussion_r157137538 --- Diff: hadoop/src/test/java/org/apache/carbondata/hadoop/ft/CarbonOutputMapperTest.java --- @@ -0,0 +1,122 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.carbondata.hadoop.ft; + +import java.io.File; +import java.io.FilenameFilter; +import java.io.IOException; + +import org.apache.carbondata.core.constants.CarbonCommonConstants; +import org.apache.carbondata.core.util.CarbonProperties; +import org.apache.carbondata.core.util.CarbonUtil; +import org.apache.carbondata.core.util.path.CarbonTablePath; +import org.apache.carbondata.hadoop.api.CarbonTableOutputFormat; +import org.apache.carbondata.hadoop.test.util.StoreCreator; +import org.apache.carbondata.processing.loading.csvinput.CSVInputFormat; +import org.apache.carbondata.processing.loading.csvinput.StringArrayWritable; +import org.apache.carbondata.processing.loading.model.CarbonLoadModel; + +import junit.framework.TestCase; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.fs.Path; +import org.apache.hadoop.io.NullWritable; +import org.apache.hadoop.mapreduce.Job; +import org.apache.hadoop.mapreduce.Mapper; +import org.apache.hadoop.mapreduce.lib.input.FileInputFormat; +import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat; +import org.junit.Test; + +public class CarbonOutputMapperTest extends TestCase { + + CarbonLoadModel carbonLoadModel; + + // changed setUp to static init block to avoid un wanted multiple time store creation + static { + CarbonProperties.getInstance(). + addProperty(CarbonCommonConstants.CARBON_BADRECORDS_LOC, "/tmp/carbon/badrecords"); + } + + + @Test public void testOutputFormat() throws Exception { + runJob(""); + String segmentPath = CarbonTablePath.getSegmentPath(carbonLoadModel.getTablePath(), "0"); + File file = new File(segmentPath); + assert (file.exists()); + File[] listFiles = file.listFiles(new FilenameFilter() { + @Override public boolean accept(File dir, String name) { + return name.endsWith(".carbondata") || name.endsWith(".carbonindex"); + } + }); + + assert (listFiles.length == 2); + + } + + + @Override public void tearDown() throws Exception { + super.tearDown(); + CarbonProperties.getInstance() + .addProperty(CarbonCommonConstants.ENABLE_QUERY_STATISTICS, "true"); + } + + @Override public void setUp() throws Exception { + super.setUp(); + CarbonProperties.getInstance() + .addProperty(CarbonCommonConstants.ENABLE_QUERY_STATISTICS, "false"); + carbonLoadModel = StoreCreator.getCarbonLoadModel(); + } + + public static class Map extends Mapper<NullWritable, StringArrayWritable, NullWritable, StringArrayWritable> { + + @Override protected void map(NullWritable key, StringArrayWritable value, Context context) --- End diff -- move @Override to previous line
---