Hi,

Today, I was running a filter query ("SELECT * FROM widetable WHERE col_long_0 = 0") on a wide table with 1187 columns and Spark started printing the below output. It spills alot of log which I want to turn off. There is any option to turn it off. I have tried both option (ERROR,INFO) in log4j.properties file. It did not work for me.

Thank you

Regards
Faisal


17/05/24 12:39:41 INFO CarbonLateDecodeRule: main Starting to optimize plan
17/05/24 12:39:41 INFO CarbonLateDecodeRule: main Skip CarbonOptimizer
17/05/24 12:39:42 INFO deprecation: mapred.job.id is deprecated. Instead, use mapreduce.job.id 17/05/24 12:39:42 INFO deprecation: mapred.tip.id is deprecated. Instead, use mapreduce.task.id 17/05/24 12:39:42 INFO deprecation: mapred.task.id is deprecated. Instead, use mapreduce.task.attempt.id 17/05/24 12:39:42 INFO deprecation: mapred.task.is.map is deprecated. Instead, use mapreduce.task.ismap 17/05/24 12:39:42 INFO deprecation: mapred.task.partition is deprecated. Instead, use mapreduce.task.partition 17/05/24 12:39:42 INFO FileOutputCommitter: File Output Committer Algorithm version is 1 17/05/24 12:39:42 INFO SQLHadoopMapReduceCommitProtocol: Using output committer class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter 17/05/24 12:39:44 ERROR CodeGenerator: failed to compile: org.codehaus.janino.JaninoRuntimeException: Code of method "processNext()V" of class "org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator" grows beyond 64 KB
/* 001 */ public Object generate(Object[] references) {
/* 002 */   return new GeneratedIterator(references);
/* 003 */ }
/* 004 */
/* 005 */ final class GeneratedIterator extends org.apache.spark.sql.execution.BufferedRowIterator {
/* 006 */   private Object[] references;
/* 007 */   private scala.collection.Iterator[] inputs;
/* 008 */   private scala.collection.Iterator scan_input;
/* 009 */ private org.apache.spark.sql.execution.metric.SQLMetric scan_numOutputRows; /* 010 */ private org.apache.spark.sql.execution.metric.SQLMetric scan_scanTime;
/* 011 */   private long scan_scanTime1;
/* 012 */ private org.apache.spark.sql.execution.vectorized.ColumnarBatch scan_batch;
/* 013 */   private int scan_batchIdx;
/* 014 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance0; /* 015 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance1; /* 016 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance2; /* 017 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance3; /* 018 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance4; /* 019 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance5; /* 020 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance6; /* 021 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance7; /* 022 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance8; /* 023 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance9; /* 024 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance10; /* 025 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance11; /* 026 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance12; /* 027 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance13; /* 028 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance14; /* 029 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance15; /* 030 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance16; /* 031 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance17; /* 032 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance18; /* 033 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance19; /* 034 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance20; /* 035 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance21; /* 036 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance22; /* 037 */ private org.apache.spark.sql.execution.vectorized.ColumnVector scan_colInstance23;

Reply via email to