[jira] [Comment Edited] (SPARK-13832) TPC-DS Query 36 fails with Parser error
[ https://issues.apache.org/jira/browse/SPARK-13832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15210585#comment-15210585 ] Roy Cecil edited comment on SPARK-13832 at 3/24/16 5:16 PM: Xin, there are other queries with grouping__id that runs through without running into this problem ( Eg. 27 ). I do not think the ORDER BY alone is the issue why kyro exception is triggered. It is somehow in the context of this query. So lets keep this open until we get to the bottom of it. was (Author: roycecil): Xin, there are other queries with grouping__id that runs through without running into this problem. I do not think the ORDER BY alone is the issue why kyro exception is triggered. It is somehow in the context of this query. So lets keep this open until we get to the bottom of it. > TPC-DS Query 36 fails with Parser error > --- > > Key: SPARK-13832 > URL: https://issues.apache.org/jira/browse/SPARK-13832 > Project: Spark > Issue Type: Bug > Components: SQL >Affects Versions: 1.6.1 > Environment: Red Hat Enterprise Linux Server release 7.1 (Maipo) > Linux bigaperf116.svl.ibm.com 3.10.0-229.el7.x86_64 #1 SMP Thu Jan 29 > 18:37:38 EST 2015 x86_64 x86_64 x86_64 GNU/Linux >Reporter: Roy Cecil > > TPC-DS query 36 fails with the following error > Analyzer error: 16/02/28 21:22:51 INFO parse.ParseDriver: Parse Completed > Exception in thread "main" org.apache.spark.sql.AnalysisException: expression > 'i_category' is neither present in the group by, nor is it an aggregate > function. Add to group by or wrap in first() (or first_value) if you don't > care which value you get.; > at > org.apache.spark.sql.catalyst.analysis.CheckAnalysis$class.failAnalysis(CheckAnalysis.scala:38) > at > org.apache.spark.sql.catalyst.analysis.Analyzer.failAnalysis(Analyzer.scala:44) > Query Text pasted here for quick reference. > select > sum(ss_net_profit)/sum(ss_ext_sales_price) as gross_margin >,i_category >,i_class >,grouping__id as lochierarchy >,rank() over ( > partition by grouping__id, > case when grouping__id = 0 then i_category end > order by sum(ss_net_profit)/sum(ss_ext_sales_price) asc) as > rank_within_parent > from > store_sales >,date_dim d1 >,item >,store > where > d1.d_year = 2001 > and d1.d_date_sk = ss_sold_date_sk > and i_item_sk = ss_item_sk > and s_store_sk = ss_store_sk > and s_state in ('TN','TN','TN','TN', > 'TN','TN','TN','TN') > group by i_category,i_class WITH ROLLUP > order by >lochierarchy desc > ,case when lochierarchy = 0 then i_category end > ,rank_within_parent > limit 100; -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Comment Edited] (SPARK-13832) TPC-DS Query 36 fails with Parser error
[ https://issues.apache.org/jira/browse/SPARK-13832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15202412#comment-15202412 ] Xin Wu edited comment on SPARK-13832 at 3/19/16 12:29 AM: -- Jesse, you are right.. With "grouping" function, the query hits the {code}com.esotericsoftware.kryo.KryoException{code}, even with text data file. So this case, we passed the analyzer. With grouping_id on column i_category, the query hits the analyzer issue. {code}Error in query: Columns of grouping_id...{code} I will continue digging in. was (Author: xwu0226): Jesse, you are right.. With "grouping" function, the query hits the {code}com.esotericsoftware.kryo.KryoException{code}, even thought with text data file. So this case, we passed the analyzer. With grouping_id on column i_category, the query hits the analyzer issue. {code}Error in query: Columns of grouping_id...{code} I will continue digging in. > TPC-DS Query 36 fails with Parser error > --- > > Key: SPARK-13832 > URL: https://issues.apache.org/jira/browse/SPARK-13832 > Project: Spark > Issue Type: Bug > Components: SQL >Affects Versions: 1.6.1 > Environment: Red Hat Enterprise Linux Server release 7.1 (Maipo) > Linux bigaperf116.svl.ibm.com 3.10.0-229.el7.x86_64 #1 SMP Thu Jan 29 > 18:37:38 EST 2015 x86_64 x86_64 x86_64 GNU/Linux >Reporter: Roy Cecil > > TPC-DS query 36 fails with the following error > Analyzer error: 16/02/28 21:22:51 INFO parse.ParseDriver: Parse Completed > Exception in thread "main" org.apache.spark.sql.AnalysisException: expression > 'i_category' is neither present in the group by, nor is it an aggregate > function. Add to group by or wrap in first() (or first_value) if you don't > care which value you get.; > at > org.apache.spark.sql.catalyst.analysis.CheckAnalysis$class.failAnalysis(CheckAnalysis.scala:38) > at > org.apache.spark.sql.catalyst.analysis.Analyzer.failAnalysis(Analyzer.scala:44) > Query Text pasted here for quick reference. > select > sum(ss_net_profit)/sum(ss_ext_sales_price) as gross_margin >,i_category >,i_class >,grouping__id as lochierarchy >,rank() over ( > partition by grouping__id, > case when grouping__id = 0 then i_category end > order by sum(ss_net_profit)/sum(ss_ext_sales_price) asc) as > rank_within_parent > from > store_sales >,date_dim d1 >,item >,store > where > d1.d_year = 2001 > and d1.d_date_sk = ss_sold_date_sk > and i_item_sk = ss_item_sk > and s_store_sk = ss_store_sk > and s_state in ('TN','TN','TN','TN', > 'TN','TN','TN','TN') > group by i_category,i_class WITH ROLLUP > order by >lochierarchy desc > ,case when lochierarchy = 0 then i_category end > ,rank_within_parent > limit 100; -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Comment Edited] (SPARK-13832) TPC-DS Query 36 fails with Parser error
[ https://issues.apache.org/jira/browse/SPARK-13832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15196377#comment-15196377 ] Xin Wu edited comment on SPARK-13832 at 3/16/16 12:39 AM: -- Trying this query in Spark 2.0 and I needed to change grouping__id to grouping_id() to pass the parser. The reported error is not reproducible in spark 2.0.. Except that I saw execution error related to com.esotericsoftware.kryo.KryoException was (Author: xwu0226): Trying this query in Spark 2.0 and I needed to change grouping__id to grouping_id() to pass the parser. The reported error is not reproducible in spark 2.0.. Except that I saw execution error maybe related to spark-13862. > TPC-DS Query 36 fails with Parser error > --- > > Key: SPARK-13832 > URL: https://issues.apache.org/jira/browse/SPARK-13832 > Project: Spark > Issue Type: Bug > Components: SQL >Affects Versions: 1.6.1 > Environment: Red Hat Enterprise Linux Server release 7.1 (Maipo) > Linux bigaperf116.svl.ibm.com 3.10.0-229.el7.x86_64 #1 SMP Thu Jan 29 > 18:37:38 EST 2015 x86_64 x86_64 x86_64 GNU/Linux >Reporter: Roy Cecil > > TPC-DS query 36 fails with the following error > Analyzer error: 16/02/28 21:22:51 INFO parse.ParseDriver: Parse Completed > Exception in thread "main" org.apache.spark.sql.AnalysisException: expression > 'i_category' is neither present in the group by, nor is it an aggregate > function. Add to group by or wrap in first() (or first_value) if you don't > care which value you get.; > at > org.apache.spark.sql.catalyst.analysis.CheckAnalysis$class.failAnalysis(CheckAnalysis.scala:38) > at > org.apache.spark.sql.catalyst.analysis.Analyzer.failAnalysis(Analyzer.scala:44) > Query Text pasted here for quick reference. > select > sum(ss_net_profit)/sum(ss_ext_sales_price) as gross_margin >,i_category >,i_class >,grouping__id as lochierarchy >,rank() over ( > partition by grouping__id, > case when grouping__id = 0 then i_category end > order by sum(ss_net_profit)/sum(ss_ext_sales_price) asc) as > rank_within_parent > from > store_sales >,date_dim d1 >,item >,store > where > d1.d_year = 2001 > and d1.d_date_sk = ss_sold_date_sk > and i_item_sk = ss_item_sk > and s_store_sk = ss_store_sk > and s_state in ('TN','TN','TN','TN', > 'TN','TN','TN','TN') > group by i_category,i_class WITH ROLLUP > order by >lochierarchy desc > ,case when lochierarchy = 0 then i_category end > ,rank_within_parent > limit 100; -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Comment Edited] (SPARK-13832) TPC-DS Query 36 fails with Parser error
[ https://issues.apache.org/jira/browse/SPARK-13832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15196377#comment-15196377 ] Xin Wu edited comment on SPARK-13832 at 3/15/16 10:43 PM: -- Trying this query in Spark 2.0 and I needed to change grouping__id to grouping_id() to pass the parser. The reported error is not reproducible in spark 2.0.. Except that I saw execution error related to spark-13862. was (Author: xwu0226): Trying this query in Spark 2.0 and I needed to change grouping__id to grouping_id() to pass the parser. The reported error is gone.. Except that I saw execution error related to kryo.serializers.. that should be a different issue and maybe related to my setup. > TPC-DS Query 36 fails with Parser error > --- > > Key: SPARK-13832 > URL: https://issues.apache.org/jira/browse/SPARK-13832 > Project: Spark > Issue Type: Bug > Components: SQL >Affects Versions: 1.6.1 > Environment: Red Hat Enterprise Linux Server release 7.1 (Maipo) > Linux bigaperf116.svl.ibm.com 3.10.0-229.el7.x86_64 #1 SMP Thu Jan 29 > 18:37:38 EST 2015 x86_64 x86_64 x86_64 GNU/Linux >Reporter: Roy Cecil > > TPC-DS query 36 fails with the following error > Analyzer error: 16/02/28 21:22:51 INFO parse.ParseDriver: Parse Completed > Exception in thread "main" org.apache.spark.sql.AnalysisException: expression > 'i_category' is neither present in the group by, nor is it an aggregate > function. Add to group by or wrap in first() (or first_value) if you don't > care which value you get.; > at > org.apache.spark.sql.catalyst.analysis.CheckAnalysis$class.failAnalysis(CheckAnalysis.scala:38) > at > org.apache.spark.sql.catalyst.analysis.Analyzer.failAnalysis(Analyzer.scala:44) > Query Text pasted here for quick reference. > select > sum(ss_net_profit)/sum(ss_ext_sales_price) as gross_margin >,i_category >,i_class >,grouping__id as lochierarchy >,rank() over ( > partition by grouping__id, > case when grouping__id = 0 then i_category end > order by sum(ss_net_profit)/sum(ss_ext_sales_price) asc) as > rank_within_parent > from > store_sales >,date_dim d1 >,item >,store > where > d1.d_year = 2001 > and d1.d_date_sk = ss_sold_date_sk > and i_item_sk = ss_item_sk > and s_store_sk = ss_store_sk > and s_state in ('TN','TN','TN','TN', > 'TN','TN','TN','TN') > group by i_category,i_class WITH ROLLUP > order by >lochierarchy desc > ,case when lochierarchy = 0 then i_category end > ,rank_within_parent > limit 100; -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Comment Edited] (SPARK-13832) TPC-DS Query 36 fails with Parser error
[ https://issues.apache.org/jira/browse/SPARK-13832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15196377#comment-15196377 ] Xin Wu edited comment on SPARK-13832 at 3/15/16 10:44 PM: -- Trying this query in Spark 2.0 and I needed to change grouping__id to grouping_id() to pass the parser. The reported error is not reproducible in spark 2.0.. Except that I saw execution error maybe related to spark-13862. was (Author: xwu0226): Trying this query in Spark 2.0 and I needed to change grouping__id to grouping_id() to pass the parser. The reported error is not reproducible in spark 2.0.. Except that I saw execution error related to spark-13862. > TPC-DS Query 36 fails with Parser error > --- > > Key: SPARK-13832 > URL: https://issues.apache.org/jira/browse/SPARK-13832 > Project: Spark > Issue Type: Bug > Components: SQL >Affects Versions: 1.6.1 > Environment: Red Hat Enterprise Linux Server release 7.1 (Maipo) > Linux bigaperf116.svl.ibm.com 3.10.0-229.el7.x86_64 #1 SMP Thu Jan 29 > 18:37:38 EST 2015 x86_64 x86_64 x86_64 GNU/Linux >Reporter: Roy Cecil > > TPC-DS query 36 fails with the following error > Analyzer error: 16/02/28 21:22:51 INFO parse.ParseDriver: Parse Completed > Exception in thread "main" org.apache.spark.sql.AnalysisException: expression > 'i_category' is neither present in the group by, nor is it an aggregate > function. Add to group by or wrap in first() (or first_value) if you don't > care which value you get.; > at > org.apache.spark.sql.catalyst.analysis.CheckAnalysis$class.failAnalysis(CheckAnalysis.scala:38) > at > org.apache.spark.sql.catalyst.analysis.Analyzer.failAnalysis(Analyzer.scala:44) > Query Text pasted here for quick reference. > select > sum(ss_net_profit)/sum(ss_ext_sales_price) as gross_margin >,i_category >,i_class >,grouping__id as lochierarchy >,rank() over ( > partition by grouping__id, > case when grouping__id = 0 then i_category end > order by sum(ss_net_profit)/sum(ss_ext_sales_price) asc) as > rank_within_parent > from > store_sales >,date_dim d1 >,item >,store > where > d1.d_year = 2001 > and d1.d_date_sk = ss_sold_date_sk > and i_item_sk = ss_item_sk > and s_store_sk = ss_store_sk > and s_state in ('TN','TN','TN','TN', > 'TN','TN','TN','TN') > group by i_category,i_class WITH ROLLUP > order by >lochierarchy desc > ,case when lochierarchy = 0 then i_category end > ,rank_within_parent > limit 100; -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Comment Edited] (SPARK-13832) TPC-DS Query 36 fails with Parser error
[ https://issues.apache.org/jira/browse/SPARK-13832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15196377#comment-15196377 ] Xin Wu edited comment on SPARK-13832 at 3/15/16 10:41 PM: -- Trying this query in Spark 2.0 and I needed to change grouping__id to grouping_id() to pass the parser. The reported error is gone.. Except that I saw execution error related to kryo.serializers.. that should be a different issue and maybe related to my setup. was (Author: xwu0226): Trying this query in Spark 2.0 and I needed to change grouping__id to grouping_id() to pass the parser. The reported error is not gone.. Except that I saw execution error related to kryo.serializers.. that should be a different issue and maybe related to my setup. > TPC-DS Query 36 fails with Parser error > --- > > Key: SPARK-13832 > URL: https://issues.apache.org/jira/browse/SPARK-13832 > Project: Spark > Issue Type: Bug > Components: SQL >Affects Versions: 1.6.1 > Environment: Red Hat Enterprise Linux Server release 7.1 (Maipo) > Linux bigaperf116.svl.ibm.com 3.10.0-229.el7.x86_64 #1 SMP Thu Jan 29 > 18:37:38 EST 2015 x86_64 x86_64 x86_64 GNU/Linux >Reporter: Roy Cecil > > TPC-DS query 36 fails with the following error > Analyzer error: 16/02/28 21:22:51 INFO parse.ParseDriver: Parse Completed > Exception in thread "main" org.apache.spark.sql.AnalysisException: expression > 'i_category' is neither present in the group by, nor is it an aggregate > function. Add to group by or wrap in first() (or first_value) if you don't > care which value you get.; > at > org.apache.spark.sql.catalyst.analysis.CheckAnalysis$class.failAnalysis(CheckAnalysis.scala:38) > at > org.apache.spark.sql.catalyst.analysis.Analyzer.failAnalysis(Analyzer.scala:44) > Query Text pasted here for quick reference. > select > sum(ss_net_profit)/sum(ss_ext_sales_price) as gross_margin >,i_category >,i_class >,grouping__id as lochierarchy >,rank() over ( > partition by grouping__id, > case when grouping__id = 0 then i_category end > order by sum(ss_net_profit)/sum(ss_ext_sales_price) asc) as > rank_within_parent > from > store_sales >,date_dim d1 >,item >,store > where > d1.d_year = 2001 > and d1.d_date_sk = ss_sold_date_sk > and i_item_sk = ss_item_sk > and s_store_sk = ss_store_sk > and s_state in ('TN','TN','TN','TN', > 'TN','TN','TN','TN') > group by i_category,i_class WITH ROLLUP > order by >lochierarchy desc > ,case when lochierarchy = 0 then i_category end > ,rank_within_parent > limit 100; -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org