[ https://issues.apache.org/jira/browse/SPARK-9955?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Apache Spark reassigned SPARK-9955: ----------------------------------- Assignee: Apache Spark (was: Wenchen Fan) > TPCDS Q8 failed in 1.5 > ----------------------- > > Key: SPARK-9955 > URL: https://issues.apache.org/jira/browse/SPARK-9955 > Project: Spark > Issue Type: Bug > Components: SQL > Reporter: Davies Liu > Assignee: Apache Spark > Priority: Critical > > {code} > -- start query 1 in stream 0 using template query8.tpl > select > s_store_name, > sum(ss_net_profit) > from > store_sales > join store on (store_sales.ss_store_sk = store.s_store_sk) > -- join date_dim on (store_sales.ss_sold_date_sk = date_dim.d_date_sk) > join > (select > a.ca_zip > from > (select > substr(ca_zip, 1, 5) ca_zip, > count( *) cnt > from > customer_address > join customer on (customer_address.ca_address_sk = > customer.c_current_addr_sk) > where > c_preferred_cust_flag = 'Y' > group by > ca_zip > having > count( *) > 10 > ) a > left semi join > (select > substr(ca_zip, 1, 5) ca_zip > from > customer_address > where > substr(ca_zip, 1, 5) in ('89436', '30868', '65085', '22977', '83927', > '77557', '58429', '40697', '80614', '10502', '32779', > '91137', '61265', '98294', '17921', '18427', '21203', '59362', '87291', > '84093', '21505', '17184', '10866', '67898', '25797', > '28055', '18377', '80332', '74535', '21757', '29742', '90885', '29898', > '17819', '40811', '25990', '47513', '89531', '91068', > '10391', '18846', '99223', '82637', '41368', '83658', '86199', '81625', > '26696', '89338', '88425', '32200', '81427', '19053', > '77471', '36610', '99823', '43276', '41249', '48584', '83550', '82276', > '18842', '78890', '14090', '38123', '40936', '34425', > '19850', '43286', '80072', '79188', '54191', '11395', '50497', '84861', > '90733', '21068', '57666', '37119', '25004', '57835', > '70067', '62878', '95806', '19303', '18840', '19124', '29785', '16737', > '16022', '49613', '89977', '68310', '60069', '98360', > '48649', '39050', '41793', '25002', '27413', '39736', '47208', '16515', > '94808', '57648', '15009', '80015', '42961', '63982', > '21744', '71853', '81087', '67468', '34175', '64008', '20261', '11201', > '51799', '48043', '45645', '61163', '48375', '36447', > '57042', '21218', '41100', '89951', '22745', '35851', '83326', '61125', > '78298', '80752', '49858', '52940', '96976', '63792', > '11376', '53582', '18717', '90226', '50530', '94203', '99447', '27670', > '96577', '57856', '56372', '16165', '23427', '54561', > '28806', '44439', '22926', '30123', '61451', '92397', '56979', '92309', > '70873', '13355', '21801', '46346', '37562', '56458', > '28286', '47306', '99555', '69399', '26234', '47546', '49661', '88601', > '35943', '39936', '25632', '24611', '44166', '56648', > '30379', '59785', '11110', '14329', '93815', '52226', '71381', '13842', > '25612', '63294', '14664', '21077', '82626', '18799', > '60915', '81020', '56447', '76619', '11433', '13414', '42548', '92713', > '70467', '30884', '47484', '16072', '38936', '13036', > '88376', '45539', '35901', '19506', '65690', '73957', '71850', '49231', > '14276', '20005', '18384', '76615', '11635', '38177', > '55607', '41369', '95447', '58581', '58149', '91946', '33790', '76232', > '75692', '95464', '22246', '51061', '56692', '53121', > '77209', '15482', '10688', '14868', '45907', '73520', '72666', '25734', > '17959', '24677', '66446', '94627', '53535', '15560', > '41967', '69297', '11929', '59403', '33283', '52232', '57350', '43933', > '40921', '36635', '10827', '71286', '19736', '80619', > '25251', '95042', '15526', '36496', '55854', '49124', '81980', '35375', > '49157', '63512', '28944', '14946', '36503', '54010', > '18767', '23969', '43905', '66979', '33113', '21286', '58471', '59080', > '13395', '79144', '70373', '67031', '38360', '26705', > '50906', '52406', '26066', '73146', '15884', '31897', '30045', '61068', > '45550', '92454', '13376', '14354', '19770', '22928', > '97790', '50723', '46081', '30202', '14410', '20223', '88500', '67298', > '13261', '14172', '81410', '93578', '83583', '46047', > '94167', '82564', '21156', '15799', '86709', '37931', '74703', '83103', > '23054', '70470', '72008', '49247', '91911', '69998', > '20961', '70070', '63197', '54853', '88191', '91830', '49521', '19454', > '81450', '89091', '62378', '25683', '61869', '51744', > '36580', '85778', '36871', '48121', '28810', '83712', '45486', '67393', > '26935', '42393', '20132', '55349', '86057', '21309', > '80218', '10094', '11357', '48819', '39734', '40758', '30432', '21204', > '29467', '30214', '61024', '55307', '74621', '11622', > '68908', '33032', '52868', '99194', '99900', '84936', '69036', '99149', > '45013', '32895', '59004', '32322', '14933', '32936', > '33562', '72550', '27385', '58049', '58200', '16808', '21360', '32961', > '18586', '79307', '15492') > ) b > on (a.ca_zip = b.ca_zip) > ) v1 on (substr(store.s_zip, 1, 2) = substr(v1.ca_zip, 1, 2)) > where > ss_date between '2002-01-01' and '2002-04-01' > -- and d_qoy = 1 > -- and d_year = 2002 > group by > s_store_name > order by > s_store_name > limit 100 > -- end query 1 in stream 0 using template query8.tpl > {code} > {code} > org.apache.spark.sql.catalyst.analysis.UnresolvedException: Invalid call to > toAttribute on unresolved object, tree: unresolvedalias('s_store_name) > at > org.apache.spark.sql.catalyst.analysis.UnresolvedAlias.toAttribute(unresolved.scala:257) > at > org.apache.spark.sql.catalyst.plans.logical.Aggregate$$anonfun$output$6.apply(basicOperators.scala:220) > at > org.apache.spark.sql.catalyst.plans.logical.Aggregate$$anonfun$output$6.apply(basicOperators.scala:220) > at > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244) > at > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244) > at scala.collection.immutable.List.foreach(List.scala:318) > at scala.collection.TraversableLike$class.map(TraversableLike.scala:244) > at scala.collection.AbstractTraversable.map(Traversable.scala:105) > at > org.apache.spark.sql.catalyst.plans.logical.Aggregate.output(basicOperators.scala:220) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$PullOutNondeterministic$$anonfun$apply$18.applyOrElse(Analyzer.scala:940) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$PullOutNondeterministic$$anonfun$apply$18.applyOrElse(Analyzer.scala:933) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:57) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:57) > at > org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:51) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:56) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:54) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:54) > at > org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:249) > at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) > at scala.collection.Iterator$class.foreach(Iterator.scala:727) > at scala.collection.AbstractIterator.foreach(Iterator.scala:1157) > at > scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48) > at > scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103) > at > scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47) > at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273) > at scala.collection.AbstractIterator.to(Iterator.scala:1157) > at > scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265) > at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157) > at > scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252) > at scala.collection.AbstractIterator.toArray(Iterator.scala:1157) > at > org.apache.spark.sql.catalyst.trees.TreeNode.transformChildren(TreeNode.scala:279) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:54) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:54) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:54) > at > org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:249) > at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) > at scala.collection.Iterator$class.foreach(Iterator.scala:727) > at scala.collection.AbstractIterator.foreach(Iterator.scala:1157) > at > scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48) > at > scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103) > at > scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47) > at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273) > at scala.collection.AbstractIterator.to(Iterator.scala:1157) > at > scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265) > at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157) > at > scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252) > at scala.collection.AbstractIterator.toArray(Iterator.scala:1157) > at > org.apache.spark.sql.catalyst.trees.TreeNode.transformChildren(TreeNode.scala:279) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:54) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$PullOutNondeterministic$.apply(Analyzer.scala:933) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$PullOutNondeterministic$.apply(Analyzer.scala:932) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:83) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:80) > at > scala.collection.IndexedSeqOptimized$class.foldl(IndexedSeqOptimized.scala:51) > at > scala.collection.IndexedSeqOptimized$class.foldLeft(IndexedSeqOptimized.scala:60) > at scala.collection.mutable.WrappedArray.foldLeft(WrappedArray.scala:34) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:80) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:72) > at scala.collection.immutable.List.foreach(List.scala:318) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:72) > at > org.apache.spark.sql.SQLContext$QueryExecution.analyzed$lzycompute(SQLContext.scala:983) > at > org.apache.spark.sql.SQLContext$QueryExecution.analyzed(SQLContext.scala:983) > at > org.apache.spark.sql.SQLContext$QueryExecution.assertAnalyzed(SQLContext.scala:981) > at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:132) > at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:51) > at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:795) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:231) > at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:379) > at py4j.Gateway.invoke(Gateway.java:259) > at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133) > at py4j.commands.CallCommand.execute(CallCommand.java:79) > at py4j.GatewayConnection.run(GatewayConnection.java:207) > at java.lang.Thread.run(Thread.java:745) > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org