[ https://issues.apache.org/jira/browse/SPARK-6898?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14640704#comment-14640704 ]
David Sabater commented on SPARK-6898: -------------------------------------- I hope I am missing something here but I am facing this issue in my current cluster (Currently running compiled version of Spark 1.4.0 with Hive support) (I tried with 1.4.0 and 1.4.1 binaries locally and had same issue). sqlContext.jsonRDD(sc.makeRDD("""{"a": {"c.b": 1}, "b.$q": [{"a@!.q": 1}], "q.w": {"w.i&": [1]}}""" :: Nil)).registerTempTable("t") sqlContext.sql("SELECT `key?number1`, `key.number2` FROM records") sqlContext.sql("SELECT a.`c.b`, `b.$q`[0].`a@!.q`, `q.w`.`w.i&`[0] FROM t") Either select is throwing the same error: scala> sqlContext.sql("SELECT a.`c.b`, `b.$q`[0].`a@!.q`, `q.w`.`w.i&`[0] FROM t") 15/07/24 17:23:12 INFO ParseDriver: Parsing command: SELECT a.`c.b`, `b.$q`[0].`a@!.q`, `q.w`.`w.i&`[0] FROM t 15/07/24 17:23:12 INFO ParseDriver: Parse Completed org.apache.spark.sql.AnalysisException: cannot resolve 'b.$q' given input columns a, b.$q, q.w; line 1 pos 16 at org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.failAnalysis(package.scala:42) at org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$checkAnalysis$1$$anonfun$apply$2.applyOrElse(CheckAnalysis. scala:63) at org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$checkAnalysis$1$$anonfun$apply$2.applyOrElse(CheckAnalysis. scala:52) at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformUp$1.apply(TreeNode.scala:286) at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformUp$1.apply(TreeNode.scala:286) > Special chars in column names is broken > --------------------------------------- > > Key: SPARK-6898 > URL: https://issues.apache.org/jira/browse/SPARK-6898 > Project: Spark > Issue Type: Bug > Components: SQL > Reporter: Wenchen Fan > Assignee: Wenchen Fan > Fix For: 1.4.0 > > > This function is added a long time ago, but it's not complete, it will fail > if we have "." inside column name. > {code} > test("SPARK-3483 Special chars in column names") { > val data = sparkContext.parallelize( > Seq("""{"key?number1": "value1", "key.number2": "value2"}""")) > jsonRDD(data).registerTempTable("records") > sql("SELECT `key?number1`, `key.number2` FROM records") > } > {code} > this test will fail. -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org