[ 
https://issues.apache.org/jira/browse/SPARK-6898?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14640704#comment-14640704
 ] 

David Sabater commented on SPARK-6898:
--------------------------------------

I hope I am missing something here but I am facing this issue in my current 
cluster (Currently running compiled version of Spark 1.4.0 with Hive support)
(I tried with 1.4.0 and 1.4.1 binaries locally and had same issue).
sqlContext.jsonRDD(sc.makeRDD("""{"a": {"c.b": 1}, "b.$q": [{"a@!.q": 1}], 
"q.w": {"w.i&": [1]}}""" :: Nil)).registerTempTable("t")
sqlContext.sql("SELECT `key?number1`, `key.number2` FROM records")
sqlContext.sql("SELECT a.`c.b`, `b.$q`[0].`a@!.q`, `q.w`.`w.i&`[0] FROM t")

Either select is throwing the same error:
scala> sqlContext.sql("SELECT a.`c.b`, `b.$q`[0].`a@!.q`, `q.w`.`w.i&`[0] FROM 
t")
15/07/24 17:23:12 INFO ParseDriver: Parsing command: SELECT a.`c.b`, 
`b.$q`[0].`a@!.q`, `q.w`.`w.i&`[0] FROM t
15/07/24 17:23:12 INFO ParseDriver: Parse Completed
org.apache.spark.sql.AnalysisException: cannot resolve 'b.$q' given input 
columns a, b.$q, q.w; line 1 pos 16
        at 
org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.failAnalysis(package.scala:42)
        at 
org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$checkAnalysis$1$$anonfun$apply$2.applyOrElse(CheckAnalysis.
scala:63)
        at 
org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$checkAnalysis$1$$anonfun$apply$2.applyOrElse(CheckAnalysis.
scala:52)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformUp$1.apply(TreeNode.scala:286)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformUp$1.apply(TreeNode.scala:286)


> Special chars in column names is broken
> ---------------------------------------
>
>                 Key: SPARK-6898
>                 URL: https://issues.apache.org/jira/browse/SPARK-6898
>             Project: Spark
>          Issue Type: Bug
>          Components: SQL
>            Reporter: Wenchen Fan
>            Assignee: Wenchen Fan
>             Fix For: 1.4.0
>
>
> This function is added a long time ago, but it's not complete, it will fail 
> if we have "." inside column name.
> {code}
> test("SPARK-3483 Special chars in column names") {
>     val data = sparkContext.parallelize(
>       Seq("""{"key?number1": "value1", "key.number2": "value2"}"""))
>     jsonRDD(data).registerTempTable("records")
>     sql("SELECT `key?number1`, `key.number2` FROM records")
>   }
> {code}
> this test will fail.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to