Liang-Chi Hsieh created SPARK-20399:
---------------------------------------

             Summary: Can't use same regex pattern between 1.6 and 2.x due to 
unescaped sql string in parser
                 Key: SPARK-20399
                 URL: https://issues.apache.org/jira/browse/SPARK-20399
             Project: Spark
          Issue Type: Bug
          Components: SQL
    Affects Versions: 2.2.0
            Reporter: Liang-Chi Hsieh


The new SQL parser is introduced into Spark 2.0. Seems it bring an issue 
regarding the regex pattern string.

The following codes can reproduce it:
{code}
val data = Seq("\u0020\u0021\u0023", "abc")
val df = data.toDF()

// 1st usage: let parser parse pattern string: works in 1.6
val rlike1 = df.filter("value rlike '^\\x20[\\x20-\\x23]+$'")
// 2nd usage: call Column.rlike so the pattern string is a literal which 
doesn't go through parser
val rlike2 = df.filter($"value".rlike("^\\x20[\\x20-\\x23]+$"))  // 2: works in 
1.6, 2.x

// To make 1st usage work, we need to add backslashes like this in 2.x:
val rlike3 = df.filter("value rlike '^\\\\x20[\\\\x20-\\\\x23]+$'")
{code}

Due to unescape SQL String in parser, the first usage working in 1.6 can't work 
in 2.0. To make it work, we need to add additional backslashes.

It is quite weird that we can't use the same regex pattern string the 2 usages.




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to