Github user HyukjinKwon commented on a diff in the pull request:

    https://github.com/apache/spark/pull/20937#discussion_r178426879
  
    --- Diff: 
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/json/JSONOptions.scala
 ---
    @@ -86,14 +86,30 @@ private[sql] class JSONOptions(
     
       val multiLine = 
parameters.get("multiLine").map(_.toBoolean).getOrElse(false)
     
    -  val lineSeparator: Option[String] = parameters.get("lineSep").map { sep 
=>
    -    require(sep.nonEmpty, "'lineSep' cannot be an empty string.")
    -    sep
    +  /**
    +   * A sequence of bytes between two consecutive json records.
    +   */
    +  val lineSeparator: Option[String] = parameters.get("lineSep")
    +
    +  /**
    +   * Standard charset name. For example UTF-8, UTF-16 and UTF-32.
    +   * If charset is not specified (None), it will be detected automatically.
    +   */
    +  val charset: Option[String] = parameters.get("charset")
    --- End diff --
    
    I thought we talked `encoding` would have higher priority like CSV.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to