Github user sureshthalamati commented on a diff in the pull request:

    https://github.com/apache/spark/pull/12904#discussion_r84373258
  
    --- Diff: 
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/csv/CSVOptions.scala
 ---
    @@ -90,6 +90,7 @@ private[csv] class CSVOptions(@transient private val 
parameters: Map[String, Str
       val permissive = ParseModes.isPermissiveMode(parseMode)
     
       val nullValue = parameters.getOrElse("nullValue", "")
    +  val emptyValue = parameters.getOrElse("emptyValue", "")
    --- End diff --
    
    Yes, null and empty can not be differentiated when they are set to same 
value.  Currently null value check has higher precedence than the empty value. 
    
    input.csv
    1,
    2,”” 
    
    Output will be: 
    1, null
    2, null
    
    
    I think this behavior  is ok.  By default  Univocity CSV parser  used in 
spark  also returns  null for empty strings.
    
    I agree we should document this behavior. 


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to