Github user goldmedal commented on a diff in the pull request:

    https://github.com/apache/spark/pull/19339#discussion_r140779203
  
    --- Diff: 
sql/core/src/main/scala/org/apache/spark/sql/DataFrameReader.scala ---
    @@ -456,6 +456,40 @@ class DataFrameReader private[sql](sparkSession: 
SparkSession) extends Logging {
       }
     
       /**
    +   * Loads a `JavaRDD[String]` storing storing CSV rows and returns the 
result as a `DataFrame`.
    +   *
    +   * If the schema is not specified using `schema` function and 
`inferSchema` option is enabled,
    +   * this function goes through the input once to determine the input 
schema.
    +   *
    +   * If the schema is not specified using `schema` function and 
`inferSchema` option is disabled,
    +   * it determines the columns as string types and it reads only the first 
line to determine the
    +   * names and the number of fields.
    +   *
    +   * @param csvRDD input RDD with one CSV row per record
    +   * @since 2.2.0
    +   */
    +  @deprecated("Use csv(Dataset[String]) instead.", "2.2.0")
    +  def csv(csvRDD: JavaRDD[String]): DataFrame = csv(csvRDD.rdd)
    +
    +  /**
    +   * Loads a `RDD[String]` storing storing CSV rows and returns the result 
as a `DataFrame`.
    +   *
    +   * If the schema is not specified using `schema` function and 
`inferSchema` option is enabled,
    +   * this function goes through the input once to determine the input 
schema.
    +   *
    +   * If the schema is not specified using `schema` function and 
`inferSchema` option is disabled,
    +   * it determines the columns as string types and it reads only the first 
line to determine the
    +   * names and the number of fields.
    +   *
    +   * @param csvRDD input RDD with one CSV row per record
    +   * @since 2.2.0
    +   */
    +  @deprecated("Use csv(Dataset[String]) instead.", "2.2.0")
    +  def csv(csvRDD: RDD[String]): DataFrame = {
    --- End diff --
    
    Thanks for your reviewing :)
    umm..  I followed `spark.read.json`'s way to add them. Although 
`json(jsonRDD :RDD[String]` has been deprecated, PySpark still use it to create 
a `DataFrame`. I think adding a private wrapper in Scala maybe better because 
not only PySpark but SparkR maybe need it.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to