Github user gatorsmile commented on a diff in the pull request:

    https://github.com/apache/spark/pull/19861#discussion_r154503549
  
    --- Diff: 
sql/core/src/main/scala/org/apache/spark/sql/DataFrameReader.scala ---
    @@ -732,3 +743,25 @@ class DataFrameReader private[sql](sparkSession: 
SparkSession) extends Logging {
       private val extraOptions = new scala.collection.mutable.HashMap[String, 
String]
     
     }
    +
    +private[sql] object DataFrameReader {
    +
    +  /**
    +   * Helper method to filter session configs with config key that matches 
at least one of the given
    +   * prefixes.
    +   *
    +   * @param cs the config key-prefixes that should be filtered.
    +   * @param conf the session conf
    +   * @return an immutable map that contains all the session configs that 
should be propagated to
    +   *         the data source.
    +   */
    +  def withSessionConfig(
    --- End diff --
    
    These helper functions need to be moved to 
org.apache.spark.sql.execution.datasources.v2 package. This will be called by 
SQL API code path. 
    
    Another more straightforward option is to provide it by `ConfigSupport`. 
WDYT? cc @cloud-fan  


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to