Github user HyukjinKwon commented on a diff in the pull request:

    https://github.com/apache/spark/pull/21878#discussion_r205350125
  
    --- Diff: 
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala
 ---
    @@ -635,12 +637,6 @@ object DataSource extends Logging {
                         "Hive built-in ORC data source must be used with Hive 
support enabled. " +
                         "Please use the native ORC data source by setting 
'spark.sql.orc.impl' to " +
                         "'native'")
    -                } else if (provider1.toLowerCase(Locale.ROOT) == "avro" ||
    -                  provider1 == "com.databricks.spark.avro") {
    -                  throw new AnalysisException(
    --- End diff --
    
    Eh, if users were using the external avro, they will likely meet the error 
if they directly upgrade Spark.
    Otherwise, users will see the release note that Avro pacakge is included in 
2.4.0, and they will not provide this jar.
    If users miss this release note, then they will try to explicitly provide 
the thirdparty jar which will give the error message above.
    
    FWIW, if it's fully qualified path, the thridparty jar will still be used 
in theory.
    
    Did I misunderstand or miss something maybe?


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to