Github user HyukjinKwon commented on a diff in the pull request: https://github.com/apache/spark/pull/21878#discussion_r205350125 --- Diff: sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala --- @@ -635,12 +637,6 @@ object DataSource extends Logging { "Hive built-in ORC data source must be used with Hive support enabled. " + "Please use the native ORC data source by setting 'spark.sql.orc.impl' to " + "'native'") - } else if (provider1.toLowerCase(Locale.ROOT) == "avro" || - provider1 == "com.databricks.spark.avro") { - throw new AnalysisException( --- End diff -- Eh, if users were using the external avro, they will likely meet the error if they directly upgrade Spark. Otherwise, users will see the release note that Avro pacakge is included in 2.4.0, and they will not provide this jar. If users miss this release note, then they will try to explicitly provide the thirdparty jar which will give the error message above. FWIW, if it's fully qualified path, the thridparty jar will still be used in theory. Did I misunderstand or miss something maybe?
--- --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org