Github user gengliangwang commented on a diff in the pull request: https://github.com/apache/spark/pull/21878#discussion_r205353836 --- Diff: sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala --- @@ -635,12 +637,6 @@ object DataSource extends Logging { "Hive built-in ORC data source must be used with Hive support enabled. " + "Please use the native ORC data source by setting 'spark.sql.orc.impl' to " + "'native'") - } else if (provider1.toLowerCase(Locale.ROOT) == "avro" || - provider1 == "com.databricks.spark.avro") { - throw new AnalysisException( --- End diff -- I totally agree with the mapping, we should do it. The comment here is about when Spark can't find any avro package, we should show a message for loading the spark-avro jar(org.apache.spark.sql.avro). Different from CSV, the package spark-avro is not loaded by default within Spark(at least as I tried spark-shell).
--- --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org