Github user rxin commented on a diff in the pull request:

    https://github.com/apache/spark/pull/15540#discussion_r83994842
  
    --- Diff: docs/sql-programming-guide.md ---
    @@ -421,9 +421,10 @@ In the simplest form, the default data source 
(`parquet` unless otherwise config
     
     You can also manually specify the data source that will be used along with 
any extra options
     that you would like to pass to the data source. Data sources are specified 
by their fully qualified
    -name (i.e., `org.apache.spark.sql.parquet`), but for built-in sources you 
can also use their short
    -names (`json`, `parquet`, `jdbc`, `orc`, `libsvm`, `csv`, `text`). 
DataFrames loaded from any data
    -source type can be converted into other types using this syntax.
    +name (i.e., `org.apache.spark.sql.parquet`) or short name specified in
    +`DataSourceRegister.shortName()` in their implementation. For built-in 
sources you can use their
    --- End diff --
    
    hm this isn't clear to the end user at all, since there is no class called 
DataSourceRegister even in the source code for data sources. I'd just leave it 
out. It should be the job of the data source to tell users how to use it.



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to