[ https://issues.apache.org/jira/browse/SPARK-15973?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15335114#comment-15335114 ]
Apache Spark commented on SPARK-15973: -------------------------------------- User 'josh-howes' has created a pull request for this issue: https://github.com/apache/spark/pull/13724 > Fix GroupedData Documentation > ----------------------------- > > Key: SPARK-15973 > URL: https://issues.apache.org/jira/browse/SPARK-15973 > Project: Spark > Issue Type: Improvement > Components: PySpark > Affects Versions: 2.0.0, 2.1.0 > Reporter: Vladimir Feinberg > Priority: Trivial > Original Estimate: 1h > Remaining Estimate: 1h > > (1) > {{GroupedData.pivot}} documenation uses {{//}} instead of {{#}} for doctest > python comments, which messes up formatting in the documentation as well as > the doctests themselves. > A PR resolving this should probably resolve the other places this happens in > pyspark. > (2) > Simple aggregation functions which take column names {{cols}} as varargs > arguments show up in documentation with the argument {{args}}, but their > documentation refers to {{cols}}. > The discrepancy is caused by an annotation, {{df_varargs_api}}, which > produces a temporary function with arguments {{args}} instead of {{cols}}, > creating the confusing documentation. > (3) > The {{pyspark.sql.GroupedData}} object calls the Java object it wraps around > as the member variable {{self._jdf}}, which is exactly the same as > {{pyspark.sql.DataFrame}}, when referring its object. > The acronym is incorrect, standing for "Java DataFrame" instead of what > should be "Java GroupedData". As such, the name should be changed to > {{self._jgd}} - in fact, in the {{DataFrame.groupBy}} implementation, the > java object is referred to as exactly {{jgd}} -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org