[GitHub] spark issue #20716: [SPARK-23566][Minor][Doc] Argument name mismatch fixed
Github user animenon commented on the issue: https://github.com/apache/spark/pull/20716 @HyukjinKwon Its minor, so may not be required. Had tagged Gator just for a check. --- - To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org
[GitHub] spark issue #20716: [SPARK-23566][Minor][Doc] Argument name mismatch fixed
Github user animenon commented on the issue: https://github.com/apache/spark/pull/20716 Do check @gatorsmile --- - To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org
[GitHub] spark pull request #20716: [SPARK-23566][Minor][Doc] Argument name mismatch ...
Github user animenon commented on a diff in the pull request: https://github.com/apache/spark/pull/20716#discussion_r172024390 --- Diff: python/pyspark/sql/dataframe.py --- @@ -612,9 +614,10 @@ def repartition(self, numPartitions, *cols): Returns a new :class:`DataFrame` partitioned by the given partitioning expressions. The resulting DataFrame is hash partitioned. -``numPartitions`` can be an int to specify the target number of partitions or a Column. -If it is a Column, it will be used as the first partitioning column. If not specified, -the default number of partitions is used. +:param numPartitions: +can be an int to specify the target number of partitions or a Column. --- End diff -- Cool, was maintaining consistency with other `:param` on page. --- - To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org
[GitHub] spark issue #20716: [SPARK-23566][Minor][Doc] Argument name mismatch fixed
Github user animenon commented on the issue: https://github.com/apache/spark/pull/20716 @HyukjinKwon Found couple more, have updated the same. --- - To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org
[GitHub] spark issue #20716: [SPARK-23566][Minor][Doc] Argument name mismatch fixed
Github user animenon commented on the issue: https://github.com/apache/spark/pull/20716 @HyukjinKwon Sure, shall do. --- - To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org
[GitHub] spark pull request #20716: [SPARK-23566][Minor][Doc] Argument name mismatch ...
GitHub user animenon opened a pull request: https://github.com/apache/spark/pull/20716 [SPARK-23566][Minor][Doc] Argument name mismatch fixed Argument name mismatch fixed. ## What changes were proposed in this pull request? `col` changed to `new` in doc string to match the argument list. Patch file added: https://issues.apache.org/jira/browse/SPARK-23566 Please review http://spark.apache.org/contributing.html before opening a pull request. You can merge this pull request into a Git repository by running: $ git pull https://github.com/animenon/spark master Alternatively you can review and apply these changes as the patch at: https://github.com/apache/spark/pull/20716.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #20716 commit 5d2bb9d0b480d7ac2a2c2d6364b87789f57856e8 Author: Anirudh <animenon@...> Date: 2018-03-02T13:58:27Z [SPARK-23566][Minor][Doc] Argument name mismatch fixed Argument name mismatch fixed. `col` changed to `new` in doc string to match the argument list. --- - To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org
[GitHub] spark pull request #19632: Added description to python spark Pi example
Github user animenon commented on a diff in the pull request: https://github.com/apache/spark/pull/19632#discussion_r148628286 --- Diff: examples/src/main/python/pi.py --- @@ -27,12 +27,16 @@ if __name__ == "__main__": """ Usage: pi [partitions] + +Monte Carlo method is used to estimate Pi in the below example. """ spark = SparkSession\ .builder\ .appName("PythonPi")\ .getOrCreate() - + +# If no arguments are passed(i.e. `len(sys.argv) < = 1` ) --- End diff -- @srowen Added a line to the [index.md](https://github.com/apache/spark/pull/19632/commits/173fd29f20eab1ca8384413548deb5aac509d629) as well, I think it would help the curious ones know what algorithm is used and why an argument(i.e. 10) is passed. --- - To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org
[GitHub] spark pull request #19632: Added description to python spark Pi example
Github user animenon commented on a diff in the pull request: https://github.com/apache/spark/pull/19632#discussion_r148620630 --- Diff: examples/src/main/python/pi.py --- @@ -27,12 +27,16 @@ if __name__ == "__main__": """ Usage: pi [partitions] + +Monte Carlo method is used to estimate Pi in the below example. """ spark = SparkSession\ .builder\ .appName("PythonPi")\ .getOrCreate() - + +# If no arguments are passed(i.e. `len(sys.argv) < = 1` ) --- End diff -- This is actually the first example on the spark doc and I wanted to know how the `pi` calculation was done. There was no mention of what algorithm is used for it, so took me a while to figure out the Monte-Carlo estimator was used and the logic is randomly generating over 10 points to finally estimate the Pi value. --- - To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org
[GitHub] spark pull request #19632: Added description
GitHub user animenon opened a pull request: https://github.com/apache/spark/pull/19632 Added description ## What changes were proposed in this pull request? Description added to better understand example. ## How was this patch tested? Not required as only comments are required. Please review http://spark.apache.org/contributing.html before opening a pull request. You can merge this pull request into a Git repository by running: $ git pull https://github.com/animenon/spark patch-2 Alternatively you can review and apply these changes as the patch at: https://github.com/apache/spark/pull/19632.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #19632 commit 74e93252763a281daeab3e859180f536beb9de66 Author: Anirudh <anime...@mail.com> Date: 2017-11-01T19:42:33Z Added documentation --- - To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org
[GitHub] spark issue #19315: [MINOR][ML]Updated english.txt word ordering
Github user animenon commented on the issue: https://github.com/apache/spark/pull/19315 Right, but if you want to check a few things together like "who", "who's", "why", "when", etc you will have to do many finds; but in an ordered document it would be much easier. --- - To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org
[GitHub] spark pull request #19315: Updated english.txt word ordering
GitHub user animenon opened a pull request: https://github.com/apache/spark/pull/19315 Updated english.txt word ordering Ordered alphabetically, for better readability. ## What changes were proposed in this pull request? Alphabetical ordering of the stop words. You can merge this pull request into a Git repository by running: $ git pull https://github.com/animenon/spark patch-1 Alternatively you can review and apply these changes as the patch at: https://github.com/apache/spark/pull/19315.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #19315 commit 57c282721c63487a82bdd6959c6ff5f6ce9f66ad Author: Anirudh <anime...@mail.com> Date: 2017-09-22T02:40:30Z Updated english.txt word ordering Ordered alphabetically, for better readability. --- - To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org