[GitHub] spark issue #20716: [SPARK-23566][Minor][Doc] Argument name mismatch fixed

2018-03-05 Thread animenon
Github user animenon commented on the issue:

https://github.com/apache/spark/pull/20716
  
@HyukjinKwon Its minor, so may not be required. Had tagged Gator just for a 
check. 


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20716: [SPARK-23566][Minor][Doc] Argument name mismatch fixed

2018-03-05 Thread animenon
Github user animenon commented on the issue:

https://github.com/apache/spark/pull/20716
  
Do check @gatorsmile 


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20716: [SPARK-23566][Minor][Doc] Argument name mismatch ...

2018-03-03 Thread animenon
Github user animenon commented on a diff in the pull request:

https://github.com/apache/spark/pull/20716#discussion_r172024390
  
--- Diff: python/pyspark/sql/dataframe.py ---
@@ -612,9 +614,10 @@ def repartition(self, numPartitions, *cols):
 Returns a new :class:`DataFrame` partitioned by the given 
partitioning expressions. The
 resulting DataFrame is hash partitioned.
 
-``numPartitions`` can be an int to specify the target number of 
partitions or a Column.
-If it is a Column, it will be used as the first partitioning 
column. If not specified,
-the default number of partitions is used.
+:param numPartitions:
+can be an int to specify the target number of partitions or a 
Column.
--- End diff --

Cool, was maintaining consistency with other `:param` on page.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20716: [SPARK-23566][Minor][Doc] Argument name mismatch fixed

2018-03-03 Thread animenon
Github user animenon commented on the issue:

https://github.com/apache/spark/pull/20716
  
@HyukjinKwon Found couple more, have updated the same.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20716: [SPARK-23566][Minor][Doc] Argument name mismatch fixed

2018-03-03 Thread animenon
Github user animenon commented on the issue:

https://github.com/apache/spark/pull/20716
  
@HyukjinKwon Sure, shall do.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20716: [SPARK-23566][Minor][Doc] Argument name mismatch ...

2018-03-02 Thread animenon
GitHub user animenon opened a pull request:

https://github.com/apache/spark/pull/20716

[SPARK-23566][Minor][Doc] Argument name mismatch fixed

Argument name mismatch fixed.

## What changes were proposed in this pull request?

`col` changed to `new` in doc string to match the argument list.

Patch file added: https://issues.apache.org/jira/browse/SPARK-23566

Please review http://spark.apache.org/contributing.html before opening a 
pull request.


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/animenon/spark master

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/20716.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #20716


commit 5d2bb9d0b480d7ac2a2c2d6364b87789f57856e8
Author: Anirudh <animenon@...>
Date:   2018-03-02T13:58:27Z

[SPARK-23566][Minor][Doc] Argument name mismatch fixed

Argument name mismatch fixed.
`col` changed to `new` in doc string to match the argument list.




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #19632: Added description to python spark Pi example

2017-11-02 Thread animenon
Github user animenon commented on a diff in the pull request:

https://github.com/apache/spark/pull/19632#discussion_r148628286
  
--- Diff: examples/src/main/python/pi.py ---
@@ -27,12 +27,16 @@
 if __name__ == "__main__":
 """
 Usage: pi [partitions]
+
+Monte Carlo method is used to estimate Pi in the below example.
 """
 spark = SparkSession\
 .builder\
 .appName("PythonPi")\
 .getOrCreate()
-
+
+# If no arguments are passed(i.e. `len(sys.argv) < = 1` ) 
--- End diff --

@srowen Added a line to the 
[index.md](https://github.com/apache/spark/pull/19632/commits/173fd29f20eab1ca8384413548deb5aac509d629)
 as well, I think it would help the curious ones know what algorithm is used 
and why an argument(i.e. 10) is passed.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #19632: Added description to python spark Pi example

2017-11-02 Thread animenon
Github user animenon commented on a diff in the pull request:

https://github.com/apache/spark/pull/19632#discussion_r148620630
  
--- Diff: examples/src/main/python/pi.py ---
@@ -27,12 +27,16 @@
 if __name__ == "__main__":
 """
 Usage: pi [partitions]
+
+Monte Carlo method is used to estimate Pi in the below example.
 """
 spark = SparkSession\
 .builder\
 .appName("PythonPi")\
 .getOrCreate()
-
+
+# If no arguments are passed(i.e. `len(sys.argv) < = 1` ) 
--- End diff --

This is actually the first example on the spark doc and I wanted to know 
how the `pi` calculation was done. There was no mention of what algorithm is 
used for it, so took me a while to figure out the Monte-Carlo estimator was 
used and the logic is randomly generating over 10 points to finally 
estimate the Pi value.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #19632: Added description

2017-11-01 Thread animenon
GitHub user animenon opened a pull request:

https://github.com/apache/spark/pull/19632

Added description 

## What changes were proposed in this pull request?
Description added to better understand example.

## How was this patch tested?

Not required as only comments are required.

Please review http://spark.apache.org/contributing.html before opening a 
pull request.


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/animenon/spark patch-2

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/19632.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #19632


commit 74e93252763a281daeab3e859180f536beb9de66
Author: Anirudh <anime...@mail.com>
Date:   2017-11-01T19:42:33Z

Added documentation




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19315: [MINOR][ML]Updated english.txt word ordering

2017-09-22 Thread animenon
Github user animenon commented on the issue:

https://github.com/apache/spark/pull/19315
  
Right, but if you want to check a few things together like "who", "who's", 
"why", "when", etc you will have to do many finds; but in an ordered document 
it would be much easier.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #19315: Updated english.txt word ordering

2017-09-21 Thread animenon
GitHub user animenon opened a pull request:

https://github.com/apache/spark/pull/19315

Updated english.txt word ordering

Ordered alphabetically, for better readability.

## What changes were proposed in this pull request?

Alphabetical ordering of the stop words.



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/animenon/spark patch-1

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/19315.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #19315


commit 57c282721c63487a82bdd6959c6ff5f6ce9f66ad
Author: Anirudh <anime...@mail.com>
Date:   2017-09-22T02:40:30Z

Updated english.txt word ordering

Ordered alphabetically, for better readability.




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org