[GitHub] spark pull request #22782: [HOTFIX] Fix PySpark pip packaging tests by non-a...

2018-10-23 Thread rvesse
Github user rvesse commented on a diff in the pull request:

https://github.com/apache/spark/pull/22782#discussion_r227373891
  
--- Diff: bin/docker-image-tool.sh ---
@@ -79,7 +79,7 @@ function build {
   fi
 
   # Verify that Spark has actually been built/is a runnable distribution
-  # i.e. the Spark JARs that the Docker files will place into the image 
are present
--- End diff --

It wasn't as if I used a non-breaking space intentionally, I just used my 
OSes default editor for Shell scripts!


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22782: [HOTFIX] Fix PySpark pip packaging tests by non-a...

2018-10-23 Thread HyukjinKwon
Github user HyukjinKwon commented on a diff in the pull request:

https://github.com/apache/spark/pull/22782#discussion_r227336895
  
--- Diff: bin/docker-image-tool.sh ---
@@ -79,7 +79,7 @@ function build {
   fi
 
   # Verify that Spark has actually been built/is a runnable distribution
-  # i.e. the Spark JARs that the Docker files will place into the image 
are present
--- End diff --

Therefore, using non-breakable spaces in the codes is obviously not a good 
practice. Please avoid next time.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22782: [HOTFIX] Fix PySpark pip packaging tests by non-a...

2018-10-23 Thread HyukjinKwon
Github user HyukjinKwon commented on a diff in the pull request:

https://github.com/apache/spark/pull/22782#discussion_r227336318
  
--- Diff: bin/docker-image-tool.sh ---
@@ -79,7 +79,7 @@ function build {
   fi
 
   # Verify that Spark has actually been built/is a runnable distribution
-  # i.e. the Spark JARs that the Docker files will place into the image 
are present
--- End diff --

For the issue itself, It's related to a historical reason for Python. 
Python 2 supported `str` type as bytes like string. It looked a mistake that 
confuses users about the concept between bytes and string, and then Python 3 
introduced `str` as unicode strings concepts like other programing languages.

`open(...).read()` reads it as `str` (which is bytes) in Python 2 but it's 
read in unicode strings in Python 3 - where we need an implicit conversion 
between bytes and strings. Looks it had to be to minimise the breaking changes 
in users codes.

So, bytes to string conversion happened here and unfortunately our 
Jenkins's system default encoding is set to ascii (even though arguably UTF-8 
is common).


For non-ascii itself, please see the justification at 
http://www.scalastyle.org/rules-dev.html in ScalaStyle.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22782: [HOTFIX] Fix PySpark pip packaging tests by non-a...

2018-10-23 Thread rvesse
Github user rvesse commented on a diff in the pull request:

https://github.com/apache/spark/pull/22782#discussion_r227274978
  
--- Diff: bin/docker-image-tool.sh ---
@@ -79,7 +79,7 @@ function build {
   fi
 
   # Verify that Spark has actually been built/is a runnable distribution
-  # i.e. the Spark JARs that the Docker files will place into the image 
are present
--- End diff --

Obvious question but why are we still using ASCII encoding for anything?


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22782: [HOTFIX] Fix PySpark pip packaging tests by non-a...

2018-10-22 Thread HyukjinKwon
Github user HyukjinKwon commented on a diff in the pull request:

https://github.com/apache/spark/pull/22782#discussion_r227173103
  
--- Diff: bin/docker-image-tool.sh ---
@@ -79,7 +79,7 @@ function build {
   fi
 
   # Verify that Spark has actually been built/is a runnable distribution
-  # i.e. the Spark JARs that the Docker files will place into the image 
are present
--- End diff --

It was more because non-breakable space was used instead instead of space. 
It was in utf-8 but non ascii compatible.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22782: [HOTFIX] Fix PySpark pip packaging tests by non-a...

2018-10-22 Thread rvesse
Github user rvesse commented on a diff in the pull request:

https://github.com/apache/spark/pull/22782#discussion_r226983785
  
--- Diff: bin/docker-image-tool.sh ---
@@ -79,7 +79,7 @@ function build {
   fi
 
   # Verify that Spark has actually been built/is a runnable distribution
-  # i.e. the Spark JARs that the Docker files will place into the image 
are present
--- End diff --

@dongjoon-hyun This was edited in Xcode on OS X so almost certainly 
defaulting to UTF-8 encoding


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22782: [HOTFIX] Fix PySpark pip packaging tests by non-a...

2018-10-20 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/spark/pull/22782


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22782: [HOTFIX] Fix PySpark pip packaging tests by non-a...

2018-10-20 Thread dongjoon-hyun
Github user dongjoon-hyun commented on a diff in the pull request:

https://github.com/apache/spark/pull/22782#discussion_r226834220
  
--- Diff: bin/docker-image-tool.sh ---
@@ -79,7 +79,7 @@ function build {
   fi
 
   # Verify that Spark has actually been built/is a runnable distribution
-  # i.e. the Spark JARs that the Docker files will place into the image 
are present
--- End diff --

Right. We need some automatic check for scripts too. Anyway, thank you for 
fixing this, @HyukjinKwon ! This blocks almost everything PRs.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22782: [HOTFIX] Fix PySpark pip packaging tests by non-a...

2018-10-20 Thread HyukjinKwon
Github user HyukjinKwon commented on a diff in the pull request:

https://github.com/apache/spark/pull/22782#discussion_r226833951
  
--- Diff: bin/docker-image-tool.sh ---
@@ -79,7 +79,7 @@ function build {
   fi
 
   # Verify that Spark has actually been built/is a runnable distribution
-  # i.e. the Spark JARs that the Docker files will place into the image 
are present
--- End diff --

Yea, some text editors insert non-breakable spaces and probably that's why. 
I dealt with similar problems at md files before. I think we have scalastyle 
for nonascii but looks not for scripts ..


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22782: [HOTFIX] Fix PySpark pip packaging tests by non-a...

2018-10-20 Thread dongjoon-hyun
Github user dongjoon-hyun commented on a diff in the pull request:

https://github.com/apache/spark/pull/22782#discussion_r226833751
  
--- Diff: bin/docker-image-tool.sh ---
@@ -79,7 +79,7 @@ function build {
   fi
 
   # Verify that Spark has actually been built/is a runnable distribution
-  # i.e. the Spark JARs that the Docker files will place into the image 
are present
--- End diff --

Hi, @rvesse .
Do you have any clue about the special characters after '#' found by @kiszk 
at [here](https://github.com/apache/spark/pull/22782#issuecomment-431599457)?
I'm wondering if we can avoid this situation in the future. Do you have any 
idea?


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22782: [HOTFIX] Fix PySpark pip packaging tests by non-a...

2018-10-20 Thread HyukjinKwon
Github user HyukjinKwon commented on a diff in the pull request:

https://github.com/apache/spark/pull/22782#discussion_r226833113
  
--- Diff: python/pyspark/__init__.py ---
@@ -16,7 +16,7 @@
 #
 
 """
-PySpark is the Python API for Spark.
+PySpark is the Python API for Spark
--- End diff --

yup. I will revert this one too. just intended to test Python tests only 
since Scala tests takes long.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22782: [HOTFIX] Fix PySpark pip packaging tests by non-a...

2018-10-20 Thread kiszk
Github user kiszk commented on a diff in the pull request:

https://github.com/apache/spark/pull/22782#discussion_r226833103
  
--- Diff: dev/run-tests.py ---
@@ -551,7 +551,8 @@ def main():
 if not changed_files or any(f.endswith(".scala")
 or f.endswith("scalastyle-config.xml")
 for f in changed_files):
-run_scala_style_checks()
+# run_scala_style_checks()
--- End diff --

Got it


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22782: [HOTFIX] Fix PySpark pip packaging tests by non-a...

2018-10-20 Thread HyukjinKwon
Github user HyukjinKwon commented on a diff in the pull request:

https://github.com/apache/spark/pull/22782#discussion_r226833051
  
--- Diff: dev/run-tests.py ---
@@ -551,7 +551,8 @@ def main():
 if not changed_files or any(f.endswith(".scala")
 or f.endswith("scalastyle-config.xml")
 for f in changed_files):
-run_scala_style_checks()
+# run_scala_style_checks()
--- End diff --

tentative workaround. yup yup. I will revert all back once the tests pass


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22782: [HOTFIX] Fix PySpark pip packaging tests by non-a...

2018-10-20 Thread kiszk
Github user kiszk commented on a diff in the pull request:

https://github.com/apache/spark/pull/22782#discussion_r226833042
  
--- Diff: python/pyspark/__init__.py ---
@@ -16,7 +16,7 @@
 #
 
 """
-PySpark is the Python API for Spark.
+PySpark is the Python API for Spark
--- End diff --

Is this intentional change?


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22782: [HOTFIX] Fix PySpark pip packaging tests by non-a...

2018-10-20 Thread kiszk
Github user kiszk commented on a diff in the pull request:

https://github.com/apache/spark/pull/22782#discussion_r226833028
  
--- Diff: dev/run-tests.py ---
@@ -551,7 +551,8 @@ def main():
 if not changed_files or any(f.endswith(".scala")
 or f.endswith("scalastyle-config.xml")
 for f in changed_files):
-run_scala_style_checks()
+# run_scala_style_checks()
--- End diff --

Is this change necessary? Or just tentative workaround?


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org