[jira] [Commented] (SPARK-18136) Make PySpark pip install works on windows

2017-11-10 Thread Felix Cheung (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-18136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16248377#comment-16248377
 ] 

Felix Cheung commented on SPARK-18136:
--

thanks [~hyukjin.kwon]

> Make PySpark pip install works on windows
> -
>
> Key: SPARK-18136
> URL: https://issues.apache.org/jira/browse/SPARK-18136
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Affects Versions: 2.1.2, 2.2.0, 2.3.0
>Reporter: holdenk
> Fix For: 2.1.3, 2.2.1, 2.3.0
>
>
> Make sure that pip installer for PySpark works on windows



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-18136) Make PySpark pip install works on windows

2017-11-10 Thread Hyukjin Kwon (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-18136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16248368#comment-16248368
 ] 

Hyukjin Kwon commented on SPARK-18136:
--

Fixed in https://github.com/apache/spark/pull/19310

> Make PySpark pip install works on windows
> -
>
> Key: SPARK-18136
> URL: https://issues.apache.org/jira/browse/SPARK-18136
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Reporter: holdenk
> Fix For: 2.1.3, 2.2.1, 2.3.0
>
>
> Make sure that pip installer for PySpark works on windows



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-18136) Make PySpark pip install works on windows

2017-11-10 Thread Hyukjin Kwon (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-18136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16248367#comment-16248367
 ] 

Hyukjin Kwon commented on SPARK-18136:
--

Let me resolve this and separately open up another JIRA for finding Spark's 
home. pip installation itself works find I believe.

> Make PySpark pip install works on windows
> -
>
> Key: SPARK-18136
> URL: https://issues.apache.org/jira/browse/SPARK-18136
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Reporter: holdenk
> Fix For: 2.1.3, 2.2.1, 2.3.0
>
>
> Make sure that pip installer for PySpark works on windows



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-18136) Make PySpark pip install works on windows

2017-11-10 Thread Felix Cheung (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-18136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16248361#comment-16248361
 ] 

Felix Cheung commented on SPARK-18136:
--

is this still being targeted for 2.2.1?

> Make PySpark pip install works on windows
> -
>
> Key: SPARK-18136
> URL: https://issues.apache.org/jira/browse/SPARK-18136
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Reporter: holdenk
> Fix For: 2.1.3, 2.2.1, 2.3.0
>
>
> Make sure that pip installer for PySpark works on windows



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-18136) Make PySpark pip install works on windows

2017-10-30 Thread Hyukjin Kwon (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-18136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16225081#comment-16225081
 ] 

Hyukjin Kwon commented on SPARK-18136:
--

Oh, no. Yes, there is something to be done to detect Spark home for pip 
installation on Windows. This is partly resolved at the first PR and second 
followup is in review. Will take an action for this soon e.g., taking over the 
second PR if it goes inactive.

> Make PySpark pip install works on windows
> -
>
> Key: SPARK-18136
> URL: https://issues.apache.org/jira/browse/SPARK-18136
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Reporter: holdenk
> Fix For: 2.1.3, 2.2.1, 2.3.0
>
>
> Make sure that pip installer for PySpark works on windows



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-18136) Make PySpark pip install works on windows

2017-10-30 Thread Sean Owen (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-18136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16224845#comment-16224845
 ] 

Sean Owen commented on SPARK-18136:
---

Is this open because it's meant to be backported further (I imagine not)? or 
just that something else needs to be done?
If it's substantially working, let's call this one resolved.

> Make PySpark pip install works on windows
> -
>
> Key: SPARK-18136
> URL: https://issues.apache.org/jira/browse/SPARK-18136
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Reporter: holdenk
> Fix For: 2.1.3, 2.2.1, 2.3.0
>
>
> Make sure that pip installer for PySpark works on windows



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-18136) Make PySpark pip install works on windows

2017-09-27 Thread Jakub Nowacki (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-18136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16183077#comment-16183077
 ] 

Jakub Nowacki commented on SPARK-18136:
---

PR 19370 (https://github.com/apache/spark/pull/19370) fixes {{SPARK_HOME}} 
issue using {{find_spark_home.py}} script. It's not maybe the most elegant way, 
but it is simple.

I think in a long run it would be better to move to some Python packaging 
mechanisms like {{console_scripts}} or related, as it will provide better 
multiplatform support; see 
https://packaging.python.org/tutorials/distributing-packages/#scripts and 
https://setuptools.readthedocs.io/en/latest/setuptools.html#automatic-script-creation.
 I'll create a separate issue with improvement proposal.

> Make PySpark pip install works on windows
> -
>
> Key: SPARK-18136
> URL: https://issues.apache.org/jira/browse/SPARK-18136
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Reporter: holdenk
> Fix For: 2.2.1, 2.3.0, 2.1.3
>
>
> Make sure that pip installer for PySpark works on windows



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-18136) Make PySpark pip install works on windows

2017-09-27 Thread Apache Spark (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-18136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16182868#comment-16182868
 ] 

Apache Spark commented on SPARK-18136:
--

User 'jsnowacki' has created a pull request for this issue:
https://github.com/apache/spark/pull/19370

> Make PySpark pip install works on windows
> -
>
> Key: SPARK-18136
> URL: https://issues.apache.org/jira/browse/SPARK-18136
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Reporter: holdenk
> Fix For: 2.2.1, 2.3.0, 2.1.3
>
>
> Make sure that pip installer for PySpark works on windows



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-18136) Make PySpark pip install works on windows

2017-09-25 Thread Jakub Nowacki (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-18136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16179746#comment-16179746
 ] 

Jakub Nowacki commented on SPARK-18136:
---

I can come back to this issue this Wednesday I think. I did some preliminary 
tests with {{find_spark_home.py}} but I won't have time to sit to it until 
Wednesday.

> Make PySpark pip install works on windows
> -
>
> Key: SPARK-18136
> URL: https://issues.apache.org/jira/browse/SPARK-18136
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Reporter: holdenk
> Fix For: 2.2.1, 2.3.0, 2.1.3
>
>
> Make sure that pip installer for PySpark works on windows



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-18136) Make PySpark pip install works on windows

2017-09-25 Thread holdenk (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-18136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16179172#comment-16179172
 ] 

holdenk commented on SPARK-18136:
-

[~fobofindia09] So currently we're working on a 2.1.2 release, and I believe 
the general consensus is to kick of 2.2.1 RC process after 2.1.2 is wrapped up, 
so it all depends on how quickly we can get 2.1.2 out the door. It's difficult 
to predict on the 2.1.2 side since we're doing it outside of the previous set 
of release managers so as to figure out what unwritten knowledge needs to get 
documented and we're running into a few unexpected issues as we go.

> Make PySpark pip install works on windows
> -
>
> Key: SPARK-18136
> URL: https://issues.apache.org/jira/browse/SPARK-18136
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Reporter: holdenk
> Fix For: 2.2.1, 2.3.0, 2.1.3
>
>
> Make sure that pip installer for PySpark works on windows



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-18136) Make PySpark pip install works on windows

2017-09-25 Thread Neel Shah (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-18136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16179166#comment-16179166
 ] 

Neel Shah commented on SPARK-18136:
---

[~hyukjin.kwon] [~jsnowacki] is this the only thing holding 2.2.1 back? Is 
there an ETA for that release?

> Make PySpark pip install works on windows
> -
>
> Key: SPARK-18136
> URL: https://issues.apache.org/jira/browse/SPARK-18136
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Reporter: holdenk
> Fix For: 2.2.1, 2.3.0, 2.1.3
>
>
> Make sure that pip installer for PySpark works on windows



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-18136) Make PySpark pip install works on windows

2017-09-23 Thread Hyukjin Kwon (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-18136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16177818#comment-16177818
 ] 

Hyukjin Kwon commented on SPARK-18136:
--

I haven't looked into the way you said but let's make the change minimised as 
possible as we can. 

> Make PySpark pip install works on windows
> -
>
> Key: SPARK-18136
> URL: https://issues.apache.org/jira/browse/SPARK-18136
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Reporter: holdenk
> Fix For: 2.1.2, 2.2.1, 2.3.0
>
>
> Make sure that pip installer for PySpark works on windows



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-18136) Make PySpark pip install works on windows

2017-09-23 Thread Jakub Nowacki (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-18136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16177795#comment-16177795
 ] 

Jakub Nowacki commented on SPARK-18136:
---

I've looked into it again and noticed the Bash script {{find_spark_home}}, 
which is used in Bash version of {{pyspark}} command. The Python script 
{{find_spark_home.py}} seems to return the correct SPARK_HOME path on Windows, 
so the all the cmd-files should be alter somehow to use it instead of 
{{%~dp0}}. I'll look into it when I have time, maybe next week, and propose 
something similar to the {{find_spark_home}} script approach.

> Make PySpark pip install works on windows
> -
>
> Key: SPARK-18136
> URL: https://issues.apache.org/jira/browse/SPARK-18136
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Reporter: holdenk
> Fix For: 2.1.2, 2.2.1, 2.3.0
>
>
> Make sure that pip installer for PySpark works on windows



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-18136) Make PySpark pip install works on windows

2017-09-21 Thread Jakub Nowacki (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-18136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16175303#comment-16175303
 ] 

Jakub Nowacki commented on SPARK-18136:
---

I've tried using Windows command {{mklink}} to create symbolic links, but it 
seems to resolve the folder in {{%~dp0}} to the Scripts folder 
{{C:\Tools\Anaconda3\Scripts\}}.

> Make PySpark pip install works on windows
> -
>
> Key: SPARK-18136
> URL: https://issues.apache.org/jira/browse/SPARK-18136
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Reporter: holdenk
>
> Make sure that pip installer for PySpark works on windows



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-18136) Make PySpark pip install works on windows

2017-09-21 Thread Jakub Nowacki (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-18136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16175276#comment-16175276
 ] 

Jakub Nowacki commented on SPARK-18136:
---

[PR|https://github.com/apache/spark/pull/19310] fixes how {{spark-class2.cmd}} 
looks for jars directory on Windows. It fails to find jars and start JVM as the 
condition for the env variable {{SPARK_JARS_DIR}} looks for 
{{%SPARK_HOME%\RELEASE}}, which is not included in the {{pip/conda}} build. 
Instead, it should look for {{%SPARK_HOME%\jars}}, which it is later referring 
to.

The above fixes the errors while importing {{pyspark}} into Python and creating 
SparkSession, but there is still an issue calling {{pyspark.cmd}}. Namely, 
normal command call on command line, without path specification fails with 
{{System cannot find the path specified.}}. It is likely due to the script link 
being resolved to Script folder in Anaconda, e.g. 
{{C:\Tools\Anaconda3\Scripts\pyspark.cmd}}. If the script is run via the full 
path to the PySpark package, e.g. 
{{\Tools\Anaconda3\Lib\site-packages\pyspark\bin\pyspark.cmd}} it works fine. 
It is likely due to the fact that {{SPARK_HOME}} is resolved as follows {{set 
SPARK_HOME=%~dp0..}}, which in case of the system call resolves (likely) to 
{{\Tools\Anaconda3\}} and should resolve to 
{{\Tools\Anaconda3\Lib\site-packages\pyspark\}}. Since I dion't know CMD 
scripting that well, I haven't found solution to this issue yet, apart from the 
workaround, i.e. calling it with full (direct) path.   

> Make PySpark pip install works on windows
> -
>
> Key: SPARK-18136
> URL: https://issues.apache.org/jira/browse/SPARK-18136
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Reporter: holdenk
>
> Make sure that pip installer for PySpark works on windows



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-18136) Make PySpark pip install works on windows

2017-09-21 Thread Apache Spark (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-18136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16175258#comment-16175258
 ] 

Apache Spark commented on SPARK-18136:
--

User 'jsnowacki' has created a pull request for this issue:
https://github.com/apache/spark/pull/19310

> Make PySpark pip install works on windows
> -
>
> Key: SPARK-18136
> URL: https://issues.apache.org/jira/browse/SPARK-18136
> Project: Spark
>  Issue Type: Improvement
>  Components: PySpark
>Reporter: holdenk
>
> Make sure that pip installer for PySpark works on windows



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org