[ https://issues.apache.org/jira/browse/SPARK-37170?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Kousuke Saruta updated SPARK-37170: ----------------------------------- Description: I noticed that the PySpark 3.1.2 is installed in the live notebook environment even though the notebook is for PySpark 3.2. http://spark.apache.org/docs/3.2.0/api/python/getting_started/index.html I guess someone accessed to Binder and built the container image with v3.2.0 before we published the pyspark package to PyPi. https://mybinder.org/ I think it's difficult to rebuild the image manually. To avoid such accident, I'll propose to pin the version of PySpark in binder/postBuild was: I noticed that the PySpark 3.1.2 is installed in the environment of live notebook even though the notebook is for PySpark 3.2. http://spark.apache.org/docs/3.2.0/api/python/getting_started/index.html I guess someone accessed to Binder and built the container image with v3.2.0 before we published the pyspark package to PyPi. https://mybinder.org/ I think it's difficult to rebuild the image manually. To avoid such accident, I'll propose to pin the version of PySpark in binder/postBuild > Pin PySpark version for Binder > ------------------------------ > > Key: SPARK-37170 > URL: https://issues.apache.org/jira/browse/SPARK-37170 > Project: Spark > Issue Type: Bug > Components: docs, PySpark > Affects Versions: 3.2.0 > Reporter: Kousuke Saruta > Assignee: Kousuke Saruta > Priority: Major > > I noticed that the PySpark 3.1.2 is installed in the live notebook > environment even though the notebook is for PySpark 3.2. > http://spark.apache.org/docs/3.2.0/api/python/getting_started/index.html > I guess someone accessed to Binder and built the container image with v3.2.0 > before we published the pyspark package to PyPi. > https://mybinder.org/ > I think it's difficult to rebuild the image manually. > To avoid such accident, I'll propose to pin the version of PySpark in > binder/postBuild > > -- This message was sent by Atlassian Jira (v8.3.4#803005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org