This is an automated email from the ASF dual-hosted git repository.
lfrolov pushed a commit to branch DATALAB-2414
in repository https://gitbox.apache.org/repos/asf/incubator-datalab.git
The following commit(s) were added to refs/heads/DATALAB-2414 by this push:
new 65d619b [DATALAB-2414]: fixed python version in kernel name
65d619b is described below
commit 65d619b9b0a26e68e54c4ec1f18fa93178400072
Author: leonidfrolov <[email protected]>
AuthorDate: Tue Jun 15 12:54:58 2021 +0300
[DATALAB-2414]: fixed python version in kernel name
---
.../src/general/scripts/os/deeplearning_dataengine_create_configs.py | 4 ++--
.../src/general/scripts/os/jupyter_install_dataengine_kernels.py | 4 ++--
.../src/general/scripts/os/tensor_dataengine_create_configs.py | 4 ++--
3 files changed, 6 insertions(+), 6 deletions(-)
diff --git
a/infrastructure-provisioning/src/general/scripts/os/deeplearning_dataengine_create_configs.py
b/infrastructure-provisioning/src/general/scripts/os/deeplearning_dataengine_create_configs.py
index 9d2f1ad..7dabd3b 100644
---
a/infrastructure-provisioning/src/general/scripts/os/deeplearning_dataengine_create_configs.py
+++
b/infrastructure-provisioning/src/general/scripts/os/deeplearning_dataengine_create_configs.py
@@ -102,8 +102,8 @@ def install_sparkamagic_kernels(args):
subprocess.run('sudo jupyter-kernelspec install
{}/sparkmagic/kernels/sparkkernel --user'.format(sparkmagic_dir), shell=True,
check=True)
subprocess.run('sudo jupyter-kernelspec install
{}/sparkmagic/kernels/pysparkkernel --user'.format(sparkmagic_dir), shell=True,
check=True)
- pyspark_kernel_name = 'PySpark (Python-3.8 / Spark-{0} )
[{1}]'.format(args.spark_version,
-
args.cluster_name)
+ pyspark_kernel_name = 'PySpark (Python-{2} / Spark-{0} )
[{1}]'.format(args.spark_version,
+
args.cluster_name, os.environ['notebook_python_venv_version'][:3])
subprocess.run('sed -i \'s|PySpark|{0}|g\'
/home/{1}/.local/share/jupyter/kernels/pysparkkernel/kernel.json'.format(
pyspark_kernel_name, args.os_user), shell=True, check=True)
scala_version = subprocess.run('spark-submit --version 2>&1 | grep -o
-P "Scala version \K.{0,7}"', capture_output=True, shell=True,
check=True).stdout.decode('UTF-8').rstrip("\n\r")
diff --git
a/infrastructure-provisioning/src/general/scripts/os/jupyter_install_dataengine_kernels.py
b/infrastructure-provisioning/src/general/scripts/os/jupyter_install_dataengine_kernels.py
index b4e172a..8fbc014 100644
---
a/infrastructure-provisioning/src/general/scripts/os/jupyter_install_dataengine_kernels.py
+++
b/infrastructure-provisioning/src/general/scripts/os/jupyter_install_dataengine_kernels.py
@@ -80,8 +80,8 @@ def install_sparkamagic_kernels(args):
datalab.fab.conn.sudo('jupyter-kernelspec install
{}/sparkmagic/kernels/sparkkernel
--prefix=/home/{}/.local/'.format(sparkmagic_dir, args.os_user))
datalab.fab.conn.sudo('jupyter-kernelspec install
{}/sparkmagic/kernels/pysparkkernel
--prefix=/home/{}/.local/'.format(sparkmagic_dir, args.os_user))
datalab.fab.conn.sudo('jupyter-kernelspec install
{}/sparkmagic/kernels/sparkrkernel
--prefix=/home/{}/.local/'.format(sparkmagic_dir, args.os_user))
- pyspark_kernel_name = 'PySpark (Python-3.8 / Spark-{0} )
[{1}]'.format(args.spark_version,
-
args.cluster_name)
+ pyspark_kernel_name = 'PySpark (Python-{2} / Spark-{0} )
[{1}]'.format(args.spark_version,
+
args.cluster_name, os.environ['notebook_python_venv_version'][:3])
datalab.fab.conn.sudo('sed -i \'s|PySpark|{0}|g\'
/home/{1}/.local/share/jupyter/kernels/pysparkkernel/kernel.json'.format(
pyspark_kernel_name, args.os_user))
scala_version = datalab.fab.conn.sudo('''bash -l -c 'spark-submit
--version 2>&1 | grep -o -P "Scala version \K.{0,7}"' ''').stdout.rstrip("\n\r")
diff --git
a/infrastructure-provisioning/src/general/scripts/os/tensor_dataengine_create_configs.py
b/infrastructure-provisioning/src/general/scripts/os/tensor_dataengine_create_configs.py
index 68f6e3f..7e77efe 100644
---
a/infrastructure-provisioning/src/general/scripts/os/tensor_dataengine_create_configs.py
+++
b/infrastructure-provisioning/src/general/scripts/os/tensor_dataengine_create_configs.py
@@ -101,8 +101,8 @@ def install_sparkamagic_kernels(args):
subprocess.run('sudo jupyter-kernelspec install
{}/sparkmagic/kernels/sparkkernel --user'.format(sparkmagic_dir), shell=True,
check=True)
subprocess.run('sudo jupyter-kernelspec install
{}/sparkmagic/kernels/pysparkkernel --user'.format(sparkmagic_dir), shell=True,
check=True)
- pyspark_kernel_name = 'PySpark (Python-3.8 / Spark-{0} )
[{1}]'.format(args.spark_version,
-
args.cluster_name)
+ pyspark_kernel_name = 'PySpark (Python-{2} / Spark-{0} )
[{1}]'.format(args.spark_version,
+
args.cluster_name, os.environ['notebook_python_venv_version'][:3])
subprocess.run('sed -i \'s|PySpark|{0}|g\'
/home/{1}/.local/share/jupyter/kernels/pysparkkernel/kernel.json'.format(
pyspark_kernel_name, args.os_user), shell=True, check=True)
scala_version = subprocess.run('spark-submit --version 2>&1 | grep -o
-P "Scala version \K.{0,7}"', capture_output=True, shell=True,
check=True).stdout.decode('UTF-8').rstrip("\n\r")
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]