This is an automated email from the ASF dual-hosted git repository.

lfrolov pushed a commit to branch DATALAB-2414
in repository https://gitbox.apache.org/repos/asf/incubator-datalab.git


The following commit(s) were added to refs/heads/DATALAB-2414 by this push:
     new 65d619b  [DATALAB-2414]: fixed python version in kernel name
65d619b is described below

commit 65d619b9b0a26e68e54c4ec1f18fa93178400072
Author: leonidfrolov <[email protected]>
AuthorDate: Tue Jun 15 12:54:58 2021 +0300

    [DATALAB-2414]: fixed python version in kernel name
---
 .../src/general/scripts/os/deeplearning_dataengine_create_configs.py  | 4 ++--
 .../src/general/scripts/os/jupyter_install_dataengine_kernels.py      | 4 ++--
 .../src/general/scripts/os/tensor_dataengine_create_configs.py        | 4 ++--
 3 files changed, 6 insertions(+), 6 deletions(-)

diff --git 
a/infrastructure-provisioning/src/general/scripts/os/deeplearning_dataengine_create_configs.py
 
b/infrastructure-provisioning/src/general/scripts/os/deeplearning_dataengine_create_configs.py
index 9d2f1ad..7dabd3b 100644
--- 
a/infrastructure-provisioning/src/general/scripts/os/deeplearning_dataengine_create_configs.py
+++ 
b/infrastructure-provisioning/src/general/scripts/os/deeplearning_dataengine_create_configs.py
@@ -102,8 +102,8 @@ def install_sparkamagic_kernels(args):
         subprocess.run('sudo jupyter-kernelspec install 
{}/sparkmagic/kernels/sparkkernel --user'.format(sparkmagic_dir), shell=True, 
check=True)
         subprocess.run('sudo jupyter-kernelspec install 
{}/sparkmagic/kernels/pysparkkernel --user'.format(sparkmagic_dir), shell=True, 
check=True)
 
-        pyspark_kernel_name = 'PySpark (Python-3.8 / Spark-{0} ) 
[{1}]'.format(args.spark_version,
-                                                                         
args.cluster_name)
+        pyspark_kernel_name = 'PySpark (Python-{2} / Spark-{0} ) 
[{1}]'.format(args.spark_version,
+                                                                         
args.cluster_name, os.environ['notebook_python_venv_version'][:3])
         subprocess.run('sed -i \'s|PySpark|{0}|g\' 
/home/{1}/.local/share/jupyter/kernels/pysparkkernel/kernel.json'.format(
             pyspark_kernel_name, args.os_user), shell=True, check=True)
         scala_version = subprocess.run('spark-submit --version 2>&1 | grep -o 
-P "Scala version \K.{0,7}"', capture_output=True, shell=True, 
check=True).stdout.decode('UTF-8').rstrip("\n\r")
diff --git 
a/infrastructure-provisioning/src/general/scripts/os/jupyter_install_dataengine_kernels.py
 
b/infrastructure-provisioning/src/general/scripts/os/jupyter_install_dataengine_kernels.py
index b4e172a..8fbc014 100644
--- 
a/infrastructure-provisioning/src/general/scripts/os/jupyter_install_dataengine_kernels.py
+++ 
b/infrastructure-provisioning/src/general/scripts/os/jupyter_install_dataengine_kernels.py
@@ -80,8 +80,8 @@ def install_sparkamagic_kernels(args):
         datalab.fab.conn.sudo('jupyter-kernelspec install 
{}/sparkmagic/kernels/sparkkernel 
--prefix=/home/{}/.local/'.format(sparkmagic_dir, args.os_user))
         datalab.fab.conn.sudo('jupyter-kernelspec install 
{}/sparkmagic/kernels/pysparkkernel 
--prefix=/home/{}/.local/'.format(sparkmagic_dir, args.os_user))
         datalab.fab.conn.sudo('jupyter-kernelspec install 
{}/sparkmagic/kernels/sparkrkernel 
--prefix=/home/{}/.local/'.format(sparkmagic_dir, args.os_user))
-        pyspark_kernel_name = 'PySpark (Python-3.8 / Spark-{0} ) 
[{1}]'.format(args.spark_version,
-                                                                         
args.cluster_name)
+        pyspark_kernel_name = 'PySpark (Python-{2} / Spark-{0} ) 
[{1}]'.format(args.spark_version,
+                                                                         
args.cluster_name, os.environ['notebook_python_venv_version'][:3])
         datalab.fab.conn.sudo('sed -i \'s|PySpark|{0}|g\' 
/home/{1}/.local/share/jupyter/kernels/pysparkkernel/kernel.json'.format(
             pyspark_kernel_name, args.os_user))
         scala_version = datalab.fab.conn.sudo('''bash -l -c 'spark-submit 
--version 2>&1 | grep -o -P "Scala version \K.{0,7}"' ''').stdout.rstrip("\n\r")
diff --git 
a/infrastructure-provisioning/src/general/scripts/os/tensor_dataengine_create_configs.py
 
b/infrastructure-provisioning/src/general/scripts/os/tensor_dataengine_create_configs.py
index 68f6e3f..7e77efe 100644
--- 
a/infrastructure-provisioning/src/general/scripts/os/tensor_dataengine_create_configs.py
+++ 
b/infrastructure-provisioning/src/general/scripts/os/tensor_dataengine_create_configs.py
@@ -101,8 +101,8 @@ def install_sparkamagic_kernels(args):
         subprocess.run('sudo jupyter-kernelspec install 
{}/sparkmagic/kernels/sparkkernel --user'.format(sparkmagic_dir), shell=True, 
check=True)
         subprocess.run('sudo jupyter-kernelspec install 
{}/sparkmagic/kernels/pysparkkernel --user'.format(sparkmagic_dir), shell=True, 
check=True)
 
-        pyspark_kernel_name = 'PySpark (Python-3.8 / Spark-{0} ) 
[{1}]'.format(args.spark_version,
-                                                                         
args.cluster_name)
+        pyspark_kernel_name = 'PySpark (Python-{2} / Spark-{0} ) 
[{1}]'.format(args.spark_version,
+                                                                         
args.cluster_name, os.environ['notebook_python_venv_version'][:3])
         subprocess.run('sed -i \'s|PySpark|{0}|g\' 
/home/{1}/.local/share/jupyter/kernels/pysparkkernel/kernel.json'.format(
             pyspark_kernel_name, args.os_user), shell=True, check=True)
         scala_version = subprocess.run('spark-submit --version 2>&1 | grep -o 
-P "Scala version \K.{0,7}"', capture_output=True, shell=True, 
check=True).stdout.decode('UTF-8').rstrip("\n\r")

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to