This is an automated email from the ASF dual-hosted git repository. zjffdu pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/zeppelin.git
The following commit(s) were added to refs/heads/master by this push: new 8e6974f ZEPPELIN-4038. Deprecate spark 2.2 and earlier 8e6974f is described below commit 8e6974fdc33e834bc01a5ee594e2cfca4ff3045f Author: Jeff Zhang <zjf...@apache.org> AuthorDate: Mon Mar 25 17:56:52 2019 +0800 ZEPPELIN-4038. Deprecate spark 2.2 and earlier ### What is this PR for? This PR is to deprecate spark 2.2 and earlier. It would display a warning message in frontend, and user can disable this message by configure `zeppelin.spark.deprecatedMsg.show` to false, by default it is true. ### What type of PR is it? [Bug Fix] ### Todos * [ ] - Task ### What is the Jira issue? * https://jira.apache.org/jira/browse/ZEPPELIN-4038 ### How should this be tested? * CI pass ### Screenshots (if appropriate) ![image](https://user-images.githubusercontent.com/164491/54910937-4ebff580-4f28-11e9-8bdf-a4327f814990.png) ### Questions: * Does the licenses files need update? * Is there breaking changes for older versions? * Does this needs documentation? Author: Jeff Zhang <zjf...@apache.org> Closes #3341 from zjffdu/ZEPPELIN-4038 and squashes the following commits: a66445678 [Jeff Zhang] ZEPPELIN-4038. Deprecate spark 2.2 and earlier --- .../zeppelin/img/docs-img/spark_deprecate.png | Bin 0 -> 176802 bytes docs/interpreter/spark.md | 6 ++++++ .../apache/zeppelin/spark/IPySparkInterpreter.java | 1 + .../apache/zeppelin/spark/PySparkInterpreter.java | 1 + .../apache/zeppelin/spark/SparkInterpreter.java | 2 ++ .../apache/zeppelin/spark/SparkRInterpreter.java | 3 ++- .../apache/zeppelin/spark/SparkSqlInterpreter.java | 2 +- .../org/apache/zeppelin/spark/SparkVersion.java | 1 + .../main/java/org/apache/zeppelin/spark/Utils.java | 22 +++++++++++++++++++++ .../src/main/resources/interpreter-setting.json | 7 +++++++ .../zeppelin/spark/IPySparkInterpreterTest.java | 1 + .../zeppelin/spark/NewSparkInterpreterTest.java | 7 +++++++ .../zeppelin/spark/NewSparkSqlInterpreterTest.java | 1 + .../zeppelin/spark/OldSparkInterpreterTest.java | 1 + .../zeppelin/spark/OldSparkSqlInterpreterTest.java | 1 + .../spark/PySparkInterpreterMatplotlibTest.java | 1 + .../zeppelin/spark/PySparkInterpreterTest.java | 1 + .../zeppelin/spark/SparkRInterpreterTest.java | 3 ++- .../zeppelin/spark/BaseSparkScalaInterpreter.scala | 2 +- .../integration/ZeppelinSparkClusterTest.java | 2 ++ 20 files changed, 61 insertions(+), 4 deletions(-) diff --git a/docs/assets/themes/zeppelin/img/docs-img/spark_deprecate.png b/docs/assets/themes/zeppelin/img/docs-img/spark_deprecate.png new file mode 100644 index 0000000..8a867cc Binary files /dev/null and b/docs/assets/themes/zeppelin/img/docs-img/spark_deprecate.png differ diff --git a/docs/interpreter/spark.md b/docs/interpreter/spark.md index 9140825..c3c9fd7 100644 --- a/docs/interpreter/spark.md +++ b/docs/interpreter/spark.md @@ -374,6 +374,12 @@ Logical setup with Zeppelin, Kerberos Key Distribution Center (KDC), and Spark o <img src="{{BASE_PATH}}/assets/themes/zeppelin/img/docs-img/kdc_zeppelin.png"> +## Deprecate Spark 2.2 and earlier versions +Starting from 0.9, Zeppelin deprecate Spark 2.2 and earlier versions. So you will see a warning message when you use Spark 2.2 and earlier. +You can get rid of this message by setting `zeppelin.spark.deprecatedMsg.show` to `false`. + +<img src="{{BASE_PATH}}/assets/themes/zeppelin/img/docs-img/spark_deprecate.png"> + ### Configuration Setup 1. On the server that Zeppelin is installed, install Kerberos client modules and configuration, krb5.conf. diff --git a/spark/interpreter/src/main/java/org/apache/zeppelin/spark/IPySparkInterpreter.java b/spark/interpreter/src/main/java/org/apache/zeppelin/spark/IPySparkInterpreter.java index 594c171..c69e6fd 100644 --- a/spark/interpreter/src/main/java/org/apache/zeppelin/spark/IPySparkInterpreter.java +++ b/spark/interpreter/src/main/java/org/apache/zeppelin/spark/IPySparkInterpreter.java @@ -85,6 +85,7 @@ public class IPySparkInterpreter extends IPythonInterpreter { @Override public InterpreterResult interpret(String st, InterpreterContext context) throws InterpreterException { + Utils.printDeprecateMessage(sparkInterpreter.getSparkVersion(), context, properties); InterpreterContext.set(context); String jobGroupId = Utils.buildJobGroupId(context); String jobDesc = Utils.buildJobDesc(context); diff --git a/spark/interpreter/src/main/java/org/apache/zeppelin/spark/PySparkInterpreter.java b/spark/interpreter/src/main/java/org/apache/zeppelin/spark/PySparkInterpreter.java index 486eca0..c365345 100644 --- a/spark/interpreter/src/main/java/org/apache/zeppelin/spark/PySparkInterpreter.java +++ b/spark/interpreter/src/main/java/org/apache/zeppelin/spark/PySparkInterpreter.java @@ -147,6 +147,7 @@ public class PySparkInterpreter extends PythonInterpreter { @Override public InterpreterResult interpret(String st, InterpreterContext context) throws InterpreterException { + Utils.printDeprecateMessage(sparkInterpreter.getSparkVersion(), context, properties); return super.interpret(st, context); } diff --git a/spark/interpreter/src/main/java/org/apache/zeppelin/spark/SparkInterpreter.java b/spark/interpreter/src/main/java/org/apache/zeppelin/spark/SparkInterpreter.java index 43b9e76..ec0be5f 100644 --- a/spark/interpreter/src/main/java/org/apache/zeppelin/spark/SparkInterpreter.java +++ b/spark/interpreter/src/main/java/org/apache/zeppelin/spark/SparkInterpreter.java @@ -28,6 +28,7 @@ import org.apache.zeppelin.interpreter.thrift.InterpreterCompletion; import org.slf4j.Logger; import org.slf4j.LoggerFactory; +import java.io.IOException; import java.util.List; import java.util.Properties; @@ -74,6 +75,7 @@ public class SparkInterpreter extends AbstractSparkInterpreter { @Override public InterpreterResult internalInterpret(String st, InterpreterContext context) throws InterpreterException { + Utils.printDeprecateMessage(delegation.getSparkVersion(), context, properties); return delegation.interpret(st, context); } diff --git a/spark/interpreter/src/main/java/org/apache/zeppelin/spark/SparkRInterpreter.java b/spark/interpreter/src/main/java/org/apache/zeppelin/spark/SparkRInterpreter.java index 7265ae4..3b14eed 100644 --- a/spark/interpreter/src/main/java/org/apache/zeppelin/spark/SparkRInterpreter.java +++ b/spark/interpreter/src/main/java/org/apache/zeppelin/spark/SparkRInterpreter.java @@ -121,7 +121,8 @@ public class SparkRInterpreter extends Interpreter { @Override public InterpreterResult interpret(String lines, InterpreterContext interpreterContext) throws InterpreterException { - + Utils.printDeprecateMessage(sparkInterpreter.getSparkVersion(), + interpreterContext, properties); String jobGroup = Utils.buildJobGroupId(interpreterContext); String jobDesc = Utils.buildJobDesc(interpreterContext); sparkInterpreter.getSparkContext().setJobGroup(jobGroup, jobDesc, false); diff --git a/spark/interpreter/src/main/java/org/apache/zeppelin/spark/SparkSqlInterpreter.java b/spark/interpreter/src/main/java/org/apache/zeppelin/spark/SparkSqlInterpreter.java index 040556b..4e63760 100644 --- a/spark/interpreter/src/main/java/org/apache/zeppelin/spark/SparkSqlInterpreter.java +++ b/spark/interpreter/src/main/java/org/apache/zeppelin/spark/SparkSqlInterpreter.java @@ -78,7 +78,7 @@ public class SparkSqlInterpreter extends AbstractInterpreter { return new InterpreterResult(Code.ERROR, "Spark " + sparkInterpreter.getSparkVersion().toString() + " is not supported"); } - + Utils.printDeprecateMessage(sparkInterpreter.getSparkVersion(), context, properties); sparkInterpreter.getZeppelinContext().setInterpreterContext(context); SQLContext sqlc = sparkInterpreter.getSQLContext(); SparkContext sc = sqlc.sparkContext(); diff --git a/spark/interpreter/src/main/java/org/apache/zeppelin/spark/SparkVersion.java b/spark/interpreter/src/main/java/org/apache/zeppelin/spark/SparkVersion.java index de65105..88b10ef 100644 --- a/spark/interpreter/src/main/java/org/apache/zeppelin/spark/SparkVersion.java +++ b/spark/interpreter/src/main/java/org/apache/zeppelin/spark/SparkVersion.java @@ -28,6 +28,7 @@ public class SparkVersion { public static final SparkVersion SPARK_1_6_0 = SparkVersion.fromVersionString("1.6.0"); public static final SparkVersion SPARK_2_0_0 = SparkVersion.fromVersionString("2.0.0"); + public static final SparkVersion SPARK_2_2_0 = SparkVersion.fromVersionString("2.2.0"); public static final SparkVersion SPARK_2_3_0 = SparkVersion.fromVersionString("2.3.0"); public static final SparkVersion SPARK_2_3_1 = SparkVersion.fromVersionString("2.3.1"); public static final SparkVersion SPARK_2_4_0 = SparkVersion.fromVersionString("2.4.0"); diff --git a/spark/interpreter/src/main/java/org/apache/zeppelin/spark/Utils.java b/spark/interpreter/src/main/java/org/apache/zeppelin/spark/Utils.java index cd6c607..1220bd7 100644 --- a/spark/interpreter/src/main/java/org/apache/zeppelin/spark/Utils.java +++ b/spark/interpreter/src/main/java/org/apache/zeppelin/spark/Utils.java @@ -18,10 +18,12 @@ package org.apache.zeppelin.spark; import org.apache.zeppelin.interpreter.InterpreterContext; +import org.apache.zeppelin.interpreter.InterpreterException; import org.apache.zeppelin.user.AuthenticationInfo; import org.slf4j.Logger; import org.slf4j.LoggerFactory; +import java.io.IOException; import java.lang.reflect.Constructor; import java.lang.reflect.InvocationTargetException; import java.util.Properties; @@ -33,6 +35,10 @@ import java.util.regex.Pattern; */ class Utils { public static Logger logger = LoggerFactory.getLogger(Utils.class); + public static String DEPRRECATED_MESSAGE = + "%html <font color=\"red\">Spark lower than 2.2 is deprecated, " + + "if you don't want to see this message, please set " + + "zeppelin.spark.deprecateMsg.show to false.</font>"; private static final String SCALA_COMPILER_VERSION = evaluateScalaCompilerVersion(); static Object invokeMethod(Object o, String name) { @@ -178,4 +184,20 @@ class Utils { } return uName; } + + public static void printDeprecateMessage(SparkVersion sparkVersion, + InterpreterContext context, + Properties properties) throws InterpreterException { + context.out.clear(); + if (sparkVersion.olderThan(SparkVersion.SPARK_2_2_0) + && Boolean.parseBoolean( + properties.getProperty("zeppelin.spark.deprecatedMsg.show", "true"))) { + try { + context.out.write(DEPRRECATED_MESSAGE); + context.out.write("%text "); + } catch (IOException e) { + throw new InterpreterException(e); + } + } + } } diff --git a/spark/interpreter/src/main/resources/interpreter-setting.json b/spark/interpreter/src/main/resources/interpreter-setting.json index 5ff5cfe..341beda 100644 --- a/spark/interpreter/src/main/resources/interpreter-setting.json +++ b/spark/interpreter/src/main/resources/interpreter-setting.json @@ -95,6 +95,13 @@ "defaultValue": true, "description": "Whether to enable color output of spark scala interpreter", "type": "checkbox" + }, + "zeppelin.spark.deprecatedMsg.show": { + "envName": null, + "propertyName": "zeppelin.spark.deprecatedMsg.show", + "defaultValue": true, + "description": "Whether show the spark deprecated message", + "type": "checkbox" } }, "editor": { diff --git a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/IPySparkInterpreterTest.java b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/IPySparkInterpreterTest.java index 20c336d..b8870e6 100644 --- a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/IPySparkInterpreterTest.java +++ b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/IPySparkInterpreterTest.java @@ -64,6 +64,7 @@ public class IPySparkInterpreterTest extends IPythonInterpreterTest { p.setProperty("zeppelin.pyspark.python", "python"); p.setProperty("zeppelin.dep.localrepo", Files.createTempDir().getAbsolutePath()); p.setProperty("zeppelin.python.gatewayserver_address", "127.0.0.1"); + p.setProperty("zeppelin.spark.deprecatedMsg.show", "false"); return p; } diff --git a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkInterpreterTest.java b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkInterpreterTest.java index 7c7dad9..773deae 100644 --- a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkInterpreterTest.java +++ b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkInterpreterTest.java @@ -86,6 +86,7 @@ public class NewSparkInterpreterTest { properties.setProperty("zeppelin.spark.uiWebUrl", "fake_spark_weburl"); // disable color output for easy testing properties.setProperty("zeppelin.spark.scala.color", "false"); + properties.setProperty("zeppelin.spark.deprecatedMsg.show", "false"); InterpreterContext context = InterpreterContext.builder() .setInterpreterOut(new InterpreterOutput(null)) @@ -382,6 +383,7 @@ public class NewSparkInterpreterTest { properties.setProperty("zeppelin.dep.localrepo", Files.createTempDir().getAbsolutePath()); // disable color output for easy testing properties.setProperty("zeppelin.spark.scala.color", "false"); + properties.setProperty("zeppelin.spark.deprecatedMsg.show", "false"); InterpreterGroup intpGroup = new InterpreterGroup(); interpreter = new SparkInterpreter(properties); @@ -413,6 +415,7 @@ public class NewSparkInterpreterTest { properties.setProperty("zeppelin.spark.printREPLOutput", "false"); // disable color output for easy testing properties.setProperty("zeppelin.spark.scala.color", "false"); + properties.setProperty("zeppelin.spark.deprecatedMsg.show", "false"); InterpreterContext.set(getInterpreterContext()); interpreter = new SparkInterpreter(properties); @@ -442,6 +445,7 @@ public class NewSparkInterpreterTest { properties.setProperty("spark.scheduler.mode", "FAIR"); // disable color output for easy testing properties.setProperty("zeppelin.spark.scala.color", "false"); + properties.setProperty("zeppelin.spark.deprecatedMsg.show", "false"); interpreter = new SparkInterpreter(properties); assertTrue(interpreter.getDelegation() instanceof NewSparkInterpreter); @@ -473,6 +477,7 @@ public class NewSparkInterpreterTest { properties.setProperty("spark.ui.enabled", "false"); // disable color output for easy testing properties.setProperty("zeppelin.spark.scala.color", "false"); + properties.setProperty("zeppelin.spark.deprecatedMsg.show", "false"); interpreter = new SparkInterpreter(properties); assertTrue(interpreter.getDelegation() instanceof NewSparkInterpreter); @@ -500,6 +505,7 @@ public class NewSparkInterpreterTest { properties.setProperty("zeppelin.spark.ui.hidden", "true"); // disable color output for easy testing properties.setProperty("zeppelin.spark.scala.color", "false"); + properties.setProperty("zeppelin.spark.deprecatedMsg.show", "false"); interpreter = new SparkInterpreter(properties); assertTrue(interpreter.getDelegation() instanceof NewSparkInterpreter); @@ -525,6 +531,7 @@ public class NewSparkInterpreterTest { properties.setProperty("zeppelin.spark.useNew", "true"); // disable color output for easy testing properties.setProperty("zeppelin.spark.scala.color", "false"); + properties.setProperty("zeppelin.spark.deprecatedMsg.show", "false"); SparkInterpreter interpreter1 = new SparkInterpreter(properties); SparkInterpreter interpreter2 = new SparkInterpreter(properties); diff --git a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkSqlInterpreterTest.java b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkSqlInterpreterTest.java index 573ce51..ac5c9e1 100644 --- a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkSqlInterpreterTest.java +++ b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkSqlInterpreterTest.java @@ -55,6 +55,7 @@ public class NewSparkSqlInterpreterTest { p.setProperty("zeppelin.spark.sql.stacktrace", "true"); p.setProperty("zeppelin.spark.useNew", "true"); p.setProperty("zeppelin.spark.useHiveContext", "true"); + p.setProperty("zeppelin.spark.deprecatedMsg.show", "false"); intpGroup = new InterpreterGroup(); sparkInterpreter = new SparkInterpreter(p); diff --git a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkInterpreterTest.java b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkInterpreterTest.java index a435272..c2a1bb0 100644 --- a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkInterpreterTest.java +++ b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkInterpreterTest.java @@ -87,6 +87,7 @@ public class OldSparkInterpreterTest { p.setProperty("zeppelin.spark.property_1", "value_1"); // disable color output for easy testing p.setProperty("zeppelin.spark.scala.color", "false"); + p.setProperty("zeppelin.spark.deprecatedMsg.show", "false"); return p; } diff --git a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkSqlInterpreterTest.java b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkSqlInterpreterTest.java index fa1e257..425651c 100644 --- a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkSqlInterpreterTest.java +++ b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkSqlInterpreterTest.java @@ -57,6 +57,7 @@ public class OldSparkSqlInterpreterTest { p.setProperty("zeppelin.spark.maxResult", "10"); p.setProperty("zeppelin.spark.concurrentSQL", "false"); p.setProperty("zeppelin.spark.sql.stacktrace", "false"); + p.setProperty("zeppelin.spark.deprecatedMsg.show", "false"); repl = new SparkInterpreter(p); intpGroup = new InterpreterGroup(); diff --git a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterMatplotlibTest.java b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterMatplotlibTest.java index 5a05ad5..305839a 100644 --- a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterMatplotlibTest.java +++ b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterMatplotlibTest.java @@ -102,6 +102,7 @@ public class PySparkInterpreterMatplotlibTest { p.setProperty("zeppelin.dep.localrepo", tmpDir.newFolder().getAbsolutePath()); p.setProperty("zeppelin.pyspark.useIPython", "false"); p.setProperty("zeppelin.python.gatewayserver_address", "127.0.0.1"); + p.setProperty("zeppelin.spark.deprecatedMsg.show", "false"); return p; } diff --git a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterTest.java b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterTest.java index 64f1ff5..5681c18 100644 --- a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterTest.java +++ b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterTest.java @@ -54,6 +54,7 @@ public class PySparkInterpreterTest extends PythonInterpreterTest { properties.setProperty("zeppelin.spark.useNew", "true"); properties.setProperty("zeppelin.spark.test", "true"); properties.setProperty("zeppelin.python.gatewayserver_address", "127.0.0.1"); + properties.setProperty("zeppelin.spark.deprecatedMsg.show", "false"); // create interpreter group intpGroup = new InterpreterGroup(); diff --git a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/SparkRInterpreterTest.java b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/SparkRInterpreterTest.java index fb9ad62..ae48cbb 100644 --- a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/SparkRInterpreterTest.java +++ b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/SparkRInterpreterTest.java @@ -56,7 +56,8 @@ public class SparkRInterpreterTest { properties.setProperty("zeppelin.spark.useNew", "true"); properties.setProperty("zeppelin.R.knitr", "true"); properties.setProperty("spark.r.backendConnectionTimeout", "10"); - + properties.setProperty("zeppelin.spark.deprecatedMsg.show", "false"); + InterpreterContext context = getInterpreterContext(); InterpreterContext.set(context); sparkRInterpreter = new SparkRInterpreter(properties); diff --git a/spark/spark-scala-parent/src/main/scala/org/apache/zeppelin/spark/BaseSparkScalaInterpreter.scala b/spark/spark-scala-parent/src/main/scala/org/apache/zeppelin/spark/BaseSparkScalaInterpreter.scala index cd241a8..3a2cd0b 100644 --- a/spark/spark-scala-parent/src/main/scala/org/apache/zeppelin/spark/BaseSparkScalaInterpreter.scala +++ b/spark/spark-scala-parent/src/main/scala/org/apache/zeppelin/spark/BaseSparkScalaInterpreter.scala @@ -95,7 +95,6 @@ abstract class BaseSparkScalaInterpreter(val conf: SparkConf, System.setOut(Console.out) interpreterOutput.setInterpreterOutput(context.out) interpreterOutput.ignoreLeadingNewLinesFromScalaReporter() - context.out.clear() val status = scalaInterpret(code) match { case succ...@scala.tools.nsc.interpreter.ir.Success => @@ -122,6 +121,7 @@ abstract class BaseSparkScalaInterpreter(val conf: SparkConf, // reset the java stdout System.setOut(originalOut) + context.out.write("") val lastStatus = _interpret(code) match { case scala.tools.nsc.interpreter.IR.Success => InterpreterResult.Code.SUCCESS diff --git a/zeppelin-interpreter-integration/src/test/java/org/apache/zeppelin/integration/ZeppelinSparkClusterTest.java b/zeppelin-interpreter-integration/src/test/java/org/apache/zeppelin/integration/ZeppelinSparkClusterTest.java index 78ff006..8d29163 100644 --- a/zeppelin-interpreter-integration/src/test/java/org/apache/zeppelin/integration/ZeppelinSparkClusterTest.java +++ b/zeppelin-interpreter-integration/src/test/java/org/apache/zeppelin/integration/ZeppelinSparkClusterTest.java @@ -112,6 +112,8 @@ public abstract class ZeppelinSparkClusterTest extends AbstractTestRestApi { new InterpreterProperty("spark.serializer", "org.apache.spark.serializer.KryoSerializer")); sparkProperties.put("zeppelin.spark.scala.color", new InterpreterProperty("zeppelin.spark.scala.color", "false")); + sparkProperties.put("zzeppelin.spark.deprecatedMsg.show", + new InterpreterProperty("zeppelin.spark.deprecatedMsg.show", "false")); TestUtils.getInstance(Notebook.class).getInterpreterSettingManager().restart(sparkIntpSetting.getId()); }