This is an automated email from the ASF dual-hosted git repository.

casion pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/linkis.git


The following commit(s) were added to refs/heads/master by this push:
     new 2fdf9d3e2e [EnginePlugin][Spark]Turn off use secure random by default 
(#5197)
2fdf9d3e2e is described below

commit 2fdf9d3e2ed2ed29a4c6a2382249286883818acb
Author: peacewong <[email protected]>
AuthorDate: Tue Nov 12 10:40:11 2024 +0800

    [EnginePlugin][Spark]Turn off use secure random by default (#5197)
    
    * Turn off use secure random by default close #5196
    
    * Update Notification Mailing List
    
    * Fix ds meta service build
---
 .asf.yaml                                                    |  4 ++--
 .../engineplugin/spark/config/SparkConfiguration.scala       |  5 ++++-
 .../engineplugin/spark/executor/SparkPythonExecutor.scala    |  9 +++++++--
 linkis-public-enhancements/distribution.xml                  |  2 +-
 .../server/src/main/assembly/distribution.xml                | 12 ++++++------
 5 files changed, 20 insertions(+), 12 deletions(-)

diff --git a/.asf.yaml b/.asf.yaml
index 79396eba55..08ec821f75 100644
--- a/.asf.yaml
+++ b/.asf.yaml
@@ -67,6 +67,6 @@ github:
         required_approving_review_count: 1
 notifications:
   commits: [email protected]
-  issues: [email protected]
-  pullrequests: [email protected]
+  issues: [email protected]
+  pullrequests: [email protected]
 
diff --git 
a/linkis-engineconn-plugins/spark/src/main/scala/org/apache/linkis/engineplugin/spark/config/SparkConfiguration.scala
 
b/linkis-engineconn-plugins/spark/src/main/scala/org/apache/linkis/engineplugin/spark/config/SparkConfiguration.scala
index a493c5ff37..429048c77f 100644
--- 
a/linkis-engineconn-plugins/spark/src/main/scala/org/apache/linkis/engineplugin/spark/config/SparkConfiguration.scala
+++ 
b/linkis-engineconn-plugins/spark/src/main/scala/org/apache/linkis/engineplugin/spark/config/SparkConfiguration.scala
@@ -156,7 +156,7 @@ object SparkConfiguration extends Logging {
     CommonVars("wds.linkis.spark.engineconn.fatal.log", "error writing 
class;OutOfMemoryError")
 
   val PYSPARK_PYTHON3_PATH =
-    CommonVars[String]("pyspark.python3.path", 
"/appcom/Install/anaconda3/bin/python")
+    CommonVars[String]("pyspark.python3.path", "python3")
 
   val ENABLE_REPLACE_PACKAGE_NAME =
     CommonVars("wds.linkis.spark.engine.scala.replace_package_header.enable", 
true)
@@ -182,6 +182,9 @@ object SparkConfiguration extends Logging {
 
   val LINKIS_SPARK_ETL_SUPPORT_HUDI = 
CommonVars[Boolean]("linkis.spark.etl.support.hudi", false)
 
+  val LINKIS_PYSPARK_USE_SECURE_RANDOM =
+    CommonVars[Boolean]("linkis.pyspark.use.secure.random", false).getValue
+
   val SCALA_PARSE_APPEND_CODE =
     CommonVars("linkis.scala.parse.append.code", "val linkisVar=1").getValue
 
diff --git 
a/linkis-engineconn-plugins/spark/src/main/scala/org/apache/linkis/engineplugin/spark/executor/SparkPythonExecutor.scala
 
b/linkis-engineconn-plugins/spark/src/main/scala/org/apache/linkis/engineplugin/spark/executor/SparkPythonExecutor.scala
index 5d4305c67c..f947db9338 100644
--- 
a/linkis-engineconn-plugins/spark/src/main/scala/org/apache/linkis/engineplugin/spark/executor/SparkPythonExecutor.scala
+++ 
b/linkis-engineconn-plugins/spark/src/main/scala/org/apache/linkis/engineplugin/spark/executor/SparkPythonExecutor.scala
@@ -46,6 +46,7 @@ import org.apache.spark.sql.execution.datasources.csv.UDF
 
 import java.io._
 import java.net.InetAddress
+import java.security.SecureRandom
 import java.util
 
 import scala.collection.JavaConverters._
@@ -76,7 +77,12 @@ class SparkPythonExecutor(val sparkEngineSession: 
SparkEngineSession, val id: In
   private val lineOutputStream = new RsOutputStream
   val sqlContext = sparkEngineSession.sqlContext
   val SUCCESS = "success"
-  private lazy val py4jToken: String = 
SecureRandomStringUtils.randomAlphanumeric(256)
+
+  private lazy val py4jToken: String = if 
(SparkConfiguration.LINKIS_PYSPARK_USE_SECURE_RANDOM) {
+    SecureRandomStringUtils.randomAlphanumeric(256)
+  } else {
+    SecureRandom.getInstance("SHA1PRNG").nextInt(100000).toString
+  }
 
   private lazy val gwBuilder: GatewayServerBuilder = {
     val builder = new GatewayServerBuilder()
@@ -152,7 +158,6 @@ class SparkPythonExecutor(val sparkEngineSession: 
SparkEngineSession, val id: In
     )
     val userDefinePythonVersion = engineCreationContext.getOptions
       .getOrDefault("spark.python.version", "python")
-      .toString
       .toLowerCase()
     val sparkPythonVersion =
       if (
diff --git a/linkis-public-enhancements/distribution.xml 
b/linkis-public-enhancements/distribution.xml
index 3e3baf6396..b4cc88582e 100644
--- a/linkis-public-enhancements/distribution.xml
+++ b/linkis-public-enhancements/distribution.xml
@@ -290,7 +290,7 @@
     <fileSets>
         <fileSet>
             <directory>
-                
./linkis-datasource/linkis-datasource-manager/server/target/out/lib/service
+                
./linkis-datasource/linkis-datasource-manager/server/target/out/service
             </directory>
             <outputDirectory>
                     lib/metadataquery-service
diff --git 
a/linkis-public-enhancements/linkis-datasource/linkis-datasource-manager/server/src/main/assembly/distribution.xml
 
b/linkis-public-enhancements/linkis-datasource/linkis-datasource-manager/server/src/main/assembly/distribution.xml
index 26130980b3..e124c84fc4 100644
--- 
a/linkis-public-enhancements/linkis-datasource/linkis-datasource-manager/server/src/main/assembly/distribution.xml
+++ 
b/linkis-public-enhancements/linkis-datasource/linkis-datasource-manager/server/src/main/assembly/distribution.xml
@@ -187,7 +187,7 @@
         <fileSet>
             
<directory>${basedir}/../service/elasticsearch/target/out/lib</directory>
             <fileMode>0755</fileMode>
-            <outputDirectory>lib/service/elasticsearch</outputDirectory>
+            <outputDirectory>service/elasticsearch</outputDirectory>
             <includes>
                 <include>*.jar</include>
             </includes>
@@ -198,7 +198,7 @@
         <fileSet>
             <directory>${basedir}/../service/hive/target/out/lib</directory>
             <fileMode>0755</fileMode>
-            <outputDirectory>lib/service/hive</outputDirectory>
+            <outputDirectory>service/hive</outputDirectory>
             <includes>
                 <include>*.jar</include>
             </includes>
@@ -209,7 +209,7 @@
         <fileSet>
             <directory>${basedir}/../service/kafka/target/out/lib</directory>
             <fileMode>0755</fileMode>
-            <outputDirectory>lib/service/kafka</outputDirectory>
+            <outputDirectory>service/kafka</outputDirectory>
             <includes>
                 <include>*.jar</include>
             </includes>
@@ -220,7 +220,7 @@
         <fileSet>
             <directory>${basedir}/../service/jdbc/target/out/lib</directory>
             <fileMode>0755</fileMode>
-            <outputDirectory>lib/service/jdbc</outputDirectory>
+            <outputDirectory>service/jdbc</outputDirectory>
             <includes>
                 <include>*.jar</include>
             </includes>
@@ -231,7 +231,7 @@
         <fileSet>
             <directory>${basedir}/../service/hdfs/target/out/lib</directory>
             <fileMode>0755</fileMode>
-            <outputDirectory>lib/service/hdfs</outputDirectory>
+            <outputDirectory>service/hdfs</outputDirectory>
             <includes>
                 <include>*.jar</include>
             </includes>
@@ -242,7 +242,7 @@
         <fileSet>
             <directory>${basedir}/../service/mongodb/target/out/lib</directory>
             <fileMode>0755</fileMode>
-            <outputDirectory>lib/service/mongodb</outputDirectory>
+            <outputDirectory>service/mongodb</outputDirectory>
             <includes>
                 <include>*.jar</include>
             </includes>


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to