[ 
https://issues.apache.org/jira/browse/SPARK-51035?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17985049#comment-17985049
 ] 

Yaniv Kunda commented on SPARK-51035:
-------------------------------------

[~LuciferYang] I don't think it's possible for user input to be passed to 
Janino, but I feel [~marinagonzalez]'s request from an operational perspective:
When faced with auditing (e.g. in case of FedRAMP validation) - the existence 
of vulnerable dependencies needs either a resolution or an explanation (to 
warrant an exemption); it's can be hard enough to prove an application cannot 
be used in a way that exploits a vulnerability in a direct dependency - but 
it's much harder to do so for transitive dependencies.
Patching these by Spark users can be problematic as well - creating possible 
runtime issues.

[~chengpan] libfb303 is indeed the latest version, but has not been released 
since 2015 - and I couldn't find any actual usages in the project.

If it makes sense, I can try and upgrade these and see how tests fare.
Regarding Derby, recent discussions hint that it's not going to be updated for 
the 10.16 branch soon -
https://issues.apache.org/jira/browse/DERBY-7147

> Upgrade dependencies exposed to critical and high CVEs
> ------------------------------------------------------
>
>                 Key: SPARK-51035
>                 URL: https://issues.apache.org/jira/browse/SPARK-51035
>             Project: Spark
>          Issue Type: Dependency upgrade
>          Components: PySpark
>    Affects Versions: 4.0.0
>            Reporter: Marina Gonzalez
>            Priority: Critical
>
> Several outdated library dependencies still referenced in *PySpark 4.0.0* 
> contain {*}high/critical security vulnerabilities (CVEs){*}. If not updated, 
> these vulnerabilities could affect users who rely on PySpark for production 
> workloads.
>  * 
> {color:#172b4d}[{*}derby{*}:|https://github.com/apache/spark/blob/b49ef2ada0fcce7e3b5559abaf067d11f324d733/pom.xml#L139]{color}
>  
> _[v.10.16.1.1|https://github.com/apache/spark/blob/b49ef2ada0fcce7e3b5559abaf067d11f324d733/pom.xml#L139]_
>  referenced by pyspark4.0.0 does [not 
> resolve|https://mvnrepository.com/artifact/org.apache.derby/derby/10.16.1.1] 
> the vulnerability ([NVD - 
> CVE-2022-46337|https://nvd.nist.gov/vuln/detail?vulnId=CVE-2022-46337])
>  * {*}libfb303-0.9.3.jar{*}: 
> [v.0.9.3|https://github.com/apache/spark/blob/b49ef2ada0fcce7e3b5559abaf067d11f324d733/pom.xml#L2486]
>  is still referenced by pyspark4.0.0 does [not 
> resolve|https://mvnrepository.com/artifact/org.apache.thrift/libfb303/0.9.3] 
> some key vulnerabilities
>  ** libfb303-0.9.3.jar comes from thrift v0.9.3 which is 10 years old (flag)
>  * {*}janino-3.1.9.jar{*}: 
> [v.3.1.9|https://github.com/apache/spark/blob/b49ef2ada0fcce7e3b5559abaf067d11f324d733/pom.xml#L200]
>  still referenced by pyspark4.0.0 does [not 
> resolve|https://nvd.nist.gov/vuln/detail/CVE-2023-33546] the vulnerability 
> ([NVD - 
> CVE-2023-33546|https://nvd.nist.gov/vuln/detail?vulnId=CVE-2023-33546])
> These vulnerabilities affect Spark users by:
>  * Exposing PySpark workflows to potential security risks.
>  * Including outdated dependencies that lack recent security fixes.
>  * Forcing users to patch Spark manually instead of using an official release.
> Expected Behaviour:
>  * Ideally, the dependencies should be upgraded to their latest secure 
> versions before the final release of PySpark 4.0.0.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to