[ https://issues.apache.org/jira/browse/SPARK-38425?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17519949#comment-17519949 ]
Apache Spark commented on SPARK-38425: -------------------------------------- User 'lyssg' has created a pull request for this issue: https://github.com/apache/spark/pull/35667 > Avoid possible errors due to incorrect file size or type supplied in hadoop > conf. > --------------------------------------------------------------------------------- > > Key: SPARK-38425 > URL: https://issues.apache.org/jira/browse/SPARK-38425 > Project: Spark > Issue Type: Improvement > Components: Kubernetes > Affects Versions: 3.2.1 > Reporter: luyuan > Priority: Major > > This would avoid failures, in case the files are a bit large or a user places > a binary file inside the HADOOP_CONF_DIR. > Both of which are not supported at the moment. > The reason is, underlying etcd store does limit the size of each entry to > only 1.5 MiB. > [https://etcd.io/docs/v3.4.0/dev-guide/limit/] > We can process this like SPARK-32221 > https://issues.apache.org/jira/browse/SPARK-32221. -- This message was sent by Atlassian Jira (v8.20.1#820001) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org