[ https://issues.apache.org/jira/browse/SPARK-18432?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Kousuke Saruta updated SPARK-18432: ----------------------------------- Assignee: Noritaka Sekiyama > Fix HDFS block size in programming guide > ---------------------------------------- > > Key: SPARK-18432 > URL: https://issues.apache.org/jira/browse/SPARK-18432 > Project: Spark > Issue Type: Documentation > Components: Documentation > Affects Versions: 2.0.1 > Reporter: Noritaka Sekiyama > Assignee: Noritaka Sekiyama > Priority: Minor > > http://spark.apache.org/docs/latest/programming-guide.html > "By default, Spark creates one partition for each block of the file (blocks > being 64MB by default in HDFS)" > Currently default block size in HDFS is 128MB. > The default value has been already increased in Hadoop 2.2.0 (the oldest > supported version of Spark). https://issues.apache.org/jira/browse/HDFS-4053 > Since it looks confusing explanation, I'd like to fix the value from 64MB to > 128MB. -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org