Noritaka Sekiyama created SPARK-18432:
-----------------------------------------

             Summary: Fix HDFS block size in programming guide
                 Key: SPARK-18432
                 URL: https://issues.apache.org/jira/browse/SPARK-18432
             Project: Spark
          Issue Type: Documentation
          Components: Documentation
    Affects Versions: 2.0.1
            Reporter: Noritaka Sekiyama
            Priority: Minor


http://spark.apache.org/docs/latest/programming-guide.html
"By default, Spark creates one partition for each block of the file (blocks 
being 64MB by default in HDFS)"

Currently default block size in HDFS is 128MB.
The default value has been already increased in Hadoop 2.2.0 (the oldest 
supported version of Spark). https://issues.apache.org/jira/browse/HDFS-4053

Since it looks confusing explanation, I'd like to fix the value from 64MB to 
128MB.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to