Assume your block size is 128MB.

Thanks
Best Regards

On Mon, May 4, 2015 at 2:38 PM, ayan guha <guha.a...@gmail.com> wrote:

> Hi
>
> How do you figure out 500gig~3900 partitions? I am trying to do the math.
> If I assume 64mb block size then 1G~16 blocks and 500g~8000 blocks. If we
> assume split and block sizes are same, shouldn't we end up with 8k
> partitions?
> On 4 May 2015 17:49, "Akhil Das" <ak...@sigmoidanalytics.com> wrote:
>
>> 500GB of data will have nearly 3900 partitions and if you can have nearly
>> that many number of cores and around 500GB of memory then things will be
>> lightening fast. :)
>>
>> Thanks
>> Best Regards
>>
>> On Sun, May 3, 2015 at 12:49 PM, sherine ahmed <
>> sherine.sha...@hotmail.com> wrote:
>>
>>> I need to use spark to upload a 500 GB data from hadoop on standalone
>>> mode
>>> cluster what are the minimum hardware requirements if it's known that it
>>> will be used for advanced analysis (social network analysis)?
>>>
>>>
>>>
>>> --
>>> View this message in context:
>>> http://apache-spark-user-list.1001560.n3.nabble.com/Hardware-requirements-tp22744.html
>>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>>
>>> ---------------------------------------------------------------------
>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>>> For additional commands, e-mail: user-h...@spark.apache.org
>>>
>>>
>>

Reply via email to