finish elementary school first. (plus, minus operations at least)

On Thu, Jan 10, 2013 at 7:23 PM, Panshul Whisper <ouchwhis...@gmail.com>wrote:

> Thank you for the response.
>
> Actually it is not a single file, I have JSON files that amount to 115 GB,
> these JSON files need to be processed and loaded into a Hbase data tables
> on the same cluster for later processing. Not considering the disk space
> required for the Hbase storage, If I reduce the replication to 3, how much
> more HDFS space will I require?
>
> Thank you,
>
>
> On Fri, Jan 11, 2013 at 4:16 AM, Ravi Mutyala <r...@hortonworks.com>wrote:
>
>> If the file is a txt file, you could get a good compression ratio.
>> Changing the replication to 3 and the file will fit. But not sure what your
>> usecase is what you want to achieve by putting this data there. Any
>> transformation on this data and you would need more space to save the
>> transformed data.
>>
>> If you have 5 nodes and they are not virtual machines, you should
>> consider adding more harddisks to your cluster.
>>
>>
>> On Thu, Jan 10, 2013 at 9:02 PM, Panshul Whisper 
>> <ouchwhis...@gmail.com>wrote:
>>
>>> Hello,
>>>
>>> I have a hadoop cluster of 5 nodes with a total of available HDFS space
>>> 130 GB with replication set to 5.
>>> I have a file of 115 GB, which needs to be copied to the HDFS and
>>> processed.
>>> Do I need to have anymore HDFS space for performing all processing
>>> without running into any problems? or is this space sufficient?
>>>
>>> --
>>> Regards,
>>> Ouch Whisper
>>> 010101010101
>>>
>>
>>
>
>
> --
> Regards,
> Ouch Whisper
> 010101010101
>

Reply via email to