Issues with large data nodes would be -

*         Nodetool repair will be impossible to run

*         Your read i/o will suffer since you will almost always go to disk 
(each read will take 3 IOPS worst case)

*         Boot-straping the node in case of failures will take days/weeks



From: Pruner, Anne (Anne) [mailto:pru...@avaya.com]
Sent: 25 July 2013 10:45
To: user@cassandra.apache.org
Subject: RE: maximum storage per node

We're storing fairly large files (about 1MB apiece) for a few months and then 
deleting the oldest to get more space to add new ones.  We have large 
requirements (maybe up to 100 TB), so having a 1TB limit would be unworkable.

What is the reason for the limit?  Does something fail after that?

If there are hardware issues, what's recommended?

BTW, we're using Cassandra 1.2

Anne

From: cem [mailto:cayiro...@gmail.com]
Sent: Thursday, July 25, 2013 11:41 AM
To: user@cassandra.apache.org<mailto:user@cassandra.apache.org>
Subject: Re: maximum storage per node

Between 500GB - 1TB is recommended.

But it depends also your hardware, traffic characteristics and requirements. 
Can you give some details on that?

Best Regards,
Cem

On Thu, Jul 25, 2013 at 5:35 PM, Pruner, Anne (Anne) 
<pru...@avaya.com<mailto:pru...@avaya.com>> wrote:
Does anyone have opinions on the maximum amount of data reasonable to store on 
one Cassandra node?  If there are limitations, what are the reasons for it?

Thanks,
Anne

Reply via email to