> Some reason with in less than an hour cassandra node is opening 32768 files > and cassandra is not responding after that. Are you using Levelled Compaction ? Is so what value did you set for min_sstable_size ? The default has changed from 5 to 160.
Increasing the file handles is the right thing to do but 32K files is a lot. Cheers ----------------- Aaron Morton New Zealand @aaronmorton Co-Founder & Principal Consultant Apache Cassandra Consulting http://www.thelastpickle.com On 8/11/2013, at 8:09 am, Arindam Barua <aba...@247-inc.com> wrote: > > I see 100 000 recommended in the Datastax documentation for thenofile limit > since Cassandra 1.2 : > > http://www.datastax.com/documentation/cassandra/2.0/webhelp/cassandra/install/installRecommendSettings.html > > -Arindam > > From: Pieter Callewaert [mailto:pieter.callewa...@be-mobile.be] > Sent: Thursday, November 07, 2013 4:22 AM > To: user@cassandra.apache.org > Subject: RE: Getting into Too many open files issues > > Hi Murthy, > > 32768 is a bit low (I know datastax docs recommend this). But our production > env is now running on 1kk, or you can even put it on unlimited. > > Pieter > > From: Murthy Chelankuri [mailto:kmurt...@gmail.com] > Sent: donderdag 7 november 2013 12:46 > To: user@cassandra.apache.org > Subject: Re: Getting into Too many open files issues > > Thanks Pieter for giving quick reply. > > I have downloaded the tar ball. And have changed the limits.conf as per the > documentation like below. > > * soft nofile 32768 > * hard nofile 32768 > root soft nofile 32768 > root hard nofile 32768 > * soft memlock unlimited > * hard memlock unlimited > root soft memlock unlimited > root hard memlock unlimited > * soft as unlimited > * hard as unlimited > root soft as unlimited > root hard as unlimited > > root soft/hard nproc 32000 > > > Some reason with in less than an hour cassandra node is opening 32768 files > and cassandra is not responding after that. > > It is still not clear why cassadra is opening that many files and not closing > properly ( does the laest cassandra 2.0.1 version have some bugs ). > > what i have been experimenting is 300 writes per sec and 500 reads per sec. > > And i have using 2 node cluster with 8 core cpu and 32GB RAM ( Virtuval > Machines) > > > Do we need to increase the nofile limts to more than 32768 ? > > > > > > > > > > > > > > > On Thu, Nov 7, 2013 at 4:55 PM, Pieter Callewaert > <pieter.callewa...@be-mobile.be> wrote: > Hi Murthy, > > Did you do a package install (.deb?) or you downloaded the tar? > If the latest, you have to adjust the limits.conf file > (/etc/security/limits.conf) to raise the nofile (number of files open) for > the cassandra user. > > If you are using the .deb package, the limit is already raised to 100 000 > files. (can be found in /etc/init.d/cassandra, FD_LIMIT). > However, with the 2.0.x I had to raise it to 1 000 000 because 100 000 was > too low. > > Kind regards, > Pieter Callewaert > > From: Murthy Chelankuri [mailto:kmurt...@gmail.com] > Sent: donderdag 7 november 2013 12:15 > To: user@cassandra.apache.org > Subject: Getting into Too many open files issues > > I have experimenting cassandra latest version for storing the huge the in our > application. > > Write are doing good. but when comes to reads i have obsereved that cassandra > is getting into too many open files issues. When i check the logs its not > able to open the cassandra data files any more before of the file descriptors > limits. > > Can some one suggest me what i am going wrong what could be issues which > causing the read operating leads to Too many open files issue.