Hi, Well, I guess knowing the disk behaviour would be useful to understand if it is really filling up and why.
- What is the disk capacity? - Does it actually fill up? - If it is filling up, it might mean that all your nodes are not running with enough available space and that any node can go down due to an heavy compaction at any time. - It might also come from an imbalance on the ownership (use 'nodetool status *<ks>*' to know that for sure) or in the load ('df -h' or 'du -sh /path_to_cassandra/' on all nodes and compare the disk in use. - Last idea that came to mind is that the streaming throughput might be too high or the compaction throughput too low. So in this last case, if your data is growing way bigger than the dataset original size, it is probably because compaction can't keep up with what is being streamed. As I believe that a joining node has an unthrottled compaction: ( https://github.com/apache/cassandra/blob/trunk/src/java/org/apache/cassandra/db/compaction/CompactionManager.java#L149-L150), it looks like a good idea to restrict the streaming throughput. C*heers, ----------------------- Alain Rodriguez - @arodream - al...@thelastpickle.com France The Last Pickle - Apache Cassandra Consulting http://www.thelastpickle.com 2017-02-04 11:11 GMT+01:00 techpyaasa . <techpya...@gmail.com>: > Cluster Information: > Name: xxxx Cluster > Snitch: org.apache.cassandra.locator.DynamicEndpointSnitch > Partitioner: org.apache.cassandra.dht.Murmur3Partitioner > Schema versions: > 397560b8-7245-3903-8828-60a97e5be4aa: [xxx.xxx.xxx.75, xxx.xxx.xxx.134, > xxx.xxx.xxx.192, xxx.xxx.xxx.132, xxx.xxx.xxx.133, xxx.xxx.xxx.115, > xxx.xxx.xxx.78, xxx.xxx.xxx.123, xxx.xxx.xxx.70, xxx.xxx.xxx.167, > xxx.xxx.xxx.168, xxx.xxx.xxx.169, xxx.xxx.xxx.146, xxx.xxx.xxx.145, > xxx.xxx.xxx.144, xxx.xxx.xxx.143, xxx.xxx.xxx.140, xxx.xxx.xxx.139, > xxx.xxx.xxx.126, xxx.xxx.xxx.136, xxx.xxx.xxx.135, xxx.xxx.xxx.191, > xxx.xxx.xxx.133, xxx.xxx.xxx.79, xxx.xxx.xxx.131, xxx.xxx.xxx.77] > > ReleaseVersion: 2.0.17 > --------------------------------------------------------------- > Note: Ownership information does not include topology; for complete > information, specify a keyspace > Datacenter: DC1 > =============== > Status=Up/Down > |/ State=Normal/Leaving/Joining/Moving > -- Address Load Tokens Owns Host ID > Rack > UN xxx.xxx.xxx.168 847.62 GB 256 4.1% > 2302491a-a8b5-4aa6-bda7-f1544064c4e3 > GRP3 > UN xxx.xxx.xxx.169 819.64 GB 256 4.2% > d5e5bc3d-38de-4043-abca-08ac09f29a46 > GRP1 > UN xxx.xxx.xxx.75 874.69 GB 256 4.1% > fdd32c67-3cea-4174-b59b-c1ea14e1a334 > GRP1 > UN xxx.xxx.xxx.78 850.07 GB 256 4.0% > a8332f22-a75f-4d7c-8b71-7284f6fe208f > GRP3 > UN xxx.xxx.xxx.126 836.88 GB 256 4.0% > 71be90d8-97db-4155-b4fc-da59d78331ef > GRP1 > UN xxx.xxx.xxx.191 751.08 GB 256 4.1% > a9023df8-a8b3-484b-a03d-0fdea35007bd > GRP3 > UN xxx.xxx.xxx.192 888.03 GB 256 3.8% > f4ad42d5-cee0-4d3e-a4f1-7cdeb5d7390a > GRP2 > UN xxx.xxx.xxx.132 688.86 GB 256 3.8% > 6a465101-29e7-4792-8269-851200a70023 > GRP2 > UN xxx.xxx.xxx.133 855.66 GB 256 4.0% > 751ce15a-10f1-44cf-9357-04da7e21b511 > GRP2 > UN xxx.xxx.xxx.134 869.32 GB 256 3.7% > bdd166fd-95a7-4119-bbae-f05fe26ddb01 > GRP3 > UN xxx.xxx.xxx.70 792.15 GB 256 4.2% > 2b6b642d-6842-47d4-bdc1-95226fd2b85d > GRP1 > UN xxx.xxx.xxx.167 732.82 GB 256 4.0% > 45f6684f-d6a0-4cba-875c-9db459646545 > GRP2 > Datacenter: DC2 > =============== > Status=Up/Down > |/ State=Normal/Leaving/Joining/Moving > -- Address Load Tokens Owns Host ID > Rack > UN xxx.xxx.xxx.136 762.85 GB 256 3.9% > ebc67006-80e6-40de-95a9-79b90b254750 > GRP3 > UN xxx.xxx.xxx.139 807.68 GB 256 4.5% > e27ca655-3186-417f-a927-cab63dc34248 > GRP3 > UN xxx.xxx.xxx.140 771.94 GB 256 4.1% > 26531fe3-7f2c-4ce5-a41b-5c79c5976141 > GRP1 > UN xxx.xxx.xxx.77 505.54 GB 256 3.5% > d1ad7194-d7fb-47cf-92d1-4206ff12f8aa > GRP1 > UN xxx.xxx.xxx.143 900.14 GB 256 4.1% > 74e1009c-0506-4d7a-b517-d37182385a21 > GRP2 > UJ xxx.xxx.xxx.79 636.08 GB 256 ? > 91b64758-67c2-48e7-86eb-43f7509c2287 > GRP3 > UN xxx.xxx.xxx.131 788 GB 256 4.0% > 5b27a680-d7c0-4ead-85cc-c295b83eda5b > GRP2 > UN xxx.xxx.xxx.133 898.27 GB 256 3.8% > 5b24f211-678e-4614-bd59-8ea13aa2397c > GRP1 > UN xxx.xxx.xxx.135 868.14 GB 256 4.1% > 8c2b5d1c-e43e-41f4-b21e-58cb525cfbab > GRP2 > UN xxx.xxx.xxx.123 848.86 GB 256 4.0% > 87cfff8f-1cfc-44c5-b608-5b40d6894182 > GRP3 > UN xxx.xxx.xxx.144 830.99 GB 256 3.6% > 31b8cf4b-dd08-4ee6-8c25-90ad6dadbdc4 > GRP3 > UN xxx.xxx.xxx.145 832.22 GB 256 4.3% > dd8c97df-7ec9-436b-8b29-4c25a8a89184 > GRP1 > UN xxx.xxx.xxx.146 830.02 GB 256 4.2% > 88b52574-8569-4d58-ba43-8fe1c742eea4 > GRP2 > UN xxx.xxx.xxx.115 878.5 GB 256 3.9% > 20817b9e-b761-437e-aa7b-49e90483c69f > GRP1 > > > > > Total keyspaces 5 with 'class': 'NetworkTopologyStrategy', and > replication with 'DC2': '3', 'DC1': '3' > > On Sat, Feb 4, 2017 at 3:22 PM, Alexander Dejanovski < > a...@thelastpickle.com> wrote: > >> Hi, >> >> could you share with us the following informations ? >> >> - "nodetool status" output >> - Keyspace definitions (we need to check the replication strategy you're >> using on all keyspaces) >> - Specifics about what you're calling "groups" in a DC. Are these racks ? >> >> Thanks >> >> On Sat, Feb 4, 2017 at 10:41 AM laxmikanth sadula < >> laxmikanth...@gmail.com> wrote: >> >>> Yes .. same number of tokens... >>> 256 >>> >>> On Sat, Feb 4, 2017 at 11:56 AM, Jonathan Haddad <j...@jonhaddad.com> >>> wrote: >>> >>> Are you using the same number of tokens on the new node as the old ones? >>> >>> On Fri, Feb 3, 2017 at 8:31 PM techpyaasa . <techpya...@gmail.com> >>> wrote: >>> >>> Hi, >>> >>> We are using c* 2.0.17 , 2 DCs , RF=3. >>> >>> When I try to add new node to one group in a DC , I got disk full. Can >>> someone please tell what is the best way to resolve this? >>> >>> Run compaction for nodes in that group(to which I'm going to add new >>> node, as data streams to new nodes from nodes of group to which it is added) >>> >>> OR >>> >>> Boootstrap/add 2(multiple nodes) at a time? >>> >>> >>> Please suggest better way to fix this. >>> >>> Thanks in advance >>> >>> Techpyaasa >>> >>> >>> >>> >>> -- >>> Regards, >>> Laxmikanth >>> 99621 38051 >>> >>> -- >> ----------------- >> Alexander Dejanovski >> France >> @alexanderdeja >> >> Consultant >> Apache Cassandra Consulting >> http://www.thelastpickle.com >> > >