Cassandra Cluster - Specific Node - specific table high Dropped Mutations

杀马特。学长 韩版系。学妹 提交于 2019-12-06 14:02:44

We faced this issue ourselves and we resolved this (as last resort) by removing the node from the cluster ( We belived there was some unknown hardware failure or memory leak of that sort )

We recommend you remove the node using nodetool removenode instead of nodetool decomission because we do not want to stream data from the failed node but instead from one of it's replica. ( This was a safe check and to avoid possibility of streaming corrupt data to other nodes. )

After we removed the node , the cluster health came back to normal and it was functioning normally.

1) You're on 3.10, you should strongly consider 3.11.2. A lot of critical bugs are fixed in 3.11.2

2) If you have one node that's misbehaving, and RF=3, then it's likely that you're treating that one node differently than the others. It may be that you're having your application only connect to one host and the cost of coordinating is overwhelming it, or you may have a disproportionate amount of data on it because of some misconfiguration (looks like you have RF=3 with 2 racks, so it's certainly possible that it's not quite distributed how you expect).

易学教程内所有资源均来自网络或用户发布的内容,如有违反法律规定的内容欢迎反馈
该文章没有解决你所遇到的问题?点击提问,说说你的问题,让更多的人一起探讨吧!