rabbitmq-server icon indicating copy to clipboard operation
rabbitmq-server copied to clipboard

A Ra cluster (Khepri or QQ) may not recover from a network partition

Open mkuratczyk opened this issue 11 months ago • 1 comments

Describe the bug

NOTE: this issue may be applicable with Mnesia as well (I haven't tried), but with Mnesia, network partition handling strategies "interfere" and potentially prevent/resolve this issue. Therefore, I focus on a khepri-enabled clusters here.

In a 3-node cluster with Khepri enabled, if the node with a Ra leader gets partitioned from the other two nodes, it may not rejoin the Ra cluster after the partition heals. It's sufficient to call net_kernel:connect_node('rabbit@another-node') or ra:members({rabbitmq_metadata, 'rabbit@<another-node>'}).) from the partitioned node, for all nodes to re-establish connections, rejoin the Ra clusters and sync the updates.

If I connect an AMQP client to a node partitioned like that, it's not able to perform any declarations, can't publish to queues etc. Attempting to do so may trigger reconnections but not quickly enough for the client to succeed (it's not exactly clear to me what triggers the reconnection - sometimes the cluster was sitting broken for 30 minutes).

Reproduction steps

  1. Deploy a 3-node RabbitMQ cluster (I'm using main from December 13th) and enable khepri_db
  2. Check which node is the leader for Khepri or a QQ
  3. Trigger a network partition so that the leader can't talk to the other 2 nodes (I'm using chaos-mesh on Kubernetes)
  4. As expected, the other two nodes elect a new leader and work correctly
  5. Resolve the network partition after 2 minutes
  6. Check rabbitmq-diagnostics metadata_store_status or rabbitmqctl list_queues leader (based on whether you are triggering this in a QQ Ra cluster or a Khepri Ra cluster) - you will see that the old leader still believes it is the leader and that the other two members are missing. The other two nodes report the correct leader and only miss the old leader's member. For example, after server-0 was partitioned, it reports:
Status of metadata store on node [email protected] ...
┌───────────────────────────────────────────┬────────────┬────────────┬────────────────┬──────────────┬──────────────┬──────────────┬────────────────┬──────┬─────────────────┐
│ Node Name                                 │ Raft State │ Membership │ Last Log Index │ Last Written │ Last Applied │ Commit Index │ Snapshot Index │ Term │ Machine Version │
├───────────────────────────────────────────┼────────────┼────────────┼────────────────┼──────────────┼──────────────┼──────────────┼────────────────┼──────┼─────────────────┤
│ [email protected] │ leader     │ voter      │ 959            │ 959          │ 959          │ 959          │ -1             │ 29   │ 1               │
└───────────────────────────────────────────┴────────────┴────────────┴────────────────┴──────────────┴──────────────┴──────────────┴────────────────┴──────┴─────────────────┘

The other two nodes report the new leader correctly:

Status of metadata store on node [email protected] ...
┌───────────────────────────────────────────┬────────────┬────────────┬────────────────┬──────────────┬──────────────┬──────────────┬────────────────┬──────┬─────────────────┐
│ Node Name                                 │ Raft State │ Membership │ Last Log Index │ Last Written │ Last Applied │ Commit Index │ Snapshot Index │ Term │ Machine Version │
├───────────────────────────────────────────┼────────────┼────────────┼────────────────┼──────────────┼──────────────┼──────────────┼────────────────┼──────┼─────────────────┤
│ [email protected] │ leader     │ voter      │ 960            │ 960          │ 960          │ 960          │ -1             │ 30   │ 1               │
├───────────────────────────────────────────┼────────────┼────────────┼────────────────┼──────────────┼──────────────┼──────────────┼────────────────┼──────┼─────────────────┤
│ [email protected] │ follower   │ voter      │ 960            │ 960          │ 960          │ 960          │ -1             │ 30   │ 1               │
└───────────────────────────────────────────┴────────────┴────────────┴────────────────┴──────────────┴──────────────┴──────────────┴────────────────┴──────┴─────────────────┘

Status of metadata store on node [email protected] ...
┌───────────────────────────────────────────┬────────────┬────────────┬────────────────┬──────────────┬──────────────┬──────────────┬────────────────┬──────┬─────────────────┐
│ Node Name                                 │ Raft State │ Membership │ Last Log Index │ Last Written │ Last Applied │ Commit Index │ Snapshot Index │ Term │ Machine Version │
├───────────────────────────────────────────┼────────────┼────────────┼────────────────┼──────────────┼──────────────┼──────────────┼────────────────┼──────┼─────────────────┤
│ [email protected] │ leader     │ voter      │ 960            │ 960          │ 960          │ 960          │ -1             │ 30   │ 1               │
├───────────────────────────────────────────┼────────────┼────────────┼────────────────┼──────────────┼──────────────┼──────────────┼────────────────┼──────┼─────────────────┤
│ [email protected] │ follower   │ voter      │ 960            │ 960          │ 960          │ 960          │ -1             │ 30   │ 1               │
└───────────────────────────────────────────┴────────────┴────────────┴────────────────┴──────────────┴──────────────┴──────────────┴────────────────┴──────┴─────────────────┘

Expected behavior

Most likely we need a component that will periodically check whether all expected nodes are present and if not, it'll try to re-establish these connections. Otherwise, at least in a fairly idle cluster, it may not happen.

Additional context

No response

mkuratczyk avatar Dec 13 '24 16:12 mkuratczyk

@mkuratczyk I personally see some parallels with https://github.com/rabbitmq/rabbitmq-server/pull/11407, https://github.com/rabbitmq/rabbitmq-server/pull/11363, except that this reconciliation process should not voluntarily stop after N runs.

michaelklishin avatar Dec 13 '24 19:12 michaelklishin

Raft ticks and hearbeat RPCs are there for this reason. There must be a code path in Ra where the tick timeout is not emitted or has no effects.

dumbbell avatar Aug 12 '25 13:08 dumbbell

I submitted a test case that attemps to reproduce the steps detailed in the description in #14384. So far, I couldn’t reproduce it.

dumbbell avatar Aug 14 '25 07:08 dumbbell

I can no longer reproduce this either (using chaos-mesh like originally, although in a different environment)

mkuratczyk avatar Aug 14 '25 07:08 mkuratczyk

Let’s close this issue for now. We couldn’t reproduce it and we now have a test case that tries to reproduce the steps above.

dumbbell avatar Aug 14 '25 12:08 dumbbell