[rabbitmq-discuss] Failover in Clustering mode
charsyam at gmail.com
Mon Sep 3 14:56:38 BST 2012
Thank Francesco and Matthias.
Oh, I'm sorry for leaving stupid question.
Yes, I killed disc node(node1), and I tried to make a queue which has the
same name that node1 has.
and Matthias. your answer is very clear.
and I understood that clustering mode is different from mirror mode. I
thought it does something like mirror mode.
and, If I want to support Failover, I have to run rabbitmq as mirror mode.
2012/8/31 Matthias Radestock <matthias at rabbitmq.com>
> On 29/08/12 12:42, CharSyam wrote:
>> I made a Rabbitmq cluster with 3 nodes.
>> Originally, there are 1 disk node and 2 ram nodes.
>> It worked well.
>> but when I kill the process of node1.
>> the queue_declare command failed( queue name is "task_queue" )
>> and I couldn't make it again unless I recovred it.
> That is the expected behaviour when the queue is declared as 'durable'.
> While a queue's "home" node is down, clients connected to other nodes and
> attempting to re-declare the queue will get a 404-NOT_FOUND.
> If other nodes were able to re-declare such queues they would then
> conflict with the original queue when that gets recovered.
-------------- next part --------------
An HTML attachment was scrubbed...
More information about the rabbitmq-discuss