Cluster Status State Unknown 127.0.0.1

Hi,

We are running a 3 node eventstore cluster in docker bridge mode.

I was in the process of upgrading a slave from 4.0.3 to 4.1.2.

Unfortunately these settings where used when creating the container. (External IP address should have been used)

EVENTSTORE_INT_IP_ADVERTISE_AS: “127.0.0.1”

EVENTSTORE_EXT_IP_ADVERTISE_AS: “127.0.0.1”

This resulted in a 4 node cluster!

From the UI:

Unknown
Alive
2019-04-3 22:28:30
L1773959 / W 1796957 / C 1796957E3 @ 1079524 : { 8fefd5fd-e943-4e94-886e-1ee0ee05e590 }
Internal : 127.0.0.1:1112
External : 127.0.0.1:1113
Internal : 127.0.0.1:2112
External : 127.0.0.1:2113

Master

Slave

Slave

How can I get rid of this Unknown but Alive node 127.0.0.1 state?

Thank you,

Bjarte Brandt

Hi Again,

So I found the solution in this thread “Ghost “manager” nodes showing up in cluster overview”.

From Greg: Nope they timeout after a while. You will notice they will show up as dead very quickly (depends on your gossip settings).

Just checked. Node is gone! Great!

Thanks,

Bjarte

onsdag 3. april 2019 22.32.56 UTC+2 skrev Bjarte Brandt følgende: