Two of three nodes failed in a pool overnight with messages like: Node 123.123.123.123 has failed - Timeout while establishing connection (the machine may be down, or the network congested; increasing the 'max_connect_time' on the pool's 'Connection Management' page may help) When I came in this morning, I was able to verify the ip/port was available and functioning on both nodes, but the pool still had them marked as failed. I stopped the virtual server and started it back up and the pool connected to all three nodes and has been happy for the last hour. It seems like it just didn't try to reconnect to those nodes. I'm new to Stingray load balancers, having only worked with F5 in the past. Is this normal? Do I really need to manually intervene to fix this? Am I missing some obvious config? One attached screenshot is the health monitor config (listed in the catalog as Connect, I'm not sure if that is a standard monitor or was created by my predecessor). The pool also has passive monitoring turned on. The second screen shot shows the connection management settings: Is something in those configs causing the LB to give up on my nodes after some outage and require me to manually intervene?
... View more