You can configure mediation servers which can fail over SNMP/CLI requests, but failover for data collection for traps, syslog and informs, but this redundancy is not supported for performance monitoring. For redundancy in performance monitoring, you can run identical data collection profiles on multiple mediation servers, however, but it creates much more network traffic.
You must configure monitored devicesto forward traps to both mediation servers.
When you cluster mediation servers for traps / informs / syslog as a High Availability (HA) pair, Redcell designates one as the active server for HA services while the other serves as a standby server. The standby server does not forward data unless the active server becomes unavailable. The mediation server that starts first is Redcell’s active server. When you shut down the active server, or it stops responding, Redcell promotes the standby server to active. If a server process shuts down, it is no longer considered active. However, if the network connection is lost to the active server, but the server process continues to run, it is still considered active, and once the network connection is restored, Redcell Synergy considers both servers temporarily active.
Both servers forward data to the application server, and the application server eliminates duplicates. When Redcell Synergy detects that the connection to the disconnected mediation server has been restored, and that both servers are now active, the server with the lowest IP address remains active, and the other server becomes standby. The detection process takes approximately two minutes and then an additional period of seconds for the sorting of active / standby servers to occur. The standby should hold any data which had not been processed during switchover and that backlog should be processed when it becomes active.
See Independent and Clustered Mediation Agents for the alternative when mediation agents have to communicate with clustered application servers.
Best practice for changing properties in the following steps assume that you override the mentioned properties with properties pasted in owareapps/installprops/lib/installed.properties, rather altered than in the mentioned properties files.
Follow these steps to configure Mediation server clusters:
1. (For MySQL installations only) If the installation has not already done so, override the database host name originally specified in oware/lib/owdatabase.properties.
2. Set oware.config.server in owareapps/installprops/medserver/lib/installed.properties, setting it to the primary mediation host’s name (for example: aaron).
3. This property need to be enabled, only if the MBeans (Trap / Data Collection) needs to be setup in Forwarding/Standby Mode so that mediation servers work as peers and have listener failover capability.
An HA “cluster” consists of only two mediation servers. You can have more than two, but that just enables more than one standby server (something we do not test). If you set the HAproperty to false, all clustered mediation servers are active at the same time, and any number of mediation servers can be in the same cluster.
4. Mediation server peers use a multicast address for communication. If the server requires another mediation server peer setup, change this multicast address for that peer.
For an alternative to multicasting, see Disabling Multicast.
5. Uncomment the following sonicmq jms properties in /oware/lib/owjms.properties:
After making the change to the above properties,
you must re-source the
Also: You can safely ignore the error messages about not finding a JMS provider in the application server shell.
6. An example of the mediation cluster’s start command:
startmedagent -c PrimarymedServer (aaron) -a primaryappserver (laika) -p ClusterName (qacluster) -m address (184.108.40.206)
For an alternative startup, see Starting Clusters Durably. Remember, start the primary first. Wait until it has completely started before starting the secondary.
When setting up a failover mediation cluster in owareapps/installprops/medserver/lib/installed.properties, do not set up an application server cluster for the mediation servers. The config server for these mediation clusters is the same machine the individual mediation server runs on.