The following sections give you information and instructions on how to cluster the message broker profile of WSO2 EI.
The clustering deployment pattern
This pattern has two WSO2 EI nodes that include the Message Broker profile to serve service requests with high availability and scalability. The following image depicts the sample pattern of this clustering deployment scenario.
This pattern uses two nodes as well-known members. It is always recommended to have all nodes of the cluster as well-known members.
When configuring your WSO2 products for clustering using Hazelcast, you need to use a specific IP address in your configurations and not localhost.
- In this guide, the IP of the first Message Broker (MB) node is referred to as
https://xxx.xxx.xxx.xx1and the IP of the second MB node is referred to as
If you want to test out clustering in a development environment using the same server, you need to port offset one of your MB nodes.
Specify the port offset value in the
This is not recommended for production environments. Change all ports used in your configurations based on the offset value if you are setting a port offset for the specific node.Click here for more information on configuring the port offset.
When you run multiple products/clusters or multiple instances of the same product on the same server or virtual machines (VMs), change their default ports with an offset value to avoid port conflicts. An offset defines the number by which all ports in the runtime (e.g., HTTP(S) ports) are increased. For example, if the default HTTP port is 9763 and the offset is 1, the effective HTTP port will change to 9764. For each additional product instance, set the port offset to a unique value. The offset of the default ports is zero.
The port value will automatically increase as shown in the Port Value column in the following table, allowing five WSO2 product instances or servers to run on the same machine.
WSO2 product instance
WSO2 server 1
WSO2 server 2
WSO2 server 3
WSO2 server 4
WSO2 server 5
Creating the databases
All profiles of WSO2 EI uses a database to store information such as user management details and registry data. All nodes in the cluster must use one central database for config and governance registry mounts. You can create the following databases and associated datasources.
|JDBC user store and authorization manager.|
|Shared database for config and governance registry mounts in the product's nodes.|
|Local registry space in Node 1.|
|Local registry space in Node 2.|
In addition to the databases listed above, the MB profile of WSO2 EI requires the following broker-specific data:
|Stores instance data that are specific to the message broker profile.|
It is recommended to use an industry-standard RDBMS such as Oracle, PostgreSQL, MySQL, MS SQL, etc. for most enterprise testing and production environments. However, you can also use the embedded H2 database only for the
Follow the steps given below to create the databases required for the MB profile of WSO2 EI.
Download and install the MySQL Server.
Download the MySQL JDBC driver.
Download and unzip the WSO2 EI binary distribution. Be sure to select the correct product version.
Throughout this guide,
<EI_HOME>refers to the extracted directory of the WSO2 EI product distribution.
- Unzip the downloaded MySQL driver, and copy the MySQL JDBC driver JAR (
mysql-connector-java-x.x.xx-bin.jar) to the
<EI_HOME>/lib/directory of both the WSO2 EI nodes.
Add the following line to the
/etc/hostsfile to define the hostname for configuring permissions for the new database:
Do this step only if your database is not on your local machine and on a separate server.
- Execute the following command in a terminal/command window, where username is the username you want to use to access the databases:
mysql -u username -p
When prompted, specify the password to access the databases.
Create the databases using the following commands:
About using MySQL in different operating systems
For users of Microsoft Windows, when creating the database in MySQL, it is important to specify the character set as latin1. Failure to do this may result in an error (error code: 1709) when starting your cluster. This error occurs in certain versions of MySQL (5.6.x) and is related to the UTF-8 encoding. MySQL originally used the latin1 character set by default, which stored characters in a 2-byte sequence. However, in recent versions, MySQL defaults to UTF-8 to be friendlier to international users. Hence, you must use latin1 as the character set as indicated below in the database creation commands to avoid this problem. Note that this may result in issues with non-latin characters (like Hebrew, Japanese, etc.). The following is how your database creation command should look.
mysql> create database <DATABASE_NAME> character set latin1;
For users of other operating systems, the standard database creation commands will suffice. For these operating systems, the following is how your database creation command should look.
mysql> create database <DATABASE_NAME>;
Create and configure the
MB_DBdatabase, which is specific to the MB profile:
<EI_HOME>, username, and password are the same as those you specified in the previous steps.
Mounting the registry
Add the following configurations to the
<EI_HOME>/wso2/broker/conf/registry.xml file of each WSO2 EI node to configure the shared registry database and mounting details. This ensures that the shared registry for governance and configurations (i.e., the
REGISTRY_DB database) mounts on both WSO2 EI nodes.
Note the following when adding these configurations:
- The existing
wso2registrymust not be removed.
- The datasource you specify in the
<dbConfig name="sharedregistry">tag must match the JNDI Config name you specify in the
The registry mount path denotes the type of registry. For example, ”
/_system/config” refers to configuration Registry, and "
/_system/governance" refers to the governance registry.
<dbconfig>entry enables you to identify the datasource you configured in the
<EI_HOME>/wso2/broker/conf/datasources/master-datasources.xmlfile. The unique name "shared registry" refers to that datasource entry.
<remoteInstance>section refers to an external registry mount. Specify the read-only/read-write nature of this instance, caching configurations and the registry root location in this section.
Also, specify the cache ID in the
<remoteInstance>section. This enables caching to function properly in the clustered environment.
Cache ID is the same as the JDBC connection URL of the registry database. This value is the Cache ID of the remote instance. It should be in the format of
[email protected]$database_url, where
$database_usernameis the username of the remote instance database and
$database_urlis the remote instance database URL. This cache ID denotes the enabled cache. In this case, the database it should connect to is
REGISTRY_DB, which is the database shared across all the nodes. You can find that in the mounting configurations of the same datasource that is being used.
Define a unique name in the
<id>tag for each remote instance. This is then referred to from mount configurations. In the above example, the unique ID for the remote instance is "
Specify the actual mount path and target mount path in each of the mounting configurations. The target path can be any meaningful name. In this instance, it is "
Configuring a message broker profile node
Time synchronising the broker nodes
The RDBMS identifies the latest record based on timestamps so a newer timestamp always gets preference. Therefore, it is vital to synchronize time across all the MB nodes in the cluster to make them function properly. Given below are instructions for time synchronizing on Linux servers. Note that you need administrator permissions to do this.
- Install the ntpdate package by executing the following commands.
For RedHat Enterprise Linux, CentOS, and Oracle Linux:
For Debian and Ubuntu:
Execute the command given below to create a script named
This command will sync the time of the local machine with the time in the pool.ntp.org time server cluster. The pool.ntp.org project is a big virtual cluster of time servers providing reliable easy-to-use NTP services for millions of clients. It's the default "time server" for most of the major Linux distributions, and many networked appliances. The logs of the time synchronization will be routed to the
- Now, you need to schedule a cron job. Every computer calculates time locally according to a clock. This may cause a considerable time gap between servers in a clustered setup. Therefore, to make sure that the time on each of the servers are synchronized, we need to run the above script from time to time in every server machine. All servers will sync to the time received by "pool.ntp.org".
First, execute the following command:
Give the cron expression. For example, if you want time synchronization every minute, use the following cron expression:
Do the following configurations for all nodes of your cluster.
Configure the datasources to point to the
WSO2_USER_DBdatabases as follows in the
<EI_HOME>/wso2/broker/conf/datasources/master-datasources.xmlfile as follows:
Replace the username, password, and database URL of your MySQL environment accordingly.
For MB node 1, configure the datasources to point to the
WSO2_USER_DBdatabases as follows:
For node 2, configure the datasources to point to the
WSO2_USER_DBdatabases as shown below. Change the username, password, and database URL as needed for your environment.
Remove or comment out the default H2-based
WSO2_CARBON_DBdatasource configurations from the
Uncomment or add the following
WSO2_MB_STORE_DBconfiguration in the
<EI_HOME>/wso2/broker/conf/datasources/master-datasources.xmlfile based on your DBMS type.
Update the JDBC URL to correctly point to your database and enter the username and password for your database user with the proper permissions.
If your MySQL username or password is not
root, make sure to update the
Add the following configuration in the
<EI_HOME>/wso2/broker/conf/user-mgt.xmlfile of both nodes to configure the user stores.
Enter the datasource information for the user store that you configured in the
<EI_HOME>/wso2/broker/conf/datasources/master-datasources.xmlfile. You can change the admin username and password as well. However, you should do this before starting the server.
Uncomment or add the following configuration in the
<EI_HOME>/wso2/broker/conf/broker.xmlfile. Enter the message store based on your database type and Andes context store.
Add the following thrift-related configurations in the
This file is the root configuration file of brokering. Do the changes you do to this file in all the broker profile nodes. Configure the
thriftServerHostvalue to point to the IP address of the MB profile node.
This configuration is related to brokering thrift communications.
In a clustered deployment, an ID is assigned to each brokering node via the cluster node identifier. This element can be used to override the cluster node identifier for this brokering node. If the value for this element is left as
default, the default node ID is generated using the IP and a universally unique identifier (UUID). The node ID of each member in a cluster must be unique.
This is a sub-element of the
<coordination>tag. The MB profile uses Apache Thrift for communications relating to message delivery. Therefore, an Apache Thrift server is started in each brokering node in a clustered deployment. This element should point to the IP address of the Apache Thrift server. This should point to the IP address of the brokering node that hosts the thrift server. The default value for this is
localhost. For example, if you are configuring a brokering node hosted in 192.168.0.102 as the thrift server, this value should be 192.168.0.102.
This is another sub-element of the
<coordination>tag. This should point to the port of the thrift server in the MB profile.
What is the effective thrift server port ?The
broker.xmlfile indicates 7611 as the default thrift server port. However, the default thrift server port that is effective in the MB profile of WSO2 EI is 7614 (7611 + 3). This is because, by default, the server port of the MB profile of WSO2 EI is offset by 3 in the
xmlfile. It is important to note that the effective thrift server port depends upon the default server port offset in the
carbon.xmlfile as well as the
thriftServerPortvalue in the
broker.xmlfile. Be sure to keep the correct thrift server port open in your system to avoid errors.
It is recommended to use the same thrift server port for all broker nodes in your cluster.
This is used to handle half-open TCP connections between the broker nodes in a cluster. In such situations, the socket may need to have a timeout value to invalidate the connection (in milliseconds). A timeout of zero is interpreted as an infinite timeout. Be sure to set this value to 180000 milliseconds as shown below.
<EI_HOME>/wso2/broker/conf/axis2/axis2.xmlfile as follows to set up the cluster configurations.
- Enable clustering for this node as follows:
<clustering class="org.wso2.carbon.core.clustering.hazelcast.HazelcastClusteringAgent" enable="true">
- Set the membership scheme to "wka" to enable the well-known address registration method. (This node sends cluster initiation messages to the WKA members):
- Specify the name of the cluster this node will join as follows:
Specify the hosts to communicate cluster messages as follows:
Specify the port to communicate cluster messages as follows:
This port number is not affected by the port offset value specified in the
xmlfile. If this port number is already assigned to another server, the clustering framework automatically increments this port number. However, if there are two servers running on the same machine, ensure that a unique port is set for each server.
Specify the well-known members as shown below. The port value for the WKA node must be the same value as it's localMemberPort (in this case it is 4100.
You can also use IP address ranges for the hostname (e.g., 192.168.1.2-10). However, you can define a range only for the last portion of the IP address. Smaller the range, the faster the time it takes to discover members since each node has to scan a lesser number of potential members. The best practice is to add all the members (including itself) in all the nodes to avoid any conflicts in configurations.
- Enable clustering for this node as follows:
Enable heartbeat messaging for each of the broker nodes by following the steps given below. This is required for handling TCP connections between the broker nodes and client systems (publishers and subscribers).
What is heartbeat messaging?Click here for more information.
When a client is connected to the broker, both the broker and the client should be able to detect problem situations where the TCP connection is half open or where the connecting client/broker is unresponsive. This can be achieved by enabling heartbeat messaging between the broker and the client.
The heartbeat messaging configuration allows both the broker and the client to verify whether the connection is inactive or whether the connecting system (broker or client) is inactive by periodically sending messages to each other.
For example, you can set the heartbeat delay time to 30 seconds. This means that if the broker (or the client) does not receive the expected response from the other system within 30 seconds, it will send a heartbeat message to check if the connection is inactive or if the other system is inactive. If the connection is inactive or is half-open, there will be a clear indication of connection failure. However, if the connection is active but the other system is unresponsive, the broker or the client will continue to send heartbeat messages (every 30 seconds). By default, the broker and clients are configured to terminate (timeout) the connection after sending two heartbeat messages; however, this time out configuration can be changed.
Note: In a scenario where the subscriber client is sending the heartbeat message to the broker node, if the connection is found to be broken or if the broker node is inactive, the connection will failover to another broker node in the cluster.
- Open the
qpid-config.xmlfile that is stored in the
Set the heartbeat delay as shown below. The recommended heartbeat delay is 30 seconds.
The elements in the above configuration are explained below.
delay The <delay> element specifies the time interval between heartbeat messages. This is the time that the broker or client will wait to receive a response from the other party. If the response is not received within this time, a heartbeat message is triggered. timeoutFactor The number of heartbeat messages the broker will send to the client before terminating the connection. That is, if the timeoutFactor is 2, the broker will send heartbeat messages every 30 seconds twice, and if a response is not received from the client, the connection will be terminated.
- Open the
Optional: Configuring Hazelcast properties
WSO2 products use Hazelcast as its default clustering engine. You can configure the hazelcast properties for the product nodes by following the steps given below.
Add the following property configurations to the
hazelcast.propertiesfile stored in the
The above configurations are explained below.
- Hazelcast shutdown hook: This configuration disables the shutdown hook in hazelcast, which ensures that the hazelcast instance shuts down gracefully whenever the product node shuts down. If the hazelcast shutdown hook is enabled (which is the default behavior of a product), you will see errors such as "Hazelcast instance is not active!" at the time of shutting down the product node: This is because the hazelcast instance shuts down too early when the shutdown hook is enabled.
- Hazelcast logging type: This configuration sets the hazelcast logging type to log4j, which allows hazelcast logs to be written to the
If you have enabled log4j for hazelcast logging as explained above, be sure to enter the configuration shown below in the
log4j.propertiesfile (stored in the
<EI_HOME>/wso2/broker/conf/directory). This can be used to configure the log level for hazelcast logging. For a clustered production environment, it is recommended to use INFO as the log level as shown below.
Optional: Additional configurations
Cluster coordination and event synchronization
The cluster coordination process controls how the nodes in an MB cluster are managed. By default, the MB uses an RDBMS (the MB database that is connected to nodes in the cluster) for cluster coordination as well as cluster event synchronization. Note that this is also the same RDBMS that is used for message persistence in the MB, and thereby cluster coordination, cluster event synchronization, and message persistence works on the same network in your cluster.
The following settings in the
broker.xml file (stored in the
<EI_HOME>/wso2/broker/conf/ directory) enables and configures RDBMS cluster coordination and event synchronization. Note that it is not recommended to disable RDBMS cluster coordination and event synchronization as it can sometimes cause network partitioning in your cluster.
The clustering mechanism of the MB profile depends on its Hazelcast engine. Therefore, Hazelcast clustering should always be enabled in the MB profile for RDBMS-based cluster coordination to work.
See the parameter descriptions given below.
Cluster coordination parameters:
This property is set to "true" by default, which indicates that RDBMS-based cluster coordination is enabled:It is not recommended to disable this setting. If you disable this setting, the hazelcast engine will handle cluster coordinatin without using the RDBMS network, which can cause network partitioning. Find out more about how to handle network partitioning in hazelcast cluster coordination.
The time interval for heartbeat messaging, which is used to check the availability of the nodes in the cluster. Declared in milliseconds.
The time the system should wait before informing other nodes about a new coordinator. Declared in milliseconds. This value should be larger than the database read time (including network latency) and should be less than the
Time interval used to poll the database for membership related events. Declared in milliseconds.
Event synchronization parameters:
This property is set to "true" by default, which indicates that RDBMS-based cluster event synchronization is enabled:It is not recommended to disable this setting.
Specifies the interval at which the cluster events will be read from the database. Declared in milliseconds. Setting this to a very low value could decrease server performance where as setting this to a large value could increase the time taken for a cluster event to be synchronized in all the nodes in a cluster.
Handling network partitioning in Hazelcast-based clustering
Network partitioning can happen in the Message Broker cluster if you have disabled RDBMS-based cluster coordination. That is, when RDBMS-based clustering is disabled, the hazelcast engine in the MB uses two separate networks: The network with the RDBMS will persist messages, subscriptions and queues etc., and the cluster network communicates information about subscriptions, queue additions, and to decide on the coordinator of the cluster.
For example, consider a cluster of three MB nodes that uses two separate networks for cluster coordination and message persistence. If one of the nodes in the cluster disconnects from the network used for cluster coordination, that node will separate from the cluster and function as a separate node/cluster (separate partition). That is, the three-node cluster will now be working as two separate clusters. However, the disconnected node will still be up and running, and message persistence will continue uninterrupted through the message persistence network. This will cause inconsistencies in the system because message persistence and cluster coordination will not be synchronized. In such a situation, it is necessary to stop the disconnected node from accepting messages.
The above situation can be prevented in the Message Broker profile by configuring the minimum node count of the cluster. For example, if the cluster size is 5 and we have configured the minimum node count to 3 nodes, during a network partition (or when nodes crash or shut down) any partition that has 3 or more nodes will keep functioning, while the other partitions will stop processing messages.
Follow the steps given below to enable this feature.
<EI_HOME>/wso2/broker/repository/conf/broker.xmlfile and set the
<networkPartitionsDetection>element shown below to
enabled="true". Note that this configuration is disabled by default as shown below.
- Update the
<minimumClusterSize>property in the above configuration to specify the minimum node count the cluster should maintain in order to operate. Note that if this value should be at least two for cluster coordination to work. If the number of nodes in the cluster becomes less than the configured value, the cluster will not accept any incoming traffic. That is, all subscriptions will be disconnected.
Testing the cluster
Follow the steps below to test the cluster.
Start the MB profile nodes. See the instructions in Running the Product.
Check for ‘
member joined’ log messages in all consoles.
Additional information on logs and new nodes
When you terminate one node, all nodes identify that the node has left the cluster. The same applies when a new node joins the cluster. If you want to add another new node, copy existing node without any changes if you are running it on a new server (such as xxx.xxx.xxx.184). If you intend to use the new node on a server where another WSO2 product is running, use a copy of node and change the port offset accordingly in the
<EI_HOME>/wso2/broker/conf/carbon.xmlfile. You also have to change
<EI_HOME>/wso2/broker/conf/axis2/axis2.xml file if that product has clustering enabled. Also, map all hostnames to the relevant IP addresses when creating a new node. The log messages indicate if the new node joins the cluster.
- Access the Management Console through the LB using the following URL:
- Sign into the management console using
adminas the username and password.
- Click Node List.
You see that the 2 nodes are up and running successfully. The message store health needs to be Healthy and only one node can be the Coordinator Node.
If you want to performance tune the WSO2 Message Broker (MB) profile, see Tuning the Message Broker Profile .