The ha-singleton-service
quickstart demonstrates how to deploy a cluster-wide singleton MSC service.
The ha-singleton-services
quickstart demonstrates two patterns, or ways, to deploy a cluster-wide singleton MSC service.
-
The first example, located in the
primary-only/
directory of the quickstart, demonstrates a singleton service and a querying service deployed on all nodes that regularly queries for the value provided by the singleton service. -
The second example, located in the
with-backups/
directory of the quickstart, demonstrates a singleton service that is installed with a backup service. The backup services are running on all nodes that were not elected to be running the singleton service itself.
Singleton service’s getValue()
always returns the value of the primary node unless a backup service is installed.
If no backup service is installed, a default backup service is used whose getValue()
returns the service value of the primary node.
Should a backup service be installed then the getValue()
is delegated to the primary or backup service depending on the state of the local node.
Make sure you inspect the activate()
method of the ServiceActivator
class for each example. Although the default election policy is used to build the singleton services for each of these examples, scripts and instructions are provided later in this document to demonstrate how to configure other election policies.
These examples are built and packaged as JAR archives.
For more information about clustered singleton services, see HA Singleton Service in the {LinkDevelopmentGuide}[{DevelopmentBookName}] for {DocInfoProductName} located on the Red Hat Customer Portal.
While you can run this example starting only one instance of the server, if you want to see the singleton behavior, you must start at least two instances of the server. Copy the entire {productName} directory to a new location to use for the second cluster member.
Note
|
You must start the server using the HA profile or the singleton service will not start correctly._ |
Start the two {productName} servers with the HA profile, passing a unique node ID. These logical node names are used in the log to identify which node is elected. If you are running the servers on the same host, you must also pass a socket binding port offset on the command line to start the second server.
$ {jbossHomeName}_1/bin/standalone.sh -c standalone-ha.xml -Djboss.node.name=node1
$ {jbossHomeName}_2/bin/standalone.sh -c standalone-ha.xml -Djboss.node.name=node2 -Djboss.socket.binding.port-offset=100
Note
|
For Windows, use the {jbossHomeName}_1\bin\standalone.bat and {jbossHomeName}_2\bin\standalone.bat scripts.
|
This example is not limited to two servers. Additional servers can be started by specifying a unique node name and port offset for each one.
This example demonstrates a singleton service and a querying service that regularly queries for the value that the singleton service provides.
-
Start the {productName} servers as described in the above section.
-
Open a terminal and navigate to the
primary-only/
directory located in the root directory of this quickstart. -
Use the following command to clean up any previously built artifacts, and to build and deploy the JAR archive.
$ mvn clean install wildfly:deploy
-
Investigate the primary-only Console Output for Server 1. Verify that the
target/ha-singleton-service-primary-only.jar
archive is deployed tonode1
, which is the first server started without port offset, by checking the server log.INFO [org.jboss.as.server.deployment] (MSC service thread 1-7) WFLYSRV0027: Starting deployment of "ha-singleton-service-primary-only.jar" (runtime-name: "ha-singleton-service-primary-only.jar") INFO [org.jboss.as.quickstarts.ha.singleton.service.primary.ServiceActivator] (MSC service thread 1-5) Singleton and querying services activated. INFO [org.jboss.as.quickstarts.ha.singleton.service.primary.QueryingService] (MSC service thread 1-3) Querying service is starting. ... INFO [org.wildfly.clustering.server] (DistributedSingletonService - 1) WFLYCLSV0001: This node will now operate as the singleton provider of the org.jboss.as.quickstarts.ha.singleton.service.primary-only service INFO [org.jboss.as.quickstarts.ha.singleton.service.primary.SingletonService] (MSC service thread 1-7) Singleton service is starting on node 'node1'. ... INFO [org.jboss.as.quickstarts.ha.singleton.service.primary.QueryingService] (pool-4-thread-1) Singleton service is running on node 'node1'.
You might see the following warnings in the server log after the applications are deployed. These warnings can be ignored in a development environment.
WARN [org.jboss.as.clustering.jgroups.protocol.UDP] (ServerService Thread Pool -- 68) JGRP000015: the receive buffer of socket MulticastSocket was set to 20MB, but the OS only allocated 6.71MB. This might lead to performance problems. Please set your max receive buffer in the OS correctly (e.g. net.core.rmem_max on Linux) WARN [org.jboss.as.clustering.jgroups.protocol.UDP] (ServerService Thread Pool -- 68) JGRP000015: the receive buffer of socket MulticastSocket was set to 25MB, but the OS only allocated 6.71MB. This might lead to performance problems. Please set your max receive buffer in the OS correctly (e.g. net.core.rmem_max on Linux)
-
Use the following command to deploy the same archive to the second server. Because the default socket binding port for deployment is
9990
and the second server ports are offset by100
, you must pass the sum,10090
, for the socket binding port as the argument to thedeploy
Maven goal.mvn wildfly:deploy -Dwildfly.port=10090
-
Investigate the primary-only console output for both servers. Verify that the
target/ha-singleton-service-primary-only.jar
archive is deployed tonode2
by checking the server log.INFO [org.jboss.as.repository] (management-handler-thread - 4) WFLYDR0001: Content added at location /Users/rhusar/wildfly/build/target/y/standalone/data/content/18/6efcc6c07b471f641cfcc97f9120505726e6bd/content INFO [org.jboss.as.server.deployment] (MSC service thread 1-1) WFLYSRV0027: Starting deployment of "ha-singleton-service-primary-only.jar" (runtime-name: "ha-singleton-service-primary-only.jar") INFO [org.jboss.as.quickstarts.ha.singleton.service.primary.ServiceActivator] (MSC service thread 1-6) Singleton and querying services activated. INFO [org.jboss.as.quickstarts.ha.singleton.service.primary.QueryingService] (MSC service thread 1-5) Querying service is starting. ... INFO [org.jboss.as.server] (management-handler-thread - 4) WFLYSRV0010: Deployed "ha-singleton-service-primary-only.jar" (runtime-name : "ha-singleton-service-primary-only.jar") ... INFO [org.jboss.as.quickstarts.ha.singleton.service.primary.QueryingService] (pool-4-thread-1) Singleton service is running on node 'node1'.
-
Inspect the server log of the first node. Since the cluster membership has changed, the election policy determines which node will run the singleton.
INFO [org.infinispan.CLUSTER] (remote-thread--p7-t1) ISPN000336: Finished cluster-wide rebalance for cache default, topology id = 5 INFO [org.wildfly.clustering.server] (DistributedSingletonService - 1) WFLYCLSV0003: node1 elected as the singleton provider of the org.jboss.as.quickstarts.ha.singleton.service.primary-only service
-
Verify that the querying service is running on all nodes and that all are querying the same singleton service instance by confirming that the same node name is printed in the log. Both nodes will output the following message every 5 seconds:
INFO [org.jboss.as.quickstarts.ha.singleton.service.primary.QueryingService] (pool-4-thread-1) Singleton service is running on node 'node1'.
-
To verify failover of the singleton service, shut down the server operating as the singleton master by using the
Ctrl
+C
key combination in the terminal. The following messages confirm that the node is shut down.INFO [org.jboss.as.quickstarts.ha.singleton.service.primary.QueryingService] (pool-4-thread-1) Singleton service is running on node 'node1'. INFO [org.jboss.as.server] (Thread-2) WFLYSRV0220: Server shutdown has been requested via an OS signal INFO [org.jboss.as.quickstarts.ha.singleton.service.primary.SingletonService] (MSC service thread 1-3) Singleton service is stopping on node 'node1'. INFO [org.jboss.as.quickstarts.ha.singleton.service.primary.QueryingService] (MSC service thread 1-6) Querying service is stopping. ... INFO [org.jboss.as] (MSC service thread 1-6) WFLYSRV0050: JBoss EAP 7.1.0.Beta1 (WildFly Core 3.0.0.Beta26-redhat-1) stopped in 66ms
Note that during shutdown of the server an exception relating to clean shutdown of the Infinispan component might appear in the log, for instance.
ERROR [org.infinispan.remoting.transport.jgroups.CommandAwareRpcDispatcher] (thread-19) ISPN000065: Exception while marshalling object: CacheNotFoundResponse: org.infinispan.IllegalLifecycleStateException: Cache marshaller has been stopped
The issue can be safely ignored for now and will be fixed in future versions of the application server.
-
Now observe the log messages on the second server. The second node is now elected as the singleton master.
INFO [org.wildfly.clustering.server] (DistributedSingletonService - 1) WFLYCLSV0003: node2 elected as the singleton provider of the org.jboss.as.quickstarts.ha.singleton.service.primary-only service INFO [org.wildfly.clustering.server] (DistributedSingletonService - 1) WFLYCLSV0001: This node will now operate as the singleton provider of the org.jboss.as.quickstarts.ha.singleton.service.primary-only service INFO [org.jboss.as.quickstarts.ha.singleton.service.primary.SingletonService] (MSC service thread 1-8) Singleton service is starting on node 'node2'.
-
Start the {productName} servers as described in the above section.
-
Open a terminal and navigate to the
primary-only/
directory located in the root directory of this quickstart. -
Use the following command to undeploy the JAR archive from Server 1.
$ mvn wildfly:undeploy
-
Use the following command to undeploy the JAR archive from Server 2.
$ mvn wildfly:undeploy -Dwildfly.port=10090
This example demonstrates a singleton service that is installed with a backup service. The backup service is running on all nodes that are not elected to be running the singleton service.
-
Start the {productName} servers as described in the above section.
-
Open a terminal and navigate to the
with-backups/
directory located in the root directory of this quickstart. -
Use the following command to clean up any previously built artifacts, and build and deploy the JAR archive.
$ mvn clean install wildfly:deploy
Use the following command to deploy the same archive to the second server. Because the default socket binding port for deployment is 9990
and the second server ports are offset by 100
, you must pass the sum, 10090
, for the socket binding port as the argument to the deploy
Maven goal.
mvn wildfly:deploy -Dwildfly.port=10090
Verify that the target/ha-singleton-service-with-backups.jar
archive is deployed to node1
, which is the first server started without port offset, by checking the server log to see that the primary service is running.
INFO [org.jboss.as.quickstarts.ha.singleton.service.backups.SingletonService] (pool-18-thread-1) Primary singleton service is running on node 'node1'.
All other nodes log that the backup singleton service is running.
INFO [org.jboss.as.quickstarts.ha.singleton.service.backups.SingletonService] (pool-20-thread-1) Backup singleton service is running on node 'node2'.
-
Start the {productName} servers as described in the above section.
-
Open a terminal and navigate to the
with-backups/
directory located in the root directory of this quickstart. -
Use the following command to undeploy the JAR archive from Server 1.
$ mvn wildfly:undeploy
-
Use the following command to undeploy the JAR archive from Server 2.
$ mvn wildfly:undeploy -Dwildfly.port=10090
As mentioned previously, the activate()
method in the ServiceActivator
class for each example in this quickstart uses the default election policy to build the singleton services. Once you have successfully deployed and verified these examples, you might want to test different election policy configurations to see how they work.
Election policies are configured using {productName} management CLI commands. Scripts are provided to configure a simple name preference election policy and a random election policy. A script is also provided to configure a quorum for the singleton policy.
This example configures the default election policy to be based on logical names.
-
If you have tested other election policies that configured the
singleton
subsystem, see Restoring the Default Singleton Subsystem Configuration for instructions to restore the singleton election policy to the default configuration. -
Start the two servers with the HA profile as described above.
-
Review the contents of the
name-preference-election-policy-add.cli
file located in the root of this quickstart directory. This script configures the default election policy to choose nodes in a preferred order ofnode3
,node2
, andnode1
using this command./subsystem=singleton/singleton-policy=default/election-policy=simple:write-attribute(name=name-preferences,value=[node3,node2,node1])
-
Open a new terminal, navigate to the root directory of this quickstart, and run the following command to execute the script for Server 1. Make sure you replace {jbossHomeName}_1 with the path to the target Server 1.
$ {jbossHomeName}_1/bin/jboss-cli.sh --connect --file=name-preference-election-policy-add.cli
NoteFor Windows, use the {jbossHomeName}_1\bin\jboss-cli.bat
script.You should see the following result when you run the script.
{ "outcome" => "success", "response-headers" => { "operation-requires-reload" => true, "process-state" => "reload-required" } }
Note that the
name-preference-election-policy-add.cli
script executes thereload
command, so a reload is not required. -
Stop the server and review the changes made to the
standalone-ha.xml
server configuration file by the management CLI commands. Thesingleton
subsystem now contains aname-preferences
element under thesimple-election-policy
that specifies the preferencesnode3 node2 node1
.<subsystem xmlns="{SingletonSubsystemNamespace}"> <singleton-policies default="default"> <singleton-policy name="default" cache-container="server"> <simple-election-policy> <name-preferences>node3 node2 node1</name-preferences> </simple-election-policy> </singleton-policy> </singleton-policies> </subsystem>
-
Repeat these steps for the second server. Note that if the second server is using a port offset, you must specify the controller address on the command line by adding
--controller=localhost:10090
.$ {jbossHomeName}_2/bin/jboss-cli.sh --connect --controller=localhost:10090 --file=name-preference-election-policy-add.cli
NoteFor Windows, use the {jbossHomeName}_2\bin\jboss-cli.bat
script. -
Make sure both servers are started, deploy one of the examples to both servers, and verify that the election policy is now in effect. The server running the election policy should now log the following message.
INFO [org.wildfly.clustering.server] (DistributedSingletonService - 1) WFLYCLSV0003: node2 elected as the singleton provider of the org.jboss.as.quickstarts.ha.singleton.service.primary-only service
The other nodes should log the following message.
INFO [org.jboss.as.quickstarts.ha.singleton.service.primary.QueryingService] (pool-7-thread-1) Singleton service is running on node 'node2'.
This example configures an election policy that elects a random cluster member when the cluster membership changes.
-
If you have tested other election policies that configured the
singleton
subsystem, see Restoring the Default Singleton Subsystem Configuration for instructions to restore the singleton election policy to the default configuration. -
Start the two servers with the HA profile as described above.
-
Review the contents of the
random-election-policy-add.cli
file located in the root of this quickstart directory. This script removes the default simple election policy and configures the default election policy to elect a random cluster member using these commands./subsystem=singleton/singleton-policy=default/election-policy=simple:remove(){allow-resource-service-restart=true} /subsystem=singleton/singleton-policy=default/election-policy=random:add()
-
Open a new terminal, navigate to the root directory of this quickstart, and run the following command to execute the script for Server 1. Make sure you replace
{jbossHomeName}_1
with the path to the target Server 1.$ {jbossHomeName}_1/bin/jboss-cli.sh --connect --file=random-election-policy-add.cli
NoteFor Windows, use the {jbossHomeName}_1\bin\jboss-cli.bat
script.You should see the following result when you run the script.
The batch executed successfully process-state: reload-required
Note that the
random-election-policy-add.cli
script executes thereload
command, so a reload is not required. -
Stop the server and review the changes made to the
standalone-ha.xml
server configuration file by the management CLI commands. Thesingleton
subsystem now contains arandom-election-policy
element under thesingleton-policy
that specifies the preferencesnode3 node2 node1
.<subsystem xmlns="{SingletonSubsystemNamespace}"> <singleton-policies default="default"> <singleton-policy name="default" cache-container="server"> <random-election-policy/> </singleton-policy> </singleton-policies> </subsystem>
-
Repeat these steps for the second server. Note that if the second server is using a port offset, you must specify the controller address on the command line by adding
--controller=localhost:10090
.$ {jbossHomeName}_2/bin/jboss-cli.sh --connect --controller=localhost:10090 --file=random-election-policy-add.cli
NoteFor Windows, use the {jbossHomeName}_2\bin\jboss-cli.bat
script. -
Make sure both servers are started, then deploy one of the examples to both servers, and verify that the election policy is now in effect.
A quorum specifies the minimum number of cluster members that must be present for the election to even begin. This mechanism is used to mitigate a split brain problem by sacrificing the availability of the singleton service. If there are less members than the specified quorum, no election is performed and the singleton service is not run on any node.
-
Quorum can be configured for any singleton policy. Optionally, if you have reconfigured the
singleton
subsystem, see Restoring the Default Singleton Subsystem Configuration for instructions to restore the singleton election policy to the default configuration. -
Start the two servers with the HA profile as described above.
-
Review the contents of the
quorum-add.cli
file located in the root of this quickstart directory. This script specifies the minimum number of cluster members required for the singleton policy using this command./subsystem=singleton/singleton-policy=default:write-attribute(name=quorum,value=2)
-
Open a new terminal, navigate to the root directory of this quickstart, and run the following command to execute the script for Server 1. Make sure you replace
{jbossHomeName}_1
with the path to the target Server 1.$ {jbossHomeName}_1/bin/jboss-cli.sh --connect --file=quorum-add.cli
NoteFor Windows, use the {jbossHomeName}_1\bin\jboss-cli.bat
script.You should see the following result when you run the script.
{ "outcome" => "success", "response-headers" => { "operation-requires-reload" => true, "process-state" => "reload-required" } }
Note that the
quorum-add.cli
script executes thereload
command, so a reload is not required. -
Review the changes made to the
standalone-ha.xml
server configuration file by the management CLI commands. Thesingleton
subsystem now contains aquorum
attribute for thesingleton-policy
element that specifies the minimum number.<subsystem xmlns="{SingletonSubsystemNamespace}"> <singleton-policies default="default"> <singleton-policy name="default" cache-container="server" quorum="2"> <simple-election-policy/> </singleton-policy> </singleton-policies> </subsystem>
-
Repeat these steps for the second server. Note that if the second server is using a port offset, you must specify the controller address on the command line by adding
--controller=localhost:10090
.$ {jbossHomeName}_2/bin/jboss-cli.sh --connect --controller=localhost:10090 --file=quorum-add.cli
NoteFor Windows, use the {jbossHomeName}_2\bin\jboss-cli.bat` script. -
Make sure both servers are started, deploy one of the examples to both servers. While both servers are running, observe the server logs. The server running the election policy should now log the following message.
INFO [org.wildfly.clustering.server] (DistributedSingletonService - 1) WFLYCLSV0007: Just reached required quorum of 2 for org.jboss.as.quickstarts.ha.singleton.service.primary-only service. If this cluster loses another member, no node will be chosen to provide this service.
-
Shut down one of the servers by using the
Ctrl
+C
key combination in the terminal to verify that no singleton service will be running after the quorum is not reached.WARN [org.wildfly.clustering.server] (DistributedSingletonService - 1) WFLYCLSV0006: Failed to reach quorum of 2 for org.jboss.as.quickstarts.ha.singleton.service.primary-only service. No singleton master will be elected. INFO [org.wildfly.clustering.server] (thread-20) WFLYCLSV0002: This node will no longer operate as the singleton provider of the org.jboss.as.quickstarts.ha.singleton.service.primary-only service INFO [org.jboss.as.quickstarts.ha.singleton.service.primary.SingletonService] (MSC service thread 1-3) Singleton service is stopping on node 'node2'. INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (thread-2) ISPN000094: Received new cluster view for channel server: [node2|4] (1) [node2] ... WARN [org.jboss.as.quickstarts.ha.singleton.service.primary.QueryingService] (pool-4-thread-1) Failed to query singleton service.
-
A
quorum-remove.cli
script is provided in the root directory of this quickstart that removes the quorum from thesingleton
subsystem.
If the singleton is running on multiple nodes, check for the following issues.
-
The most common cause of this problem is starting the servers with the
standalone.xml
orstandalone-full.xml
profile instead of with thestandalone-ha.xml
orstandalone-full-ha.xml
profile. Make sure to start the server with an HA profile using-c standalone-ha.xml
. -
Another common cause is because the server instances did not discover each other and each server is operating as a singleton cluster. Ensure that
multicast
is enabled or change thejgroups
subsystem configuration to use a different discovery mechanism. Confirm the following message in the server log to ensure that the discovery was successful.INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (MSC service thread 1-3) ISPN000094: Received new cluster view for channel server: [node1|1] (2) [node1, node2]
If you have not yet done so, you can undeploy all of the deployed artifacts by following these steps.
-
Start the two servers with the HA profile as described above.
-
Open a terminal and navigate to the root directory of this quickstart.
-
Use the following commands to undeploy all of the artifacts.
$ mvn wildfly:undeploy $ mvn wildfly:undeploy -Dwildfly.port=10090
Some of these examples require that you modify the election policies for the singleton
subsystem by running management CLI scripts. After you have completed testing each configuration, it is important to restore the singleton
subsystem to its default configuration before you run any other examples.
-
Start both servers with the HA profile as described above.
-
Open a terminal and navigate to the root directory of this quickstart.
-
Restore your default server configurations by running these commands.
$ {jbossHomeName}_1/bin/jboss-cli.sh --connect --file=restore-singleton-subsystem.cli $ {jbossHomeName}_2/bin/jboss-cli.sh --connect --controller=localhost:10090 --file=restore-singleton-subsystem.cli
NoteFor Windows, use the {jbossHomeName}_1\bin\jboss-cli.bat
and{jbossHomeName}_2\bin\jboss-cli.bat
scripts.