KEYCLOAK-5710 Change cache-server to use backups based caches

This commit is contained in:
mposolda 2017-10-19 08:53:04 +02:00
parent f0bbcbf0fd
commit 9a19e95b60
10 changed files with 258 additions and 79 deletions

View file

@ -21,23 +21,25 @@ So 2 infinispan servers and 4 Keycloak servers are totally in the testing setup.
* Site2 consists of infinispan server `jdg2` and 2 Keycloak servers `node21` and `node22` .
* Infinispan servers `jdg1` and `jdg2` forms cluster with each other. The communication between them is the only communication between the 2 datacenters.
* Infinispan servers `jdg1` and `jdg2` are connected with each other through the RELAY protocol and "backup" based infinispan caches in
similar way as described in the infinispan documentation - https://access.redhat.com/documentation/en-us/red_hat_jboss_data_grid/7.1/html-single/administration_and_configuration_guide/#configure_cross_datacenter_replication_remote_client_server_mode .
* Keycloak servers `node11` and `node12` forms cluster with each other, but they don't communicate with any server in `site2` . They communicate with infinispan server `jdg1` through the HotRod protocol (Remote cache).
* Keycloak servers `node11` and `node12` forms cluster with each other, but they don't communicate with any server in `site2` .
They communicate with infinispan server `jdg1` through the HotRod protocol (Remote cache).
* Same applies for `node21` and `node22` . They have cluster with each other and communicate just with `jdg2` server through the HotRod protocol.
TODO: Picture on blog
* For example when some object (realm, client, role, user, ...) is updated on `node11`, the `node11` will send invalidation message. It does it by saving special cache entry to the remote cache `work` on `jdg1` .
The `jdg1` notifies client listeners in same DC (hence on `node12`) and propagate the message to it. But `jdg1` is in replicated cache with `jdg2` .
The `jdg1` notifies client listeners in same DC (hence on `node12`) and propagate the message to it. But `jdg1` is connected through backup with `jdg2` too.
So the entry is saved on `jdg2` too and `jdg2` will notify client listeners on nodes `node21` and `node22`.
All the nodes know that they should invalidate the updated object from their caches. The caches with the actual data (`realms`, `users` and `authorization`) are infinispan local caches.
TODO: Picture and better explanation?
* For example when some userSession is created/updated/removed on `node11` it is saved in cluster on current DC, so the `node12` can see it. But it's saved also to remote cache on `jdg1` server.
The userSession is then automatically seen on `jdg2` server because there is replicated cache `sessions` between `jdg1` and `jdg2` . Server `jdg2` then notifies nodes `node21` and `node22` through
The userSession is then automatically seen on `jdg2` server through the backup cache `sessions` between `jdg1` and `jdg2` . Server `jdg2` then notifies nodes `node21` and `node22` through
the client listeners (Feature of Remote Cache and HotRod protocol. See infinispan docs for details). The node, who is owner of the userSession (either `node21` or `node22`) will update userSession in the cluster
on `site2` . Hence any user requests coming to Keycloak nodes on `site2` will see latest updates.
@ -49,29 +51,74 @@ Example setup assumes all 6 servers are bootstrapped on localhost, but each on d
Infinispan Server setup
-----------------------
1) Download Infinispan 8.2.6 server and unzip to some folder
1) Download Infinispan 8.2.8 server and unzip to some folder
2) Add this into `JDG1_HOME/standalone/configuration/clustered.xml` under cache-container named `clustered` :
2) Change those things in the `JDG1_HOME/standalone/configuration/clustered.xml` in the configuration of JGroups subsystem:
2.a) Add the `xsite` channel, which will use `tcp` stack, under `channels` element:
```xml
<channels default="cluster">
<channel name="cluster"/>
<channel name="xsite" stack="tcp"/>
</channels>
```
2.b) Add `relay` element to the end of the `udp` stack:
```xml
<stack name="udp">
...
<relay site="site1">
<remote-site name="site2" channel="xsite"/>
</relay>
</stack>
```
2.c) Configure `tcp` stack to use TCPPING instead of MPING . Just remove MPING element and replace with the TCPPING like this:
```xml
<stack name="tcp">
<transport type="TCP" socket-binding="jgroups-tcp"/>
<protocol type="TCPPING">
<property name="initial_hosts">localhost[8610],localhost[9610]"</property>
<property name="ergonomics">false</property>
</protocol>
<protocol type="MERGE3"/>
...
</stack>
```
3) Add this into `JDG1_HOME/standalone/configuration/clustered.xml` under cache-container named `clustered` :
```xml
<cache-container name="clustered" default-cache="default" statistics="true">
...
<replicated-cache-configuration name="sessions-cfg" mode="ASYNC" start="EAGER" batching="false">
<transaction mode="NON_XA" locking="PESSIMISTIC"/>
</replicated-cache-configuration>
<replicated-cache name="work" configuration="sessions-cfg" />
<replicated-cache name="sessions" configuration="sessions-cfg" />
<replicated-cache name="offlineSessions" configuration="sessions-cfg" />
<replicated-cache name="actionTokens" configuration="sessions-cfg" />
<replicated-cache name="loginFailures" configuration="sessions-cfg" />
<replicated-cache-configuration name="sessions-cfg" mode="SYNC" start="EAGER" batching="false">
<transaction mode="NON_XA" locking="PESSIMISTIC"/>
<backups>
<backup site="site2" failure-policy="FAIL" strategy="SYNC" enabled="true"/>
</backups>
</replicated-cache-configuration>
<replicated-cache name="work" configuration="sessions-cfg"/>
<replicated-cache name="sessions" configuration="sessions-cfg"/>
<replicated-cache name="offlineSessions" configuration="sessions-cfg"/>
<replicated-cache name="actionTokens" configuration="sessions-cfg"/>
<replicated-cache name="loginFailures" configuration="sessions-cfg"/>
</cache-container>
```
3) Copy the server into the second location referred later as `JDG2_HOME`
4) Copy the server into the second location referred later as `JDG2_HOME`
4) Start server `jdg1`:
5) In the `JDG2_HOME/standalone/configuration/clustered.xml` exchange `site1` with `site2` and viceversa in the configuration of `relay` in the
JGroups subsystem and in configuration of `backups` in the cache-subsystem.
NOTE: It's currently needed to have different configuration files for both sites as Infinispan subsystem doesn't support
replacing site name with expressions. See https://issues.jboss.org/browse/WFLY-9458 for more details.
6) Start server `jdg1`:
```
cd JDG1_HOME/bin
@ -80,19 +127,22 @@ cd JDG1_HOME/bin
-Djboss.node.name=jdg1
```
5) Start server `jdg2`:
7) Start server `jdg2` . There is different multicast address, so the `jdg1` and `jdg2` servers are not in "direct" cluster with each other,
but they are just connected through the RELAY protocol and TCP JGroups stack is used for communication between them. So the startup command is like this:
```
cd JDG2_HOME/bin
./standalone.sh -c clustered.xml -Djava.net.preferIPv4Stack=true \
-Djboss.socket.binding.port-offset=2010 -Djboss.default.multicast.address=234.56.78.99 \
-Djboss.socket.binding.port-offset=2010 -Djboss.default.multicast.address=234.56.78.100 \
-Djboss.node.name=jdg2
```
6) There should be message in the log that nodes are in cluster with each other:
8) To verify that channel works at this point, you may need to use JConsole and connect either to JDG1 or JDG2 running server. When
use the MBean `jgroups:type=protocol,cluster="cluster",protocol=RELAY2` and operation `printRoutes`, you should see the output like this:
```
Received new cluster view for channel clustered: [jdg1|1] (2) [jdg1, jdg2]
site1 --> _jdg1:site1
site2 --> _jdg2:site2
```
Keycloak servers setup

View file

@ -113,9 +113,13 @@ public class ConcurrencyJDGRemoteCacheClientListenersTest {
Assert.assertEquals(info.val.get(), info.dc2Created.get());
Assert.assertEquals(info.val.get() * 2, info.dc1Updated.get());
Assert.assertEquals(info.val.get() * 2, info.dc2Updated.get());
worker1.cache.remove(entry.getKey());
}
} finally {
// Remove items
for (Map.Entry<String, EntryInfo> entry : state.entrySet()) {
worker1.cache.remove(entry.getKey());
}
// Finish JVM
worker1.cache.getCacheManager().stop();
worker2.cache.getCacheManager().stop();

View file

@ -49,7 +49,7 @@ public class ConcurrencyJDGRemoteCacheTest {
public static void main(String[] args) throws Exception {
// Init map somehow
for (int i=0 ; i<100 ; i++) {
for (int i=0 ; i<30 ; i++) {
String key = "key-" + i;
state.put(key, new EntryInfo());
}
@ -77,8 +77,8 @@ public class ConcurrencyJDGRemoteCacheTest {
}
private static Worker createWorker(int threadId) {
EmbeddedCacheManager manager = new TestCacheManagerFactory().createManager(threadId, InfinispanConnectionProvider.WORK_CACHE_NAME, RemoteStoreConfigurationBuilder.class);
Cache<String, Integer> cache = manager.getCache(InfinispanConnectionProvider.WORK_CACHE_NAME);
EmbeddedCacheManager manager = new TestCacheManagerFactory().createManager(threadId, InfinispanConnectionProvider.SESSION_CACHE_NAME, RemoteStoreConfigurationBuilder.class);
Cache<String, Integer> cache = manager.getCache(InfinispanConnectionProvider.SESSION_CACHE_NAME);
System.out.println("Retrieved cache: " + threadId);
@ -142,19 +142,33 @@ public class ConcurrencyJDGRemoteCacheTest {
}
public static int getClusterStartupTime(Cache<String, Integer> cache, String cacheKey, EntryInfo wrapper) {
int startupTime = new Random().nextInt(1024);
Integer startupTime = new Random().nextInt(1024);
// Concurrency doesn't work correctly with this
//Integer existingClusterStartTime = (Integer) cache.putIfAbsent(cacheKey, startupTime);
// Concurrency works fine with this
RemoteCache remoteCache = cache.getAdvancedCache().getComponentRegistry().getComponent(PersistenceManager.class).getStores(RemoteStore.class).iterator().next().getRemoteCache();
Integer existingClusterStartTime = (Integer) remoteCache.withFlags(Flag.FORCE_RETURN_VALUE).putIfAbsent(cacheKey, startupTime);
if (existingClusterStartTime == null) {
Integer existingClusterStartTime = null;
for (int i=0 ; i<10 ; i++) {
try {
existingClusterStartTime = (Integer) remoteCache.withFlags(Flag.FORCE_RETURN_VALUE).putIfAbsent(cacheKey, startupTime);
} catch (Exception ce) {
if (i == 9) {
throw ce;
//break;
} else {
System.err.println("EXception: i=" + i);
}
}
}
if (existingClusterStartTime == null || startupTime.equals(remoteCache.get(cacheKey))) {
wrapper.successfulInitializations.incrementAndGet();
return startupTime;
} else {
System.err.println("Not equal!!! startupTime=" + startupTime + ", existingClusterStartTime=" + existingClusterStartTime );
return existingClusterStartTime;
}
}

View file

@ -17,8 +17,12 @@
package org.keycloak.cluster.infinispan;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.HashSet;
import java.util.List;
import java.util.concurrent.ExecutorService;
import java.util.concurrent.Executors;
import java.util.concurrent.atomic.AtomicInteger;
import org.infinispan.Cache;
@ -58,6 +62,8 @@ public class ConcurrencyJDGSessionsCacheTest {
private static RemoteCache remoteCache1;
private static RemoteCache remoteCache2;
private static List<ExecutorService> executors = new ArrayList<>();
private static final AtomicInteger failedReplaceCounter = new AtomicInteger(0);
private static final AtomicInteger failedReplaceCounter2 = new AtomicInteger(0);
@ -144,6 +150,7 @@ public class ConcurrencyJDGSessionsCacheTest {
// Explicitly call put on remoteCache (KcRemoteCache.write ignores remote writes)
InfinispanUtil.getRemoteCache(cache1).put("123", session);
InfinispanUtil.getRemoteCache(cache2).replace("123", session);
// Create caches, listeners and finally worker threads
Thread worker1 = createWorker(cache1, 1);
@ -172,14 +179,19 @@ public class ConcurrencyJDGSessionsCacheTest {
System.out.println("Sleeping before other report");
Thread.sleep(1000);
Thread.sleep(2000);
System.out.println("Finished. Took: " + took + " ms. Notes: " + cache1.get("123").getEntity().getNotes().size() +
", successfulListenerWrites: " + successfulListenerWrites.get() + ", successfulListenerWrites2: " + successfulListenerWrites2.get() +
", failedReplaceCounter: " + failedReplaceCounter.get() + ", failedReplaceCounter2: " + failedReplaceCounter2.get());
System.out.println("Histogram: ");
histogram.dumpStats();
//histogram.dumpStats();
// shutdown pools
for (ExecutorService ex : executors) {
ex.shutdown();
}
// Finish JVM
cache1.getCacheManager().stop();
@ -218,10 +230,15 @@ public class ConcurrencyJDGSessionsCacheTest {
private RemoteCache remoteCache;
private AtomicInteger listenerCount;
private ExecutorService executor;
public HotRodListener(Cache<String, SessionEntityWrapper<UserSessionEntity>> origCache, RemoteCache remoteCache, AtomicInteger listenerCount) {
this.listenerCount = listenerCount;
this.remoteCache = remoteCache;
this.origCache = origCache;
executor = Executors.newCachedThreadPool();
executors.add(executor);
}
@ClientCacheEntryCreated
@ -235,25 +252,37 @@ public class ConcurrencyJDGSessionsCacheTest {
String cacheKey = (String) event.getKey();
listenerCount.incrementAndGet();
// TODO: can be optimized - object sent in the event
VersionedValue<SessionEntity> versionedVal = remoteCache.getVersioned(cacheKey);
executor.submit(() -> {
// TODO: can be optimized - object sent in the event
VersionedValue<SessionEntity> versionedVal = remoteCache.getVersioned(cacheKey);
for (int i = 0; i < 10; i++) {
if (versionedVal.getVersion() < event.getVersion()) {
System.err.println("INCOMPATIBLE VERSION. event version: " + event.getVersion() + ", entity version: " + versionedVal.getVersion());
return;
}
if (versionedVal.getVersion() < event.getVersion()) {
System.err.println("INCOMPATIBLE VERSION. event version: " + event.getVersion() + ", entity version: " + versionedVal.getVersion() + ", i=" + i);
try {
Thread.sleep(100);
} catch (InterruptedException ie) {
throw new RuntimeException(ie);
}
SessionEntity session = (SessionEntity) remoteCache.get(cacheKey);
SessionEntityWrapper sessionWrapper = new SessionEntityWrapper(session);
versionedVal = remoteCache.getVersioned(cacheKey);
} else {
break;
}
}
if (listenerCount.get() % 100 == 0) {
logger.infof("Listener count: " + listenerCount.get());
}
SessionEntity session = (SessionEntity) versionedVal.getValue();
SessionEntityWrapper sessionWrapper = new SessionEntityWrapper(session);
// TODO: for distributed caches, ensure that it is executed just on owner OR if event.isCommandRetried
origCache
.getAdvancedCache().withFlags(Flag.SKIP_CACHE_LOAD, Flag.SKIP_CACHE_STORE)
.replace(cacheKey, sessionWrapper);
if (listenerCount.get() % 100 == 0) {
logger.infof("Listener count: " + listenerCount.get());
}
// TODO: for distributed caches, ensure that it is executed just on owner OR if event.isCommandRetried
origCache
.getAdvancedCache().withFlags(Flag.SKIP_CACHE_LOAD, Flag.SKIP_CACHE_STORE)
.replace(cacheKey, sessionWrapper);
});
}
@ -299,7 +328,7 @@ public class ConcurrencyJDGSessionsCacheTest {
RemoteCache secondDCRemoteCache = myThreadId == 1 ? remoteCache2 : remoteCache1;
UserSessionEntity thatSession = (UserSessionEntity) secondDCRemoteCache.get("123");
Assert.assertEquals("someVal", thatSession.getNotes().get(noteKey));
//Assert.assertEquals("someVal", thatSession.getNotes().get(noteKey));
//System.out.println("Passed");
}
@ -308,7 +337,8 @@ public class ConcurrencyJDGSessionsCacheTest {
private boolean cacheReplace(VersionedValue<UserSessionEntity> oldSession, UserSessionEntity newSession) {
try {
boolean replaced = remoteCache.replaceWithVersion("123", newSession, oldSession.getVersion());
//cache.replace("123", newSession);
//boolean replaced = true;
//remoteCache.replace("123", newSession);
if (!replaced) {
failedReplaceCounter.incrementAndGet();
//return false;

View file

@ -31,6 +31,7 @@
<outputDirectory>cache-server-${cache.server}</outputDirectory>
<excludes>
<exclude>**/*.sh</exclude>
<exclude>**/clustered.xml</exclude>
</excludes>
</fileSet>
<fileSet>

View file

@ -23,34 +23,23 @@
<xsl:output method="xml" version="1.0" encoding="UTF-8" indent="yes" xalan:indent-amount="4" standalone="no"/>
<xsl:strip-space elements="*"/>
<xsl:param name="local.site" />
<xsl:param name="remote.site" />
<xsl:variable name="nsCacheServer" select="'urn:infinispan:server:core:'"/>
<xsl:variable name="nsJGroups" select="'urn:infinispan:server:jgroups:'"/>
<xsl:template match="//*[local-name()='subsystem' and starts-with(namespace-uri(), $nsCacheServer)]
/*[local-name()='cache-container' and starts-with(namespace-uri(), $nsCacheServer) and @name='local']">
<xsl:copy>
<xsl:apply-templates select="@* | node()" />
<local-cache-configuration name="sessions-cfg" start="EAGER" batching="false">
<transaction mode="NON_XA" locking="PESSIMISTIC"/>
</local-cache-configuration>
<local-cache name="sessions" configuration="sessions-cfg" />
<local-cache name="offlineSessions" configuration="sessions-cfg" />
<local-cache name="loginFailures" configuration="sessions-cfg" />
<local-cache name="actionTokens" configuration="sessions-cfg" />
<local-cache name="work" configuration="sessions-cfg" />
<local-cache name="employee-distributable-cache.ssoCache" configuration="sessions-cfg"/>
<local-cache name="employee-distributable-cache" configuration="sessions-cfg"/>
</xsl:copy>
</xsl:template>
<!-- Configuration of infinispan caches in infinispan-subsystem -->
<xsl:template match="//*[local-name()='subsystem' and starts-with(namespace-uri(), $nsCacheServer)]
/*[local-name()='cache-container' and starts-with(namespace-uri(), $nsCacheServer) and @name='clustered']">
<xsl:copy>
<xsl:apply-templates select="@* | node()" />
<replicated-cache-configuration name="sessions-cfg" mode="ASYNC" start="EAGER" batching="false">
<replicated-cache-configuration name="sessions-cfg" mode="SYNC" start="EAGER" batching="false">
<transaction mode="NON_XA" locking="PESSIMISTIC"/>
<backups>
<backup site="{$remote.site}" failure-policy="FAIL" strategy="SYNC" enabled="true"/>
</backups>
</replicated-cache-configuration>
@ -64,6 +53,42 @@
</xsl:copy>
</xsl:template>
<!-- Add "xsite" channel in JGroups subsystem -->
<xsl:template match="//*[local-name()='subsystem' and starts-with(namespace-uri(), $nsJGroups)]
/*[local-name()='channels' and starts-with(namespace-uri(), $nsJGroups) and @default='cluster']">
<xsl:copy>
<xsl:apply-templates select="@* | node()" />
<channel name="xsite" stack="tcp"/>
</xsl:copy>
</xsl:template>
<!-- Add "relay" to JGroups stack "udp" -->
<xsl:template match="//*[local-name()='subsystem' and starts-with(namespace-uri(), $nsJGroups)]
/*[local-name()='stacks' and starts-with(namespace-uri(), $nsJGroups)]
/*[local-name()='stack' and @name='udp']">
<xsl:copy>
<xsl:apply-templates select="@* | node()" />
<relay site="{$local.site}">
<remote-site name="{$remote.site}" channel="xsite"/>
</relay>
</xsl:copy>
</xsl:template>
<!-- Replace MPING with TCPPING in JGroups stack "tcp" -->
<xsl:template match="//*[local-name()='subsystem' and starts-with(namespace-uri(), $nsJGroups)]
/*[local-name()='stacks' and starts-with(namespace-uri(), $nsJGroups)]
/*[local-name()='stack' and @name='tcp']
/*[local-name()='protocol' and @type='MPING']">
<protocol type="TCPPING">
<property name="initial_hosts">localhost[8610],localhost[9610]</property>
<property name="ergonomics">false</property>
</protocol>
</xsl:template>
<xsl:template match="@*|node()">
<xsl:copy>
<xsl:apply-templates select="@*|node()" />

View file

@ -107,18 +107,63 @@
</goals>
<configuration>
<transformationSets>
<!-- Configure dc-0 site in file clustered-1.xml -->
<transformationSet>
<dir>${cache.server.jboss.home}/standalone/configuration</dir>
<includes>
<include>standalone.xml</include>
<include>clustered.xml</include>
</includes>
<stylesheet>${common.resources}/add-keycloak-caches.xsl</stylesheet>
<parameters>
<parameter>
<name>local.site</name>
<value>dc-0</value>
</parameter>
<parameter>
<name>remote.site</name>
<value>dc-1</value>
</parameter>
</parameters>
<outputDir>${cache.server.jboss.home}/standalone/configuration</outputDir>
<fileMappers>
<fileMapper implementation="org.codehaus.plexus.components.io.filemappers.RegExpFileMapper">
<pattern>^(.*)\.xml$</pattern>
<replacement>$1-1.xml</replacement>
</fileMapper>
</fileMappers>
</transformationSet>
<!-- Configure dc-1 site in file clustered-2.xml -->
<transformationSet>
<dir>${cache.server.jboss.home}/standalone/configuration</dir>
<includes>
<include>clustered.xml</include>
</includes>
<stylesheet>${common.resources}/add-keycloak-caches.xsl</stylesheet>
<parameters>
<parameter>
<name>local.site</name>
<value>dc-1</value>
</parameter>
<parameter>
<name>remote.site</name>
<value>dc-0</value>
</parameter>
</parameters>
<outputDir>${cache.server.jboss.home}/standalone/configuration</outputDir>
<fileMappers>
<fileMapper implementation="org.codehaus.plexus.components.io.filemappers.RegExpFileMapper">
<pattern>^(.*)\.xml$</pattern>
<replacement>$1-2.xml</replacement>
</fileMapper>
</fileMappers>
</transformationSet>
</transformationSets>
</configuration>
</execution>
<execution>
<id>io-worker-threads</id>
<phase>process-resources</phase>

View file

@ -66,6 +66,16 @@ public class ConcurrentLoginCrossDCTest extends ConcurrentLoginTest {
.forEach(loadBalancerCtrl::enableBackendNodeByName);
}
// TODO: We know that this test won't work in cross-dc setup based on "backup "caches. But we need to add the test that clientSessions
// are invalidated after attempt of reuse the same code multiple times
@Test
@Override
@Ignore
public void concurrentCodeReuseShouldFail() throws Throwable {
}
@Test
public void concurrentLoginWithRandomDcFailures() throws Throwable {
log.info("*********************************************");

View file

@ -117,7 +117,7 @@ public class SessionExpirationCrossDCTest extends AbstractAdminCrossDCTest {
// Assert sessions removed on node1 and node2 and on remote caches. Assert that count of messages sent between DCs is not too big.
assertStatisticsExpected("After realm remove", InfinispanConnectionProvider.SESSION_CACHE_NAME, cacheDc1Statistics, cacheDc2Statistics, channelStatisticsCrossDc,
sessions01, sessions02, remoteSessions01, remoteSessions02, 40l);
sessions01, sessions02, remoteSessions01, remoteSessions02, 100l);
}
@ -207,7 +207,7 @@ public class SessionExpirationCrossDCTest extends AbstractAdminCrossDCTest {
// Assert sessions removed on node1 and node2 and on remote caches. Assert that count of messages sent between DCs is not too big.
assertStatisticsExpected("After realm remove", InfinispanConnectionProvider.OFFLINE_SESSION_CACHE_NAME, cacheDc1Statistics, cacheDc2Statistics, channelStatisticsCrossDc,
sessions01, sessions02, remoteSessions01, remoteSessions02, 70l); // Might be bigger messages as online sessions removed too.
sessions01, sessions02, remoteSessions01, remoteSessions02, 200l); // Might be bigger messages as online sessions removed too.
}
@ -226,7 +226,7 @@ public class SessionExpirationCrossDCTest extends AbstractAdminCrossDCTest {
// Assert sessions removed on node1 and node2 and on remote caches. Assert that count of messages sent between DCs is not too big.
assertStatisticsExpected("After realm logout", InfinispanConnectionProvider.SESSION_CACHE_NAME, cacheDc1Statistics, cacheDc2Statistics, channelStatisticsCrossDc,
sessions01, sessions02, remoteSessions01, remoteSessions02, 40l);
sessions01, sessions02, remoteSessions01, remoteSessions02, 100l);
}
@ -269,7 +269,7 @@ public class SessionExpirationCrossDCTest extends AbstractAdminCrossDCTest {
// Assert sessions removed on node1 and node2 and on remote caches. Assert that count of messages sent between DCs is not too big.
assertStatisticsExpected("After remove expired - 2", InfinispanConnectionProvider.SESSION_CACHE_NAME, cacheDc1Statistics, cacheDc2Statistics, channelStatisticsCrossDc,
sessions01, sessions02, remoteSessions01, remoteSessions02, 40l);
sessions01, sessions02, remoteSessions01, remoteSessions02, 100l);
}
@ -293,7 +293,7 @@ public class SessionExpirationCrossDCTest extends AbstractAdminCrossDCTest {
// Assert sessions removed on node1 and node2 and on remote caches. Assert that count of messages sent between DCs is not too big.
assertStatisticsExpected("After user remove", InfinispanConnectionProvider.SESSION_CACHE_NAME, cacheDc1Statistics, cacheDc2Statistics, channelStatisticsCrossDc,
sessions01, sessions02, remoteSessions01, remoteSessions02, 40l);
sessions01, sessions02, remoteSessions01, remoteSessions02, 100l);
}
@ -315,7 +315,7 @@ public class SessionExpirationCrossDCTest extends AbstractAdminCrossDCTest {
// Assert sessions removed on node1 and node2 and on remote caches. Assert that count of messages sent between DCs is not too big.
assertStatisticsExpected("After user remove", InfinispanConnectionProvider.OFFLINE_SESSION_CACHE_NAME, cacheDc1Statistics, cacheDc2Statistics, channelStatisticsCrossDc,
sessions01, sessions02, remoteSessions01, remoteSessions02, 40l);
sessions01, sessions02, remoteSessions01, remoteSessions02, 100l);
}
@ -343,7 +343,7 @@ public class SessionExpirationCrossDCTest extends AbstractAdminCrossDCTest {
// Assert sessions removed on node1 and node2 and on remote caches. Assert that count of messages sent between DCs is not too big.
assertStatisticsExpected("After user logout", InfinispanConnectionProvider.SESSION_CACHE_NAME, cacheDc1Statistics, cacheDc2Statistics, channelStatisticsCrossDc,
sessions01, sessions02, remoteSessions01, remoteSessions02, 40l);
sessions01, sessions02, remoteSessions01, remoteSessions02, 100l);
}
@ -495,7 +495,7 @@ public class SessionExpirationCrossDCTest extends AbstractAdminCrossDCTest {
// Assert sessions removed on node1 and node2 and on remote caches. Assert that count of messages sent between DCs is not too big, however there are some messages due to removed realm
assertAuthSessionsStatisticsExpected("After realm removed", channelStatisticsCrossDc,
0, 40l);
0, 100l);
}

View file

@ -193,7 +193,7 @@
<property name="enabled">${auth.server.crossdc} &amp;&amp; ! ${cache.server.lifecycle.skip}</property>
<property name="adapterImplClass">org.jboss.as.arquillian.container.managed.ManagedDeployableContainer</property>
<property name="jbossHome">${cache.server.home}</property>
<property name="serverConfig">clustered.xml</property>
<property name="serverConfig">clustered-1.xml</property>
<property name="jbossArguments">
-Djboss.socket.binding.port-offset=${cache.server.port.offset}
-Djboss.default.multicast.address=234.56.78.99
@ -219,10 +219,10 @@
<property name="jbossHome">${cache.server.home}</property>
<property name="setupCleanServerBaseDir">true</property>
<property name="cleanServerBaseDir">${cache.server.home}/standalone-dc-2</property>
<property name="serverConfig">clustered.xml</property>
<property name="serverConfig">clustered-2.xml</property>
<property name="jbossArguments">
-Djboss.socket.binding.port-offset=${cache.server.2.port.offset}
-Djboss.default.multicast.address=234.56.78.99
-Djboss.default.multicast.address=234.56.78.100
-Djboss.node.name=cache-server-dc-2
${adapter.test.props}
${auth.server.profile}