/* * Licensed to the Apache Software Foundation (ASF) under one or more contributor license * agreements. See the NOTICE file distributed with this work for additional information regarding * copyright ownership. The ASF licenses this file to You under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance with the License. You may obtain a * copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software distributed under the License * is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express * or implied. See the License for the specific language governing permissions and limitations under * the License. */ package org.apache.geode.internal.cache.wan.misc; import org.junit.Ignore; import org.junit.experimental.categories.Category; import org.junit.Test; import static org.junit.Assert.*; import org.apache.geode.test.dunit.cache.internal.JUnit4CacheTestCase; import org.apache.geode.test.dunit.internal.JUnit4DistributedTestCase; import org.apache.geode.test.junit.categories.DistributedTest; import org.apache.geode.cache.wan.GatewaySender; import org.apache.geode.internal.cache.BucketRegion; import org.apache.geode.internal.cache.PartitionedRegion; import org.apache.geode.internal.cache.RegionQueue; import org.apache.geode.internal.cache.wan.AbstractGatewaySender; import org.apache.geode.internal.cache.wan.WANTestBase; import org.apache.geode.internal.cache.wan.parallel.ConcurrentParallelGatewaySenderQueue; import org.apache.geode.test.dunit.Assert; import org.apache.geode.test.dunit.AsyncInvocation; import org.apache.geode.test.dunit.IgnoredException; import org.apache.geode.test.dunit.LogWriterUtils; import org.apache.geode.test.dunit.Wait; import org.apache.geode.test.dunit.WaitCriterion; import java.util.Set; /** * */ @Category(DistributedTest.class) public class CommonParallelGatewaySenderDUnitTest extends WANTestBase { @Test public void testSameSenderWithNonColocatedRegions() throws Exception { IgnoredException.addIgnoredException("cannot have the same parallel"); Integer lnPort = (Integer) vm0.invoke(() -> WANTestBase.createFirstLocatorWithDSId(1)); vm4.invoke(() -> WANTestBase.createCache(lnPort)); vm4.invoke(() -> WANTestBase.createSender("ln", 2, true, 100, 10, false, false, null, true)); vm4.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR1", "ln", 1, 100, isOffHeap())); try { vm4.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR2", "ln", 1, 100, isOffHeap())); fail("Expected IllegalStateException : cannot have the same parallel gateway sender"); } catch (Exception e) { if (!(e.getCause() instanceof IllegalStateException) || !(e.getCause().getMessage() .contains("cannot have the same parallel gateway sender id"))) { Assert.fail("Expected IllegalStateException", e); } } } /** * Simple scenario. Two regions attach the same PGS * * @throws Exception Below test is disabled intentionally 1> In this release 8.0, for rolling * upgrade support queue name is changed to old style 2>Common parallel sender for * different non colocated regions is not supported in 8.0 so no need to bother about * ParallelGatewaySenderQueue#convertPathToName 3> We have to enabled it in next release * 4> Version based rolling upgrade support should be provided. based on the version of * the gemfire QSTRING should be used between 8.0 and version prior to 8.0 */ @Test @Ignore("TODO") public void testParallelPropagation() throws Exception { Integer lnPort = (Integer) vm0.invoke(() -> WANTestBase.createFirstLocatorWithDSId(1)); Integer nyPort = (Integer) vm1.invoke(() -> WANTestBase.createFirstRemoteLocator(2, lnPort)); createCacheInVMs(nyPort, vm2, vm3); createReceiverInVMs(vm2, vm3); createCacheInVMs(lnPort, vm4, vm5, vm6, vm7); vm4.invoke(() -> WANTestBase.createSender("ln", 2, true, 100, 10, false, false, null, true)); vm5.invoke(() -> WANTestBase.createSender("ln", 2, true, 100, 10, false, false, null, true)); vm6.invoke(() -> WANTestBase.createSender("ln", 2, true, 100, 10, false, false, null, true)); vm7.invoke(() -> WANTestBase.createSender("ln", 2, true, 100, 10, false, false, null, true)); vm4.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR1", "ln", 1, 100, isOffHeap())); vm5.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR1", "ln", 1, 100, isOffHeap())); vm6.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR1", "ln", 1, 100, isOffHeap())); vm7.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR1", "ln", 1, 100, isOffHeap())); vm4.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR2", "ln", 1, 100, isOffHeap())); vm5.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR2", "ln", 1, 100, isOffHeap())); vm6.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR2", "ln", 1, 100, isOffHeap())); vm7.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR2", "ln", 1, 100, isOffHeap())); startSenderInVMs("ln", vm4, vm5, vm6, vm7); vm2.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR1", null, 1, 100, isOffHeap())); vm3.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR1", null, 1, 100, isOffHeap())); vm2.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR2", null, 1, 100, isOffHeap())); vm3.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR2", null, 1, 100, isOffHeap())); // before doing any puts, let the senders be running in order to ensure that // not a single event will be lost vm4.invoke(() -> WANTestBase.waitForSenderRunningState("ln")); vm5.invoke(() -> WANTestBase.waitForSenderRunningState("ln")); vm6.invoke(() -> WANTestBase.waitForSenderRunningState("ln")); vm7.invoke(() -> WANTestBase.waitForSenderRunningState("ln")); vm4.invoke(() -> WANTestBase.doPuts(getTestMethodName() + "_PR1", 1000)); vm4.invoke(() -> WANTestBase.doPuts(getTestMethodName() + "_PR2", 1000)); // verify all buckets drained on all sender nodes. vm4.invoke(() -> CommonParallelGatewaySenderDUnitTest .validateParallelSenderQueueAllBucketsDrained("ln")); vm5.invoke(() -> CommonParallelGatewaySenderDUnitTest .validateParallelSenderQueueAllBucketsDrained("ln")); vm6.invoke(() -> CommonParallelGatewaySenderDUnitTest .validateParallelSenderQueueAllBucketsDrained("ln")); vm7.invoke(() -> CommonParallelGatewaySenderDUnitTest .validateParallelSenderQueueAllBucketsDrained("ln")); vm2.invoke(() -> WANTestBase.validateRegionSize(getTestMethodName() + "_PR1", 1000)); vm2.invoke(() -> WANTestBase.validateRegionSize(getTestMethodName() + "_PR2", 1000)); } /** * The PGS is persistence enabled but not the Regions Below test is disabled intentionally 1> In * this release 8.0, for rolling upgrade support queue name is changed to old style 2>Common * parallel sender for different non colocated regions is not supported in 8.0 so no need to * bother about ParallelGatewaySenderQueue#convertPathToName 3> We have to enabled it in next * release 4> Version based rolling upgrade support should be provided. based on the version of * the gemfire QSTRING should be used between 8.0 and version prior to 8.0 */ @Test @Ignore("TODO") public void testParallelPropagationPersistenceEnabled() throws Exception { Integer lnPort = (Integer) vm0.invoke(() -> WANTestBase.createFirstLocatorWithDSId(1)); Integer nyPort = (Integer) vm1.invoke(() -> WANTestBase.createFirstRemoteLocator(2, lnPort)); createCacheInVMs(nyPort, vm2, vm3); createReceiverInVMs(vm2, vm3); createCacheInVMs(lnPort, vm4, vm5, vm6, vm7); vm4.invoke(() -> WANTestBase.createSender("ln", 2, true, 100, 10, false, true, null, true)); vm5.invoke(() -> WANTestBase.createSender("ln", 2, true, 100, 10, false, true, null, true)); vm6.invoke(() -> WANTestBase.createSender("ln", 2, true, 100, 10, false, true, null, true)); vm7.invoke(() -> WANTestBase.createSender("ln", 2, true, 100, 10, false, true, null, true)); vm4.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR1", "ln", 1, 100, isOffHeap())); vm5.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR1", "ln", 1, 100, isOffHeap())); vm6.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR1", "ln", 1, 100, isOffHeap())); vm7.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR1", "ln", 1, 100, isOffHeap())); vm4.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR2", "ln", 1, 100, isOffHeap())); vm5.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR2", "ln", 1, 100, isOffHeap())); vm6.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR2", "ln", 1, 100, isOffHeap())); vm7.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR2", "ln", 1, 100, isOffHeap())); startSenderInVMs("ln", vm4, vm5, vm6, vm7); vm2.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR1", null, 1, 100, isOffHeap())); vm3.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR1", null, 1, 100, isOffHeap())); vm2.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR2", null, 1, 100, isOffHeap())); vm3.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "_PR2", null, 1, 100, isOffHeap())); // before doing any puts, let the senders be running in order to ensure that // not a single event will be lost vm4.invoke(() -> WANTestBase.waitForSenderRunningState("ln")); vm5.invoke(() -> WANTestBase.waitForSenderRunningState("ln")); vm6.invoke(() -> WANTestBase.waitForSenderRunningState("ln")); vm7.invoke(() -> WANTestBase.waitForSenderRunningState("ln")); vm4.invoke(() -> WANTestBase.doPuts(getTestMethodName() + "_PR1", 1000)); vm4.invoke(() -> WANTestBase.doPuts(getTestMethodName() + "_PR2", 1000)); // verify all buckets drained on all sender nodes. vm4.invoke(() -> CommonParallelGatewaySenderDUnitTest .validateParallelSenderQueueAllBucketsDrained("ln")); vm5.invoke(() -> CommonParallelGatewaySenderDUnitTest .validateParallelSenderQueueAllBucketsDrained("ln")); vm6.invoke(() -> CommonParallelGatewaySenderDUnitTest .validateParallelSenderQueueAllBucketsDrained("ln")); vm7.invoke(() -> CommonParallelGatewaySenderDUnitTest .validateParallelSenderQueueAllBucketsDrained("ln")); vm2.invoke(() -> WANTestBase.validateRegionSize(getTestMethodName() + "_PR1", 1000)); vm2.invoke(() -> WANTestBase.validateRegionSize(getTestMethodName() + "_PR2", 1000)); } /** * Enable persistence for GatewaySender. Pause the sender and do some puts in local region. Close * the local site and rebuild the region and sender from disk store. Dispatcher should not start * dispatching events recovered from persistent sender. Check if the remote site receives all the * events. Below test is disabled intentionally 1> In this release 8.0, for rolling upgrade * support queue name is changed to old style 2>Common parallel sender for different non colocated * regions is not supported in 8.0 so no need to bother about * ParallelGatewaySenderQueue#convertPathToName 3> We have to enabled it in next release 4> * Version based rolling upgrade support should be provided. based on the version of the gemfire * QSTRING should be used between 8.0 and version prior to 8.0 */ @Test @Ignore("TODO") public void testPRWithGatewaySenderPersistenceEnabled_Restart() { // create locator on local site Integer lnPort = (Integer) vm0.invoke(() -> WANTestBase.createFirstLocatorWithDSId(1)); // create locator on remote site Integer nyPort = (Integer) vm1.invoke(() -> WANTestBase.createFirstRemoteLocator(2, lnPort)); // create receiver on remote site createCacheInVMs(nyPort, vm2, vm3); createReceiverInVMs(vm2, vm3); // create cache in local site createCacheInVMs(lnPort, vm4, vm5, vm6, vm7); // create senders with disk store String diskStore1 = (String) vm4.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, null, true)); String diskStore2 = (String) vm5.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, null, true)); String diskStore3 = (String) vm6.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, null, true)); String diskStore4 = (String) vm7.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, null, true)); LogWriterUtils.getLogWriter() .info("The DS are: " + diskStore1 + "," + diskStore2 + "," + diskStore3 + "," + diskStore4); // create PR on remote site vm2.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "PR1", null, 1, 100, isOffHeap())); vm3.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "PR1", null, 1, 100, isOffHeap())); // create PR on remote site vm2.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "PR2", null, 1, 100, isOffHeap())); vm3.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "PR2", null, 1, 100, isOffHeap())); // create PR on local site vm4.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "PR1", "ln", 1, 100, isOffHeap())); vm5.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "PR1", "ln", 1, 100, isOffHeap())); vm6.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "PR1", "ln", 1, 100, isOffHeap())); vm7.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "PR1", "ln", 1, 100, isOffHeap())); // create PR on local site vm4.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "PR2", "ln", 1, 100, isOffHeap())); vm5.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "PR2", "ln", 1, 100, isOffHeap())); vm6.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "PR2", "ln", 1, 100, isOffHeap())); vm7.invoke(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "PR2", "ln", 1, 100, isOffHeap())); // start the senders on local site startSenderInVMs("ln", vm4, vm5, vm6, vm7); // wait for senders to become running vm4.invoke(() -> WANTestBase.waitForSenderRunningState("ln")); vm5.invoke(() -> WANTestBase.waitForSenderRunningState("ln")); vm6.invoke(() -> WANTestBase.waitForSenderRunningState("ln")); vm7.invoke(() -> WANTestBase.waitForSenderRunningState("ln")); // pause the senders vm4.invoke(() -> WANTestBase.pauseSender("ln")); vm5.invoke(() -> WANTestBase.pauseSender("ln")); vm6.invoke(() -> WANTestBase.pauseSender("ln")); vm7.invoke(() -> WANTestBase.pauseSender("ln")); // start puts in region on local site vm4.invoke(() -> WANTestBase.doPuts(getTestMethodName() + "PR1", 3000)); vm4.invoke(() -> WANTestBase.doPuts(getTestMethodName() + "PR2", 5000)); LogWriterUtils.getLogWriter().info("Completed puts in the region"); // --------------------close and rebuild local site // ------------------------------------------------- // kill the senders vm4.invoke(() -> WANTestBase.killSender()); vm5.invoke(() -> WANTestBase.killSender()); vm6.invoke(() -> WANTestBase.killSender()); vm7.invoke(() -> WANTestBase.killSender()); LogWriterUtils.getLogWriter().info("Killed all the senders."); // restart the vm vm4.invoke(() -> WANTestBase.createCache(lnPort)); vm5.invoke(() -> WANTestBase.createCache(lnPort)); vm6.invoke(() -> WANTestBase.createCache(lnPort)); vm7.invoke(() -> WANTestBase.createCache(lnPort)); LogWriterUtils.getLogWriter().info("Created back the cache"); // create senders with disk store vm4.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, diskStore1, true)); vm5.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, diskStore2, true)); vm6.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, diskStore3, true)); vm7.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, diskStore4, true)); LogWriterUtils.getLogWriter().info("Created the senders back from the disk store."); // create PR on local site AsyncInvocation inv1 = vm4.invokeAsync(() -> WANTestBase .createPartitionedRegion(getTestMethodName() + "PR1", "ln", 1, 100, isOffHeap())); AsyncInvocation inv2 = vm5.invokeAsync(() -> WANTestBase .createPartitionedRegion(getTestMethodName() + "PR1", "ln", 1, 100, isOffHeap())); AsyncInvocation inv3 = vm6.invokeAsync(() -> WANTestBase .createPartitionedRegion(getTestMethodName() + "PR1", "ln", 1, 100, isOffHeap())); AsyncInvocation inv4 = vm7.invokeAsync(() -> WANTestBase .createPartitionedRegion(getTestMethodName() + "PR1", "ln", 1, 100, isOffHeap())); try { inv1.join(); inv2.join(); inv3.join(); inv4.join(); } catch (InterruptedException e) { e.printStackTrace(); fail(); } inv1 = vm4.invokeAsync(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "PR2", "ln", 1, 100, isOffHeap())); inv2 = vm5.invokeAsync(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "PR2", "ln", 1, 100, isOffHeap())); inv3 = vm6.invokeAsync(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "PR2", "ln", 1, 100, isOffHeap())); inv4 = vm7.invokeAsync(() -> WANTestBase.createPartitionedRegion(getTestMethodName() + "PR2", "ln", 1, 100, isOffHeap())); try { inv1.join(); inv2.join(); inv3.join(); inv4.join(); } catch (InterruptedException e) { e.printStackTrace(); fail(); } LogWriterUtils.getLogWriter().info("Created back the partitioned regions"); // start the senders in async mode. This will ensure that the // node of shadow PR that went down last will come up first startSenderInVMsAsync("ln", vm4, vm5, vm6, vm7); LogWriterUtils.getLogWriter().info("Waiting for senders running."); // wait for senders running vm4.invoke(() -> WANTestBase.waitForSenderRunningState("ln")); vm5.invoke(() -> WANTestBase.waitForSenderRunningState("ln")); vm6.invoke(() -> WANTestBase.waitForSenderRunningState("ln")); vm7.invoke(() -> WANTestBase.waitForSenderRunningState("ln")); LogWriterUtils.getLogWriter().info("All the senders are now running..."); // ---------------------------------------------------------------------------------------------------- vm2.invoke(() -> WANTestBase.validateRegionSize(getTestMethodName() + "PR1", 3000)); vm3.invoke(() -> WANTestBase.validateRegionSize(getTestMethodName() + "PR1", 3000)); vm2.invoke(() -> WANTestBase.validateRegionSize(getTestMethodName() + "PR2", 5000)); vm3.invoke(() -> WANTestBase.validateRegionSize(getTestMethodName() + "PR2", 5000)); } public static void validateParallelSenderQueueAllBucketsDrained(final String senderId) { Set<GatewaySender> senders = cache.getGatewaySenders(); GatewaySender sender = null; for (GatewaySender s : senders) { if (s.getId().equals(senderId)) { sender = s; break; } } ConcurrentParallelGatewaySenderQueue regionQueue = (ConcurrentParallelGatewaySenderQueue) ((AbstractGatewaySender) sender).getQueues() .toArray(new RegionQueue[1])[0]; Set<PartitionedRegion> shadowPRs = (Set<PartitionedRegion>) regionQueue.getRegions(); for (PartitionedRegion shadowPR : shadowPRs) { Set<BucketRegion> buckets = shadowPR.getDataStore().getAllLocalBucketRegions(); for (final BucketRegion bucket : buckets) { WaitCriterion wc = new WaitCriterion() { public boolean done() { if (bucket.keySet().size() == 0) { LogWriterUtils.getLogWriter().info("Bucket " + bucket.getId() + " is empty"); return true; } return false; } public String description() { return "Expected bucket entries for bucket: " + bucket.getId() + " is: 0 but actual entries: " + bucket.keySet().size() + " This bucket isPrimary: " + bucket.getBucketAdvisor().isPrimary() + " KEYSET: " + bucket.keySet(); } }; Wait.waitForCriterion(wc, 180000, 50, true); } // for loop ends } } }