/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.accumulo.test.replication;
import static org.easymock.EasyMock.createMock;
import static org.easymock.EasyMock.expectLastCall;
import static org.easymock.EasyMock.replay;
import static org.easymock.EasyMock.verify;
import java.util.HashSet;
import java.util.Set;
import java.util.UUID;
import org.apache.accumulo.core.client.BatchWriter;
import org.apache.accumulo.core.client.Connector;
import org.apache.accumulo.core.data.Mutation;
import org.apache.accumulo.core.protobuf.ProtobufUtil;
import org.apache.accumulo.core.replication.ReplicationSchema.OrderSection;
import org.apache.accumulo.core.replication.ReplicationSchema.WorkSection;
import org.apache.accumulo.core.replication.ReplicationTable;
import org.apache.accumulo.core.replication.ReplicationTarget;
import org.apache.accumulo.core.security.TablePermission;
import org.apache.accumulo.master.replication.UnorderedWorkAssigner;
import org.apache.accumulo.server.replication.DistributedWorkQueueWorkAssignerHelper;
import org.apache.accumulo.server.replication.StatusUtil;
import org.apache.accumulo.server.replication.proto.Replication.Status;
import org.apache.accumulo.server.zookeeper.DistributedWorkQueue;
import org.apache.accumulo.server.zookeeper.ZooCache;
import org.apache.accumulo.test.functional.ConfigurableMacBase;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.Text;
import org.junit.Before;
import org.junit.Test;
public class UnorderedWorkAssignerIT extends ConfigurableMacBase {
private Connector conn;
private MockUnorderedWorkAssigner assigner;
private static class MockUnorderedWorkAssigner extends UnorderedWorkAssigner {
public MockUnorderedWorkAssigner(Connector conn) {
super(null, conn);
}
@Override
protected void setQueuedWork(Set<String> queuedWork) {
super.setQueuedWork(queuedWork);
}
@Override
protected void setWorkQueue(DistributedWorkQueue workQueue) {
super.setWorkQueue(workQueue);
}
@Override
protected boolean queueWork(Path path, ReplicationTarget target) {
return super.queueWork(path, target);
}
@Override
protected void initializeQueuedWork() {
super.initializeQueuedWork();
}
@Override
protected Set<String> getQueuedWork() {
return super.getQueuedWork();
}
@Override
protected void setConnector(Connector conn) {
super.setConnector(conn);
}
@Override
protected void setMaxQueueSize(int maxQueueSize) {
super.setMaxQueueSize(maxQueueSize);
}
@Override
protected void createWork() {
super.createWork();
}
@Override
protected void setZooCache(ZooCache zooCache) {
super.setZooCache(zooCache);
}
@Override
protected void cleanupFinishedWork() {
super.cleanupFinishedWork();
}
}
@Before
public void init() throws Exception {
conn = getConnector();
assigner = new MockUnorderedWorkAssigner(conn);
ReplicationTable.setOnline(conn);
conn.securityOperations().grantTablePermission(conn.whoami(), ReplicationTable.NAME, TablePermission.WRITE);
conn.securityOperations().grantTablePermission(conn.whoami(), ReplicationTable.NAME, TablePermission.READ);
}
@Test
public void createWorkForFilesNeedingIt() throws Exception {
ReplicationTarget target1 = new ReplicationTarget("cluster1", "table1", "1"), target2 = new ReplicationTarget("cluster1", "table2", "2");
Text serializedTarget1 = target1.toText(), serializedTarget2 = target2.toText();
String keyTarget1 = target1.getPeerName() + DistributedWorkQueueWorkAssignerHelper.KEY_SEPARATOR + target1.getRemoteIdentifier()
+ DistributedWorkQueueWorkAssignerHelper.KEY_SEPARATOR + target1.getSourceTableId(), keyTarget2 = target2.getPeerName()
+ DistributedWorkQueueWorkAssignerHelper.KEY_SEPARATOR + target2.getRemoteIdentifier() + DistributedWorkQueueWorkAssignerHelper.KEY_SEPARATOR
+ target2.getSourceTableId();
Status.Builder builder = Status.newBuilder().setBegin(0).setEnd(0).setInfiniteEnd(true).setClosed(false).setCreatedTime(5l);
Status status1 = builder.build();
builder.setCreatedTime(10l);
Status status2 = builder.build();
// Create two mutations, both of which need replication work done
BatchWriter bw = ReplicationTable.getBatchWriter(conn);
String filename1 = UUID.randomUUID().toString(), filename2 = UUID.randomUUID().toString();
String file1 = "/accumulo/wal/tserver+port/" + filename1, file2 = "/accumulo/wal/tserver+port/" + filename2;
Mutation m = new Mutation(file1);
WorkSection.add(m, serializedTarget1, ProtobufUtil.toValue(status1));
bw.addMutation(m);
m = OrderSection.createMutation(file1, status1.getCreatedTime());
OrderSection.add(m, target1.getSourceTableId(), ProtobufUtil.toValue(status1));
bw.addMutation(m);
m = new Mutation(file2);
WorkSection.add(m, serializedTarget2, ProtobufUtil.toValue(status2));
bw.addMutation(m);
m = OrderSection.createMutation(file2, status2.getCreatedTime());
OrderSection.add(m, target2.getSourceTableId(), ProtobufUtil.toValue(status2));
bw.addMutation(m);
bw.close();
DistributedWorkQueue workQueue = createMock(DistributedWorkQueue.class);
HashSet<String> queuedWork = new HashSet<>();
assigner.setQueuedWork(queuedWork);
assigner.setWorkQueue(workQueue);
assigner.setMaxQueueSize(Integer.MAX_VALUE);
// Make sure we expect the invocations in the order they were created
String key = filename1 + "|" + keyTarget1;
workQueue.addWork(key, file1);
expectLastCall().once();
key = filename2 + "|" + keyTarget2;
workQueue.addWork(key, file2);
expectLastCall().once();
replay(workQueue);
assigner.createWork();
verify(workQueue);
}
@Test
public void doNotCreateWorkForFilesNotNeedingIt() throws Exception {
ReplicationTarget target1 = new ReplicationTarget("cluster1", "table1", "1"), target2 = new ReplicationTarget("cluster1", "table2", "2");
Text serializedTarget1 = target1.toText(), serializedTarget2 = target2.toText();
// Create two mutations, both of which need replication work done
BatchWriter bw = ReplicationTable.getBatchWriter(conn);
String filename1 = UUID.randomUUID().toString(), filename2 = UUID.randomUUID().toString();
String file1 = "/accumulo/wal/tserver+port/" + filename1, file2 = "/accumulo/wal/tserver+port/" + filename2;
Mutation m = new Mutation(file1);
WorkSection.add(m, serializedTarget1, StatusUtil.fileCreatedValue(5));
bw.addMutation(m);
m = new Mutation(file2);
WorkSection.add(m, serializedTarget2, StatusUtil.fileCreatedValue(10));
bw.addMutation(m);
bw.close();
DistributedWorkQueue workQueue = createMock(DistributedWorkQueue.class);
HashSet<String> queuedWork = new HashSet<>();
assigner.setQueuedWork(queuedWork);
assigner.setMaxQueueSize(Integer.MAX_VALUE);
replay(workQueue);
assigner.createWork();
verify(workQueue);
}
@Test
public void workNotReAdded() throws Exception {
Set<String> queuedWork = new HashSet<>();
assigner.setQueuedWork(queuedWork);
ReplicationTarget target = new ReplicationTarget("cluster1", "table1", "1");
String serializedTarget = target.getPeerName() + DistributedWorkQueueWorkAssignerHelper.KEY_SEPARATOR + target.getRemoteIdentifier()
+ DistributedWorkQueueWorkAssignerHelper.KEY_SEPARATOR + target.getSourceTableId();
queuedWork.add("wal1|" + serializedTarget.toString());
// Create two mutations, both of which need replication work done
BatchWriter bw = ReplicationTable.getBatchWriter(conn);
String file1 = "/accumulo/wal/tserver+port/wal1";
Mutation m = new Mutation(file1);
WorkSection.add(m, target.toText(), StatusUtil.openWithUnknownLengthValue());
bw.addMutation(m);
bw.close();
DistributedWorkQueue workQueue = createMock(DistributedWorkQueue.class);
assigner.setWorkQueue(workQueue);
assigner.setMaxQueueSize(Integer.MAX_VALUE);
replay(workQueue);
assigner.createWork();
verify(workQueue);
}
}