/*
* Copyright (c) 2016 Ericsson Systems, Inc. and others. All rights reserved.
*
* This program and the accompanying materials are made available under the
* terms of the Eclipse Public License v1.0 which accompanies this distribution,
* and is available at http://www.eclipse.org/legal/epl-v10.html
*/
package org.opendaylight.openflowplugin.applications.bulk.o.matic;
import com.google.common.util.concurrent.FutureCallback;
import com.google.common.util.concurrent.Futures;
import java.util.concurrent.ExecutorService;
import java.util.concurrent.atomic.AtomicInteger;
import java.util.concurrent.atomic.AtomicLong;
import org.opendaylight.controller.md.sal.binding.api.DataBroker;
import org.opendaylight.controller.md.sal.binding.api.WriteTransaction;
import org.opendaylight.controller.md.sal.common.api.data.LogicalDatastoreType;
import org.opendaylight.yang.gen.v1.urn.opendaylight.flow.inventory.rev130819.tables.table.Flow;
import org.opendaylight.yang.gen.v1.urn.opendaylight.flow.types.rev131026.flow.Match;
import org.opendaylight.yangtools.yang.binding.InstanceIdentifier;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
public class FlowWriterSequential implements FlowCounterMBean {
private static final Logger LOG = LoggerFactory.getLogger(FlowWriterSequential.class);
private final DataBroker dataBroker;
private final ExecutorService flowPusher;
protected int dpnCount;
private long startTime;
private AtomicInteger writeOpStatus = new AtomicInteger(FlowCounter.OperationStatus.INIT.status());
private AtomicInteger countDpnWriteCompletion = new AtomicInteger();
private AtomicLong taskCompletionTime = new AtomicLong();
public FlowWriterSequential(final DataBroker dataBroker, ExecutorService flowPusher) {
this.dataBroker = dataBroker;
this.flowPusher = flowPusher;
LOG.info("Using Sequential implementation of Flow Writer.");
}
public void addFlows(Integer dpnCount, Integer flowsPerDPN, int batchSize, int sleepMillis,
short startTableId, short endTableId, boolean isCreateParents) {
LOG.info("Using Sequential implementation of Flow Writer.");
this.dpnCount = dpnCount;
countDpnWriteCompletion.set(dpnCount);
startTime = System.nanoTime();
for (int i = 1; i <= dpnCount; i++) {
FlowHandlerTask task = new FlowHandlerTask(Integer.toString(i), flowsPerDPN, true, batchSize,
sleepMillis, startTableId, endTableId, isCreateParents);
flowPusher.execute(task);
}
}
public void deleteFlows(Integer dpnCount, Integer flowsPerDPN, int batchSize, short startTableId,
short endTableId) {
LOG.info("Using Sequential implementation of Flow Writer.");
countDpnWriteCompletion.set(dpnCount);
for (int i = 1; i <= dpnCount; i++) {
FlowHandlerTask task = new FlowHandlerTask(Integer.toString(i), flowsPerDPN, false, batchSize, 0,
startTableId, endTableId, false);
flowPusher.execute(task);
}
}
@Override
public int getWriteOpStatus() {
return writeOpStatus.get();
}
@Override
public long getTaskCompletionTime() {
return taskCompletionTime.get();
}
private class FlowHandlerTask implements Runnable {
private final String dpId;
private final int flowsPerDpn;
private final boolean add;
private final int batchSize;
private final int sleepMillis;
private final short startTableId;
private final short endTableId;
private final boolean isCreateParents;
public FlowHandlerTask(final String dpId,
final int flowsPerDpn,
final boolean add,
final int batchSize,
int sleepMillis,
final short startTableId,
final short endTableId,
final boolean isCreateParents){
this.dpId = BulkOMaticUtils.DEVICE_TYPE_PREFIX + dpId;
this.add = add;
this.flowsPerDpn = flowsPerDpn;
this.batchSize = batchSize;
this.sleepMillis = sleepMillis;
this.startTableId = startTableId;
this.endTableId = endTableId;
this.isCreateParents = isCreateParents;
}
@Override
public void run() {
LOG.info("Starting flow writer task for dpid: {}. Number of transactions: {}", dpId, flowsPerDpn/batchSize);
writeOpStatus.set(FlowCounter.OperationStatus.IN_PROGRESS.status());
Short tableId = startTableId;
Integer sourceIp = 1;
WriteTransaction writeTransaction = dataBroker.newWriteOnlyTransaction();
short k = tableId;
for (; sourceIp <= batchSize; sourceIp++) {
String flowId = "Flow-" + dpId + "." + k + "." + sourceIp;
LOG.debug("Adding flow with id: {}", flowId);
Flow flow = null;
if (add) {
Match match = BulkOMaticUtils.getMatch(sourceIp);
flow = BulkOMaticUtils.buildFlow(k, flowId, match);
}
addFlowToTx(writeTransaction, flowId,
BulkOMaticUtils.getFlowInstanceIdentifier(k, flowId, dpId), flow);
if (sourceIp < batchSize) {
short a = 1;
short b = (short)(endTableId - startTableId + 1);
k = (short) (((k + a) % b) + startTableId);
}
}
LOG.debug("Submitting Txn for dpId: {}, begin tableId: {}, end tableId: {}, sourceIp: {}", dpId, tableId, k, sourceIp);
Futures.addCallback(writeTransaction.submit(), new DsCallBack(dpId, sourceIp, k));
}
private void addFlowToTx(WriteTransaction writeTransaction, String flowId, InstanceIdentifier<Flow> flowIid,
Flow flow) {
if (add) {
LOG.trace("Adding flow for flowId: {}, flowIid: {}", flowId, flowIid);
writeTransaction.put(LogicalDatastoreType.CONFIGURATION, flowIid, flow, isCreateParents);
} else {
LOG.trace("Deleting flow for flowId: {}, flowIid: {}", flowId, flowIid);
writeTransaction.delete(LogicalDatastoreType.CONFIGURATION, flowIid);
}
}
private class DsCallBack implements FutureCallback {
private String dpId;
private Integer sourceIp;
private Short tableId;
public DsCallBack(String dpId, Integer sourceIp, Short tableId) {
this.dpId = dpId;
this.sourceIp = sourceIp;
short a = 1;
short b = (short)(endTableId - startTableId + 1);
this.tableId = (short) (((tableId + a) % b) + startTableId);
}
@Override
public void onSuccess(Object o) {
if (sourceIp > flowsPerDpn) {
long dur = System.nanoTime() - startTime;
LOG.info("Completed all flows installation for: dpid: {}, tableId: {}, sourceIp: {} in {}ns", dpId,
tableId, sourceIp, dur);
if(0 == countDpnWriteCompletion.decrementAndGet() &&
writeOpStatus.get() != FlowCounter.OperationStatus.FAILURE.status()) {
writeOpStatus.set(FlowCounter.OperationStatus.SUCCESS.status());
taskCompletionTime.set(dur);
}
return;
}
try {
if (sleepMillis > 0) {
Thread.sleep(sleepMillis);
}
} catch (InterruptedException e) {
LOG.error("Writer Thread Interrupted while sleeping: {}", e.getMessage());
}
WriteTransaction writeTransaction = dataBroker.newWriteOnlyTransaction();
int newBatchSize = sourceIp + batchSize - 1;
short k = tableId;
for (; sourceIp <= newBatchSize; sourceIp++) {
String flowId = "Flow-" + dpId + "." + k + "." + sourceIp;
Flow flow = null;
if (add) {
Match match = BulkOMaticUtils.getMatch(sourceIp);
flow = BulkOMaticUtils.buildFlow(k, flowId, match);
}
LOG.debug("Adding flow with id: {}", flowId);
addFlowToTx(writeTransaction, flowId,
BulkOMaticUtils.getFlowInstanceIdentifier(k, flowId, dpId), flow);
if (sourceIp < newBatchSize) {
short a = 1;
short b = (short)(endTableId - startTableId + 1);
k = (short) (((k + a) % b) + startTableId);
}
}
LOG.debug("OnSuccess: Submitting Txn for dpId: {}, begin tableId: {}, end tableId: {}, sourceIp: {}",
dpId, tableId, k, sourceIp);
Futures.addCallback(writeTransaction.submit(), new DsCallBack(dpId, sourceIp, k));
}
public void onFailure(Throwable error) {
LOG.error("Error: {} in Datastore write operation: dpid: {}, tableId: {}, sourceIp: {}",
error, dpId, tableId, sourceIp);
writeOpStatus.set(FlowCounter.OperationStatus.FAILURE.status());
}
}
}
}