/*
* Copyright (c) 2011-2015 EPFL DATA Laboratory
* Copyright (c) 2014-2015 The Squall Collaboration (see NOTICE)
*
* All rights reserved.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package ch.epfl.data.squall.storm_components.hybrid_hypercube;
import java.util.ArrayList;
import java.util.List;
import java.util.Map;
import org.apache.log4j.Logger;
import backtype.storm.generated.GlobalStreamId;
import backtype.storm.grouping.CustomStreamGrouping;
import backtype.storm.task.WorkerTopologyContext;
import ch.epfl.data.squall.thetajoin.matrix_assignment.HybridHyperCubeAssignment;
import ch.epfl.data.squall.storm_components.hash_hypercube.HashHyperCubeGrouping.EmitterDesc;
import ch.epfl.data.squall.utilities.MyUtilities;
import org.apache.log4j.Logger;
import java.util.ArrayList;
import java.util.List;
import java.util.Map;
import java.util.HashMap;
public class HybridHyperCubeGrouping implements CustomStreamGrouping {
private static final long serialVersionUID = 1L;
private static Logger LOG = Logger
.getLogger(HybridHyperCubeGrouping.class);
private final HybridHyperCubeAssignment _assignment;
private final List<EmitterDesc> _emitters;
private List<Integer> _targetTasks;
private final Map _map;
public HybridHyperCubeGrouping(List<EmitterDesc> emitters, HybridHyperCubeAssignment assignment, Map map) {
_emitters = emitters;
_assignment = assignment;
_map = map;
}
// @Override
@Override
public List<Integer> chooseTasks(int taskId, List<Object> stormTuple) {
// the following is checking for FinalAck
if (!MyUtilities.isManualBatchingMode(_map)) {
final List<String> tuple = (List<String>) stormTuple.get(1); // TUPLE
if (MyUtilities.isFinalAck(tuple, _map))
return _targetTasks;
} else {
final String tupleBatch = (String) stormTuple.get(1); // TUPLE
if (MyUtilities.isFinalAckManualBatching(tupleBatch, _map))
// send to everyone
return _targetTasks;
}
// the following picks tasks for Non-FinalAck
return chooseTasksNonFinalAck(stormTuple);
}
private List<Integer> chooseTasksNonFinalAck(List<Object> stormTuple) {
List<Integer> tasks = null;
final String tableName = (String) stormTuple.get(0);
final List<String> tuple = (List<String>) stormTuple.get(1); // TUPLE
for (EmitterDesc emitter : _emitters) {
if (emitter.index.equals(tableName)) {
List<Integer> regionsID = _assignment.getRegionIDs(emitter.name, createColumns(emitter, tuple));
tasks = translateIdsToTasks(regionsID);
break;
}
}
if (tasks == null) {
LOG.info(_emitters);
LOG.info("Table Name: " + tableName);
LOG.info("WRONG ASSIGNMENT");
}
return tasks;
}
// Creates <ColumnName, Value>
public Map<String, String> createColumns(EmitterDesc emitter, List<String> stormTuple) {
HashMap<String, String> columns = new HashMap<String, String>();
int i = 0;
for (String columnName : emitter.columnNames) {
columns.put(columnName, stormTuple.get(i++));
}
return columns;
}
@Override
public void prepare(WorkerTopologyContext wtc, GlobalStreamId gsi,
List<Integer> targetTasks) {
_targetTasks = targetTasks;
}
private List<Integer> translateIdsToTasks(List<Integer> ids) {
final List<Integer> converted = new ArrayList<Integer>();
for (final int id : ids)
converted.add(_targetTasks.get(id));
return converted;
}
}