/** * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package com.alibaba.jstorm.daemon.worker; import backtype.storm.generated.StormTopology; import backtype.storm.task.TopologyContext; import backtype.storm.task.WorkerTopologyContext; import backtype.storm.tuple.Fields; import com.alibaba.jstorm.cluster.StormConfig; import com.alibaba.jstorm.utils.JStormUtils; import com.alibaba.jstorm.utils.PathUtils; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import java.io.IOException; import java.util.HashMap; import java.util.List; import java.util.Map; /** * This class is used to create TopologyContext * * @author yannian/Longda */ public class ContextMaker { private static Logger LOG = LoggerFactory.getLogger(ContextMaker.class); private WorkerData workerData; private String resourcePath; private String workerId; private String workerIdDir; private String pidDir; private String codeDir; private List<Integer> workerTasks; @SuppressWarnings("rawtypes") public ContextMaker(WorkerData workerData) { this.workerData = workerData; this.workerTasks = JStormUtils.mk_list(workerData.getTaskIds()); try { Map stormConf = workerData.getStormConf(); String topologyId = workerData.getTopologyId(); workerId = workerData.getWorkerId(); String distRoot = StormConfig.supervisor_stormdist_root(workerData.getConf(), topologyId); resourcePath = StormConfig.supervisor_storm_resources_path(distRoot); workerIdDir = StormConfig.worker_root(stormConf, workerId); pidDir = StormConfig.worker_pids_root(stormConf, workerId); String codePath = StormConfig.stormcode_path(distRoot); codeDir = PathUtils.parent_path(codePath); } catch (IOException e) { LOG.error("Failed to create ContextMaker", e); throw new RuntimeException(e); } } public TopologyContext makeTopologyContext(StormTopology topology, Integer taskId, clojure.lang.Atom openOrPrepareWasCalled) { Map stormConf = new HashMap(); stormConf.putAll(workerData.getStormConf()); String topologyId = workerData.getTopologyId(); HashMap<String, Map<String, Fields>> componentToStreamToFields = workerData.generateComponentToStreamToFields(topology); return new TopologyContext(topology, stormConf, workerData.getTasksToComponent(), workerData.getComponentToSortedTasks(), componentToStreamToFields, topologyId, resourcePath, workerId, taskId, workerData.getPort(), workerTasks, workerData.getDefaultResources(), workerData.getUserResources(), workerData.getExecutorData(), workerData.getRegisteredMetrics(), openOrPrepareWasCalled, workerData.getZkCluster()); } public WorkerTopologyContext makeWorkerTopologyContext(StormTopology topology) { Map stormConf = workerData.getStormConf(); String topologyId = workerData.getTopologyId(); HashMap<String, Map<String, Fields>> componentToStreamToFields = workerData.generateComponentToStreamToFields(topology); return new WorkerTopologyContext(topology, stormConf, workerData.getTasksToComponent(), workerData.getComponentToSortedTasks(), componentToStreamToFields, topologyId, resourcePath, workerId, workerData.getPort(), workerTasks, workerData.getDefaultResources(), workerData.getUserResources()); } }