/** * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.falcon.entity; import org.apache.falcon.FalconException; import org.apache.falcon.entity.store.ConfigurationStore; import org.apache.falcon.entity.v0.EntityType; import org.apache.falcon.entity.v0.cluster.*; import org.apache.falcon.hadoop.HadoopClientFactory; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import java.util.HashMap; import java.util.List; import java.util.Map; /** * Helper to get end points relating to the cluster. */ public final class ClusterHelper { public static final String DEFAULT_BROKER_IMPL_CLASS = "org.apache.activemq.ActiveMQConnectionFactory"; private ClusterHelper() { } public static Cluster getCluster(String cluster) throws FalconException { return ConfigurationStore.get().get(EntityType.CLUSTER, cluster); } public static FileSystem getFileSystem(String cluster) throws FalconException { Cluster clusterEntity = getCluster(cluster); Configuration conf = ClusterHelper.getConfiguration(clusterEntity); return HadoopClientFactory.get().createProxiedFileSystem(conf); } public static Configuration getConfiguration(Cluster cluster) { Configuration conf = new Configuration(); final String storageUrl = getStorageUrl(cluster); conf.set(HadoopClientFactory.FS_DEFAULT_NAME_KEY, storageUrl); final String executeEndPoint = getMREndPoint(cluster); conf.set(HadoopClientFactory.MR_JOB_TRACKER_KEY, executeEndPoint); conf.set(HadoopClientFactory.YARN_RM_ADDRESS_KEY, executeEndPoint); if (cluster.getProperties() != null) { for (Property prop : cluster.getProperties().getProperties()) { conf.set(prop.getName(), prop.getValue()); } } return conf; } public static String getOozieUrl(Cluster cluster) { return getInterface(cluster, Interfacetype.WORKFLOW).getEndpoint(); } public static String getStorageUrl(Cluster cluster) { return getNormalizedUrl(cluster, Interfacetype.WRITE); } public static String getReadOnlyStorageUrl(Cluster cluster) { return getNormalizedUrl(cluster, Interfacetype.READONLY); } public static String getMREndPoint(Cluster cluster) { return getInterface(cluster, Interfacetype.EXECUTE).getEndpoint(); } public static String getRegistryEndPoint(Cluster cluster) { final Interface catalogInterface = getInterface(cluster, Interfacetype.REGISTRY); return catalogInterface == null ? null : catalogInterface.getEndpoint(); } public static String getMessageBrokerUrl(Cluster cluster) { return getInterface(cluster, Interfacetype.MESSAGING).getEndpoint(); } public static String getMessageBrokerImplClass(Cluster cluster) { if (cluster.getProperties() != null) { for (Property prop : cluster.getProperties().getProperties()) { if (prop.getName().equals("brokerImplClass")) { return prop.getValue(); } } } return DEFAULT_BROKER_IMPL_CLASS; } public static Interface getInterface(Cluster cluster, Interfacetype type) { for (Interface interf : cluster.getInterfaces().getInterfaces()) { if (interf.getType() == type) { return interf; } } return null; } private static String getNormalizedUrl(Cluster cluster, Interfacetype type) { String normalizedUrl = getInterface(cluster, type).getEndpoint(); String normalizedPath = new Path(normalizedUrl + "/").toString(); return normalizedPath.substring(0, normalizedPath.length() - 1); } public static String getCompleteLocation(Cluster cluster, String locationKey) { return getStorageUrl(cluster) + "/" + getLocation(cluster, locationKey); } public static String getLocation(Cluster cluster, String locationKey) { for (Location loc : cluster.getLocations().getLocations()) { if (loc.getName().equals(locationKey)) { return loc.getPath(); } } return null; } public static String getPropertyValue(Cluster cluster, String propName) { if (cluster.getProperties() != null) { for (Property prop : cluster.getProperties().getProperties()) { if (prop.getName().equals(propName)) { return prop.getValue(); } } } return null; } public static Map<String, String> getHiveProperties(Cluster cluster) { if (cluster.getProperties() != null) { List<Property> properties = cluster.getProperties().getProperties(); if (properties != null && !properties.isEmpty()) { Map<String, String> hiveProperties = new HashMap<String, String>(); for (Property prop : properties) { if (prop.getName().startsWith("hive.")) { hiveProperties.put(prop.getName(), prop.getValue()); } } return hiveProperties; } } return null; } }