/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.ambari.server.checks; import java.util.ArrayList; import java.util.Arrays; import java.util.List; import java.util.Map; import org.apache.ambari.server.AmbariException; import org.apache.ambari.server.controller.PrereqCheckRequest; import org.apache.ambari.server.state.Cluster; import org.apache.ambari.server.state.Config; import org.apache.ambari.server.state.DesiredConfig; import org.apache.ambari.server.state.stack.PrereqCheckStatus; import org.apache.ambari.server.state.stack.PrerequisiteCheck; import org.apache.ambari.server.state.stack.UpgradePack.PrerequisiteCheckConfig; import org.apache.commons.lang.StringUtils; import com.google.inject.Singleton; /** * Checks that Tez jobs reference hadoop libraries from the distributed cache. */ @Singleton @UpgradeCheck(group = UpgradeCheckGroup.NAMENODE_HA, order = 21.0f) public class ServicesTezDistributedCacheCheck extends AbstractCheckDescriptor { static final String KEY_LIB_URI_MISSING = "tez_lib_uri_missing"; static final String KEY_USE_HADOOP_LIBS = "tez_use_hadoop_libs"; static final String KEY_LIB_NOT_DFS = "lib_not_dfs"; static final String KEY_LIB_NOT_TARGZ = "lib_not_targz"; static final String KEY_USE_HADOOP_LIBS_FALSE = "tez_use_hadoop_libs_false"; static final String DFS_PROTOCOLS_REGEX_PROPERTY_NAME = "dfs-protocols-regex"; static final String DFS_PROTOCOLS_REGEX_DEFAULT = "^([^:]*dfs|wasb|ecs):.*"; @Override public boolean isApplicable(PrereqCheckRequest request) throws AmbariException { if (!super.isApplicable(request, Arrays.asList("TEZ"), true)) { return false; } PrereqCheckStatus ha = request.getResult(CheckDescription.SERVICES_NAMENODE_HA); if (null != ha && ha == PrereqCheckStatus.FAIL) { return false; } return true; } /** * Constructor. */ public ServicesTezDistributedCacheCheck() { super(CheckDescription.SERVICES_TEZ_DISTRIBUTED_CACHE); } @Override public void perform(PrerequisiteCheck prerequisiteCheck, PrereqCheckRequest request) throws AmbariException { String dfsProtocolsRegex = DFS_PROTOCOLS_REGEX_DEFAULT; PrerequisiteCheckConfig prerequisiteCheckConfig = request.getPrerequisiteCheckConfig(); Map<String, String> checkProperties = null; if(prerequisiteCheckConfig != null) { checkProperties = prerequisiteCheckConfig.getCheckProperties(this.getClass().getName()); } if(checkProperties != null && checkProperties.containsKey(DFS_PROTOCOLS_REGEX_PROPERTY_NAME)) { dfsProtocolsRegex = checkProperties.get(DFS_PROTOCOLS_REGEX_PROPERTY_NAME); } final String clusterName = request.getClusterName(); final Cluster cluster = clustersProvider.get().getCluster(clusterName); final String tezConfigType = "tez-site"; final String coreSiteConfigType = "core-site"; final Map<String, DesiredConfig> desiredConfigs = cluster.getDesiredConfigs(); final DesiredConfig tezDesiredConfig = desiredConfigs.get(tezConfigType); final Config tezConfig = cluster.getConfig(tezConfigType, tezDesiredConfig.getTag()); final DesiredConfig coreSiteDesiredConfig = desiredConfigs.get(coreSiteConfigType); final Config coreSiteConfig = cluster.getConfig(coreSiteConfigType, coreSiteDesiredConfig.getTag()); final String libUris = tezConfig.getProperties().get("tez.lib.uris"); final String useHadoopLibs = tezConfig.getProperties().get("tez.use.cluster.hadoop-libs"); final String defaultFS = coreSiteConfig.getProperties().get("fs.defaultFS"); List<String> errorMessages = new ArrayList<>(); if (libUris == null || libUris.isEmpty()) { errorMessages.add(getFailReason(KEY_LIB_URI_MISSING, prerequisiteCheck, request)); } if (useHadoopLibs == null || useHadoopLibs.isEmpty()) { errorMessages.add(getFailReason(KEY_USE_HADOOP_LIBS, prerequisiteCheck, request)); } if (!errorMessages.isEmpty()) { prerequisiteCheck.getFailedOn().add("TEZ"); prerequisiteCheck.setStatus(PrereqCheckStatus.FAIL); prerequisiteCheck.setFailReason(StringUtils.join(errorMessages, " ")); return; } if (!libUris.matches(dfsProtocolsRegex) && (defaultFS == null || !defaultFS.matches(dfsProtocolsRegex))) { errorMessages.add(getFailReason(KEY_LIB_NOT_DFS, prerequisiteCheck, request)); } if (!libUris.contains("tar.gz")) { errorMessages.add(getFailReason(KEY_LIB_NOT_TARGZ, prerequisiteCheck, request)); } if (Boolean.parseBoolean(useHadoopLibs)) { errorMessages.add(getFailReason(KEY_USE_HADOOP_LIBS_FALSE, prerequisiteCheck, request)); } if (!errorMessages.isEmpty()) { prerequisiteCheck.getFailedOn().add("TEZ"); prerequisiteCheck.setStatus(PrereqCheckStatus.FAIL); prerequisiteCheck.setFailReason(StringUtils.join(errorMessages, " ")); } } }