/** * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.hadoop.hbase.util; import java.io.FileNotFoundException; import java.io.IOException; import java.io.InterruptedIOException; import java.lang.reflect.Method; import java.net.InetSocketAddress; import java.net.URI; import java.util.HashSet; import java.util.Map; import java.util.Set; import java.util.Collection; import com.google.common.collect.Sets; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hdfs.DistributedFileSystem; import org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException; /** * Implementation for hdfs */ @InterfaceAudience.Private @InterfaceStability.Evolving public class FSHDFSUtils extends FSUtils { private static final Log LOG = LogFactory.getLog(FSHDFSUtils.class); private static Class dfsUtilClazz; private static Method getNNAddressesMethod; /** * @param fs * @param conf * @return A set containing all namenode addresses of fs */ private static Set<InetSocketAddress> getNNAddresses(DistributedFileSystem fs, Configuration conf) { Set<InetSocketAddress> addresses = new HashSet<>(); String serviceName = fs.getCanonicalServiceName(); if (serviceName.startsWith("ha-hdfs")) { try { if (dfsUtilClazz == null) { dfsUtilClazz = Class.forName("org.apache.hadoop.hdfs.DFSUtil"); } if (getNNAddressesMethod == null) { try { // getNNServiceRpcAddressesForCluster is available only in version // equal to or later than Hadoop 2.6 getNNAddressesMethod = dfsUtilClazz.getMethod("getNNServiceRpcAddressesForCluster", Configuration.class); } catch (NoSuchMethodException e) { // If hadoop version is older than hadoop 2.6 getNNAddressesMethod = dfsUtilClazz.getMethod("getNNServiceRpcAddresses", Configuration.class); } } Map<String, Map<String, InetSocketAddress>> addressMap = (Map<String, Map<String, InetSocketAddress>>) getNNAddressesMethod .invoke(null, conf); String nameService = serviceName.substring(serviceName.indexOf(":") + 1); if (addressMap.containsKey(nameService)) { Map<String, InetSocketAddress> nnMap = addressMap.get(nameService); for (Map.Entry<String, InetSocketAddress> e2 : nnMap.entrySet()) { InetSocketAddress addr = e2.getValue(); addresses.add(addr); } } } catch (Exception e) { LOG.warn("DFSUtil.getNNServiceRpcAddresses failed. serviceName=" + serviceName, e); } } else { URI uri = fs.getUri(); int port = uri.getPort(); if (port < 0) { int idx = serviceName.indexOf(':'); port = Integer.parseInt(serviceName.substring(idx+1)); } InetSocketAddress addr = new InetSocketAddress(uri.getHost(), port); addresses.add(addr); } return addresses; } /** * @param conf the Configuration of HBase * @param srcFs * @param desFs * @return Whether srcFs and desFs are on same hdfs or not */ public static boolean isSameHdfs(Configuration conf, FileSystem srcFs, FileSystem desFs) { // By getCanonicalServiceName, we could make sure both srcFs and desFs // show a unified format which contains scheme, host and port. String srcServiceName = srcFs.getCanonicalServiceName(); String desServiceName = desFs.getCanonicalServiceName(); if (srcServiceName == null || desServiceName == null) { return false; } if (srcServiceName.equals(desServiceName)) { return true; } if (srcServiceName.startsWith("ha-hdfs") && desServiceName.startsWith("ha-hdfs")) { Collection<String> internalNameServices = conf.getTrimmedStringCollection("dfs.internal.nameservices"); if (!internalNameServices.isEmpty()) { if (internalNameServices.contains(srcServiceName.split(":")[1])) { return true; } else { return false; } } } if (srcFs instanceof DistributedFileSystem && desFs instanceof DistributedFileSystem) { //If one serviceName is an HA format while the other is a non-HA format, // maybe they refer to the same FileSystem. //For example, srcFs is "ha-hdfs://nameservices" and desFs is "hdfs://activeNamenode:port" Set<InetSocketAddress> srcAddrs = getNNAddresses((DistributedFileSystem) srcFs, conf); Set<InetSocketAddress> desAddrs = getNNAddresses((DistributedFileSystem) desFs, conf); if (Sets.intersection(srcAddrs, desAddrs).size() > 0) { return true; } } return false; } /** * Recover the lease from HDFS, retrying multiple times. */ @Override public void recoverFileLease(final FileSystem fs, final Path p, Configuration conf, CancelableProgressable reporter) throws IOException { // lease recovery not needed for local file system case. if (!(fs instanceof DistributedFileSystem)) return; recoverDFSFileLease((DistributedFileSystem)fs, p, conf, reporter); } /* * Run the dfs recover lease. recoverLease is asynchronous. It returns: * -false when it starts the lease recovery (i.e. lease recovery not *yet* done) * - true when the lease recovery has succeeded or the file is closed. * But, we have to be careful. Each time we call recoverLease, it starts the recover lease * process over from the beginning. We could put ourselves in a situation where we are * doing nothing but starting a recovery, interrupting it to start again, and so on. * The findings over in HBASE-8354 have it that the namenode will try to recover the lease * on the file's primary node. If all is well, it should return near immediately. But, * as is common, it is the very primary node that has crashed and so the namenode will be * stuck waiting on a socket timeout before it will ask another datanode to start the * recovery. It does not help if we call recoverLease in the meantime and in particular, * subsequent to the socket timeout, a recoverLease invocation will cause us to start * over from square one (possibly waiting on socket timeout against primary node). So, * in the below, we do the following: * 1. Call recoverLease. * 2. If it returns true, break. * 3. If it returns false, wait a few seconds and then call it again. * 4. If it returns true, break. * 5. If it returns false, wait for what we think the datanode socket timeout is * (configurable) and then try again. * 6. If it returns true, break. * 7. If it returns false, repeat starting at step 5. above. * * If HDFS-4525 is available, call it every second and we might be able to exit early. */ boolean recoverDFSFileLease(final DistributedFileSystem dfs, final Path p, final Configuration conf, final CancelableProgressable reporter) throws IOException { LOG.info("Recover lease on dfs file " + p); long startWaiting = EnvironmentEdgeManager.currentTime(); // Default is 15 minutes. It's huge, but the idea is that if we have a major issue, HDFS // usually needs 10 minutes before marking the nodes as dead. So we're putting ourselves // beyond that limit 'to be safe'. long recoveryTimeout = conf.getInt("hbase.lease.recovery.timeout", 900000) + startWaiting; // This setting should be a little bit above what the cluster dfs heartbeat is set to. long firstPause = conf.getInt("hbase.lease.recovery.first.pause", 4000); // This should be set to how long it'll take for us to timeout against primary datanode if it // is dead. We set it to 61 seconds, 1 second than the default READ_TIMEOUT in HDFS, the // default value for DFS_CLIENT_SOCKET_TIMEOUT_KEY. If recovery is still failing after this // timeout, then further recovery will take liner backoff with this base, to avoid endless // preemptions when this value is not properly configured. long subsequentPauseBase = conf.getLong("hbase.lease.recovery.dfs.timeout", 61 * 1000); Method isFileClosedMeth = null; // whether we need to look for isFileClosed method boolean findIsFileClosedMeth = true; boolean recovered = false; // We break the loop if we succeed the lease recovery, timeout, or we throw an exception. for (int nbAttempt = 0; !recovered; nbAttempt++) { recovered = recoverLease(dfs, nbAttempt, p, startWaiting); if (recovered) break; checkIfCancelled(reporter); if (checkIfTimedout(conf, recoveryTimeout, nbAttempt, p, startWaiting)) break; try { // On the first time through wait the short 'firstPause'. if (nbAttempt == 0) { Thread.sleep(firstPause); } else { // Cycle here until (subsequentPause * nbAttempt) elapses. While spinning, check // isFileClosed if available (should be in hadoop 2.0.5... not in hadoop 1 though. long localStartWaiting = EnvironmentEdgeManager.currentTime(); while ((EnvironmentEdgeManager.currentTime() - localStartWaiting) < subsequentPauseBase * nbAttempt) { Thread.sleep(conf.getInt("hbase.lease.recovery.pause", 1000)); if (findIsFileClosedMeth) { try { isFileClosedMeth = dfs.getClass().getMethod("isFileClosed", new Class[]{ Path.class }); } catch (NoSuchMethodException nsme) { LOG.debug("isFileClosed not available"); } finally { findIsFileClosedMeth = false; } } if (isFileClosedMeth != null && isFileClosed(dfs, isFileClosedMeth, p)) { recovered = true; break; } checkIfCancelled(reporter); } } } catch (InterruptedException ie) { InterruptedIOException iioe = new InterruptedIOException(); iioe.initCause(ie); throw iioe; } } return recovered; } boolean checkIfTimedout(final Configuration conf, final long recoveryTimeout, final int nbAttempt, final Path p, final long startWaiting) { if (recoveryTimeout < EnvironmentEdgeManager.currentTime()) { LOG.warn("Cannot recoverLease after trying for " + conf.getInt("hbase.lease.recovery.timeout", 900000) + "ms (hbase.lease.recovery.timeout); continuing, but may be DATALOSS!!!; " + getLogMessageDetail(nbAttempt, p, startWaiting)); return true; } return false; } /** * Try to recover the lease. * @param dfs * @param nbAttempt * @param p * @param startWaiting * @return True if dfs#recoverLease came by true. * @throws FileNotFoundException */ boolean recoverLease(final DistributedFileSystem dfs, final int nbAttempt, final Path p, final long startWaiting) throws FileNotFoundException { boolean recovered = false; try { recovered = dfs.recoverLease(p); LOG.info((recovered? "Recovered lease, ": "Failed to recover lease, ") + getLogMessageDetail(nbAttempt, p, startWaiting)); } catch (IOException e) { if (e instanceof LeaseExpiredException && e.getMessage().contains("File does not exist")) { // This exception comes out instead of FNFE, fix it throw new FileNotFoundException("The given WAL wasn't found at " + p); } else if (e instanceof FileNotFoundException) { throw (FileNotFoundException)e; } LOG.warn(getLogMessageDetail(nbAttempt, p, startWaiting), e); } return recovered; } /** * @param nbAttempt * @param p * @param startWaiting * @return Detail to append to any log message around lease recovering. */ private String getLogMessageDetail(final int nbAttempt, final Path p, final long startWaiting) { return "attempt=" + nbAttempt + " on file=" + p + " after " + (EnvironmentEdgeManager.currentTime() - startWaiting) + "ms"; } /** * Call HDFS-4525 isFileClosed if it is available. * @param dfs * @param m * @param p * @return True if file is closed. */ private boolean isFileClosed(final DistributedFileSystem dfs, final Method m, final Path p) { try { return (Boolean) m.invoke(dfs, p); } catch (SecurityException e) { LOG.warn("No access", e); } catch (Exception e) { LOG.warn("Failed invocation for " + p.toString(), e); } return false; } void checkIfCancelled(final CancelableProgressable reporter) throws InterruptedIOException { if (reporter == null) return; if (!reporter.progress()) throw new InterruptedIOException("Operation cancelled"); } }