/** * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.hadoop.mapred; import java.io.File; import java.io.IOException; import java.util.Collections; import java.util.HashMap; import java.util.Iterator; import java.util.Map; import java.util.Map.Entry; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.mapred.CleanupQueue.PathDeletionContext; import org.apache.hadoop.mapreduce.JobContext; import org.apache.hadoop.mapreduce.JobID; import org.apache.hadoop.mapreduce.server.tasktracker.userlogs.*; /** * This is used only in UserLogManager, to manage cleanup of user logs. */ public class UserLogCleaner extends Thread { private static final Log LOG = LogFactory.getLog(UserLogCleaner.class); static final String USERLOGCLEANUP_SLEEPTIME = "mapreduce.tasktracker.userlogcleanup.sleeptime"; static final int DEFAULT_USER_LOG_RETAIN_HOURS = 24; // 1 day static final long DEFAULT_THREAD_SLEEP_TIME = 1000 * 60 * 60; // 1 hour private UserLogManager userLogManager; private Map<JobID, Long> completedJobs = Collections .synchronizedMap(new HashMap<JobID, Long>()); private final long threadSleepTime; private CleanupQueue cleanupQueue; private Clock clock; private FileSystem localFs; public UserLogCleaner(UserLogManager userLogManager, Configuration conf) throws IOException { this.userLogManager = userLogManager; threadSleepTime = conf.getLong(USERLOGCLEANUP_SLEEPTIME, DEFAULT_THREAD_SLEEP_TIME); cleanupQueue = new CleanupQueue(); localFs = FileSystem.getLocal(conf); setClock(new Clock()); setDaemon(true); } void setClock(Clock clock) { this.clock = clock; } Clock getClock() { return this.clock; } CleanupQueue getCleanupQueue() { return cleanupQueue; } void setCleanupQueue(CleanupQueue cleanupQueue) { this.cleanupQueue = cleanupQueue; } @Override public void run() { // This thread wakes up after every threadSleepTime interval // and deletes if there are any old logs. while (true) { try { // sleep Thread.sleep(threadSleepTime); processCompletedJobs(); } catch (Throwable e) { LOG.warn(getClass().getSimpleName() + " encountered an exception while monitoring :", e); LOG.info("Ingoring the exception and continuing monitoring."); } } } void processCompletedJobs() throws IOException { long now = clock.getTime(); // iterate through completedJobs and remove old logs. synchronized (completedJobs) { Iterator<Entry<JobID, Long>> completedJobIter = completedJobs.entrySet() .iterator(); while (completedJobIter.hasNext()) { Entry<JobID, Long> entry = completedJobIter.next(); // see if the job is old enough if (entry.getValue().longValue() <= now) { // add the job for deletion userLogManager.addLogEvent(new DeleteJobEvent(entry.getKey())); completedJobIter.remove(); } } } } public void deleteJobLogs(JobID jobid) throws IOException { deleteLogPath(TaskLog.getJobDir(jobid).getAbsolutePath()); } /** * Clears all the logs in userlog directory. * * Adds the job directories for deletion with default retain hours. Deletes * all other directories, if any. This is usually called on reinit/restart of * the TaskTracker * * @param conf * @throws IOException */ public void clearOldUserLogs(Configuration conf) throws IOException { File userLogDir = TaskLog.getUserLogDir(); if (userLogDir.exists()) { String[] logDirs = userLogDir.list(); if (logDirs.length > 0) { // add all the log dirs to taskLogsMnonitor. long now = clock.getTime(); for (String logDir : logDirs) { JobID jobid = null; try { jobid = JobID.forName(logDir); } catch (IllegalArgumentException ie) { // if the directory is not a jobid, delete it immediately deleteLogPath(new File(userLogDir, logDir).getAbsolutePath()); continue; } // add the job log directory for deletion with default retain hours, // if it is not already added if (!completedJobs.containsKey(jobid)) { JobCompletedEvent jce = new JobCompletedEvent(jobid, now, getUserlogRetainHours(conf)); userLogManager.addLogEvent(jce); } } } } } /** * If the configuration is null or user-log retain hours is not configured, * the retain hours are {@value UserLogCleaner#DEFAULT_USER_LOG_RETAIN_HOURS} */ static int getUserlogRetainHours(Configuration conf) { return (conf == null ? DEFAULT_USER_LOG_RETAIN_HOURS : conf.getInt( JobContext.USER_LOG_RETAIN_HOURS, DEFAULT_USER_LOG_RETAIN_HOURS)); } /** * Adds job user-log directory to cleanup thread to delete logs after user-log * retain hours. * * @param jobCompletionTime * job completion time in millis * @param retainHours * the user-log retain hours for the job * @param jobid * JobID for which user logs should be deleted */ public void markJobLogsForDeletion(long jobCompletionTime, int retainHours, org.apache.hadoop.mapreduce.JobID jobid) { long retainTimeStamp = jobCompletionTime + (retainHours * 1000L * 60L * 60L); LOG.info("Adding " + jobid + " for user-log deletion with retainTimeStamp:" + retainTimeStamp); completedJobs.put(jobid, Long.valueOf(retainTimeStamp)); } /** * Remove job from user log deletion. * * @param jobid */ public void unmarkJobFromLogDeletion(JobID jobid) { if (completedJobs.remove(jobid) != null) { LOG.info("Removing " + jobid + " from user-log deletion"); } } /** * Deletes the log path. * * This path will be removed through {@link CleanupQueue} * * @param logPath * @throws IOException */ private void deleteLogPath(String logPath) throws IOException { LOG.info("Deleting user log path " + logPath); PathDeletionContext context = new PathDeletionContext(localFs, logPath); cleanupQueue.addToQueue(context); } }