/** * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.hadoop.hdfs.server.namenode.snapshot; import static org.junit.Assert.*; import java.io.File; import java.io.IOException; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileUtil; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hdfs.DFSConfigKeys; import org.apache.hadoop.hdfs.HdfsConfiguration; import org.apache.hadoop.hdfs.MiniDFSCluster; import org.apache.hadoop.hdfs.client.HdfsAdmin; import org.apache.hadoop.hdfs.server.namenode.NameNodeAdapter; import org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode; import org.junit.Before; import org.junit.Test; public class TestCheckpointsWithSnapshots { private static final Path TEST_PATH = new Path("/foo"); private static final Configuration conf = new HdfsConfiguration(); static { conf.set(DFSConfigKeys.DFS_NAMENODE_SECONDARY_HTTP_ADDRESS_KEY, "0.0.0.0:0"); } @Before public void setUp() { FileUtil.fullyDeleteContents(new File(MiniDFSCluster.getBaseDirectory())); } /** * Regression test for HDFS-5433 - "When reloading fsimage during * checkpointing, we should clear existing snapshottable directories" */ @Test public void testCheckpoint() throws IOException { MiniDFSCluster cluster = null; SecondaryNameNode secondary = null; try { cluster = new MiniDFSCluster.Builder(conf).build(); cluster.waitActive(); secondary = new SecondaryNameNode(conf); SnapshotManager nnSnapshotManager = cluster.getNamesystem().getSnapshotManager(); SnapshotManager secondarySnapshotManager = secondary.getFSNamesystem().getSnapshotManager(); FileSystem fs = cluster.getFileSystem(); HdfsAdmin admin = new HdfsAdmin(FileSystem.getDefaultUri(conf), conf); assertEquals(0, nnSnapshotManager.getNumSnapshots()); assertEquals(0, nnSnapshotManager.getNumSnapshottableDirs()); assertEquals(0, secondarySnapshotManager.getNumSnapshots()); assertEquals(0, secondarySnapshotManager.getNumSnapshottableDirs()); // 1. Create a snapshottable directory foo on the NN. fs.mkdirs(TEST_PATH); admin.allowSnapshot(TEST_PATH); assertEquals(0, nnSnapshotManager.getNumSnapshots()); assertEquals(1, nnSnapshotManager.getNumSnapshottableDirs()); // 2. Create a snapshot of the dir foo. This will be referenced both in // the SnapshotManager as well as in the file system tree. The snapshot // count will go up to 1. Path snapshotPath = fs.createSnapshot(TEST_PATH); assertEquals(1, nnSnapshotManager.getNumSnapshots()); assertEquals(1, nnSnapshotManager.getNumSnapshottableDirs()); // 3. Start up a 2NN and have it do a checkpoint. It will have foo and its // snapshot in its list of snapshottable dirs referenced from the // SnapshotManager, as well as in the file system tree. secondary.doCheckpoint(); assertEquals(1, secondarySnapshotManager.getNumSnapshots()); assertEquals(1, secondarySnapshotManager.getNumSnapshottableDirs()); // 4. Disallow snapshots on and delete foo on the NN. The snapshot count // will go down to 0 and the snapshottable dir will be removed from the fs // tree. fs.deleteSnapshot(TEST_PATH, snapshotPath.getName()); admin.disallowSnapshot(TEST_PATH); assertEquals(0, nnSnapshotManager.getNumSnapshots()); assertEquals(0, nnSnapshotManager.getNumSnapshottableDirs()); // 5. Have the NN do a saveNamespace, writing out a new fsimage with // snapshot count 0. NameNodeAdapter.enterSafeMode(cluster.getNameNode(), false); NameNodeAdapter.saveNamespace(cluster.getNameNode()); NameNodeAdapter.leaveSafeMode(cluster.getNameNode()); // 6. Have the still-running 2NN do a checkpoint. It will notice that the // fsimage has changed on the NN and redownload/reload from that image. // This will replace all INodes in the file system tree as well as reset // the snapshot counter to 0 in the SnapshotManager. However, it will not // clear the list of snapshottable dirs referenced from the // SnapshotManager. When it writes out an fsimage, the 2NN will write out // 0 for the snapshot count, but still serialize the snapshottable dir // referenced in the SnapshotManager even though it no longer appears in // the file system tree. The NN will not be able to start up with this. secondary.doCheckpoint(); assertEquals(0, secondarySnapshotManager.getNumSnapshots()); assertEquals(0, secondarySnapshotManager.getNumSnapshottableDirs()); } finally { if (cluster != null) { cluster.shutdown(); } if (secondary != null) { secondary.shutdown(); } } } }