/** * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.hadoop.hdfs.server.namenode; import junit.framework.TestCase; import java.io.*; import java.util.List; import java.util.Random; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hdfs.MiniDFSCluster; import org.apache.hadoop.hdfs.server.namenode.NNStorage.NameNodeDirType; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileUtil; import org.apache.hadoop.fs.Path; /** * This class tests various combinations of dfs.name.dir * and dfs.name.edits.dir configurations. */ public class TestNameEditsConfigs extends TestCase { static final long SEED = 0xDEADBEEFL; static final int BLOCK_SIZE = 4096; static final int FILE_SIZE = 8192; static final int NUM_DATA_NODES = 3; static final String FILE_IMAGE = "current/fsimage"; static final String FILE_EDITS = "current/edits"; short replication = 3; private File base_dir = new File( System.getProperty("test.build.data", "build/test/data"), "dfs/"); protected void setUp() throws java.lang.Exception { if(base_dir.exists()) tearDown(); } protected void tearDown() throws java.lang.Exception { if (base_dir.exists() && !FileUtil.fullyDelete(base_dir)) throw new IOException("Cannot remove directory " + base_dir); } private void writeFile(FileSystem fileSys, Path name, int repl) throws IOException { FSDataOutputStream stm = fileSys.create(name, true, fileSys.getConf().getInt("io.file.buffer.size", 4096), (short)repl, (long)BLOCK_SIZE); byte[] buffer = new byte[FILE_SIZE]; Random rand = new Random(SEED); rand.nextBytes(buffer); stm.write(buffer); stm.close(); } void checkImageAndEditsFilesExistence(File dir, boolean shouldHaveImages, boolean shouldHaveEdits) throws IOException { FSImageTransactionalStorageInspector ins = inspect(dir); if (shouldHaveImages) { assertTrue("Expect images in " + dir, ins.foundImages.size() > 0); } else { assertTrue("Expect no images in " + dir, ins.foundImages.isEmpty()); } List<FileJournalManager.EditLogFile> editlogs = FileJournalManager .matchEditLogs(new File(dir, "current").listFiles()); if (shouldHaveEdits) { assertTrue("Expect edits in " + dir, editlogs.size() > 0); } else { assertTrue("Expect no edits in " + dir, editlogs.isEmpty()); } } private FSImageTransactionalStorageInspector inspect(File storageDir) throws IOException { return FSImageTestUtil.inspectStorageDirectory( new File(storageDir, "current"), NameNodeDirType.IMAGE_AND_EDITS); } private void checkFile(FileSystem fileSys, Path name, int repl) throws IOException { assertTrue(fileSys.exists(name)); int replication = fileSys.getFileStatus(name).getReplication(); assertEquals("replication for " + name, repl, replication); long size = fileSys.getContentSummary(name).getLength(); assertEquals("file size for " + name, size, (long)FILE_SIZE); } private void cleanupFile(FileSystem fileSys, Path name) throws IOException { assertTrue(fileSys.exists(name)); fileSys.delete(name, true); assertTrue(!fileSys.exists(name)); } SecondaryNameNode startSecondaryNameNode(Configuration conf ) throws IOException { conf.set("dfs.secondary.http.address", "0.0.0.0:0"); return new SecondaryNameNode(conf); } // Test dfs.name.dir.policy configuration with 0: success public void testNameDirPolicy0() throws IOException { MiniDFSCluster cluster = null; SecondaryNameNode secondary = null; Configuration conf = null; FileSystem fileSys = null; File nameAndEdits = new File(base_dir, "name_and_edits"); File checkpointNameAndEdits = new File(base_dir, "second_name_and_edits"); Path file1 = new Path("TestNameEditsConfigs1"); conf = new Configuration(); conf.set("dfs.name.dir.policy", "0"); conf.set("dfs.name.dir", nameAndEdits.getPath()); conf.set("dfs.name.edits.dir", nameAndEdits.getPath()); conf.set("fs.checkpoint.dir", checkpointNameAndEdits.getPath()); conf.set("fs.checkpoint.edits.dir", checkpointNameAndEdits.getPath()); replication = (short)conf.getInt("dfs.replication", 3); cluster = new MiniDFSCluster(0, conf, NUM_DATA_NODES, true, false, true, null, null, null, null); cluster.waitActive(); secondary = startSecondaryNameNode(conf); fileSys = cluster.getFileSystem(); try { assertTrue(!fileSys.exists(file1)); writeFile(fileSys, file1, replication); checkFile(fileSys, file1, replication); secondary.doCheckpoint(); } finally { fileSys.close(); cluster.shutdown(); secondary.shutdown(); } } // Test dfs.name.dir.policy configuration failure cases private void testNameDirPolicyFailure(int policy, boolean useUri) throws IOException { Configuration conf = null; MiniDFSCluster cluster = null; String prefix = useUri ? "file:" : ""; File nameAndEdits = new File(base_dir, "name_and_edits"); String policyStr = Integer.toString(policy); conf = new Configuration(); conf.set("dfs.name.dir.policy", policyStr); conf.set("dfs.name.dir", prefix + nameAndEdits.getPath()); conf.set("dfs.name.edits.dir", prefix + nameAndEdits.getPath()); try { cluster = new MiniDFSCluster(0, conf, NUM_DATA_NODES, false, false, true, null, null, null, null); fail("The startup should fail"); } catch (IOException e) { // expect to fail System.out.println("cluster start failed due to name/edits dir " + "violating policy " + policyStr); } finally { cluster = null; } } /** * Test various configuration options of dfs.name.dir and dfs.name.edits.dir * The test creates files and restarts cluster with different configs. * 1. Starts cluster with shared name and edits dirs * 2. Restarts cluster by adding additional (different) name and edits dirs * 3. Restarts cluster by removing shared name and edits dirs by allowing to * start using separate name and edits dirs * 4. Restart cluster by adding shared directory again, but make sure we * do not read any stale image or edits. * All along the test, we create and delete files at reach restart to make * sure we are reading proper edits and image. */ public void testNameEditsConfigs() throws IOException { Path file1 = new Path("TestNameEditsConfigs1"); Path file2 = new Path("TestNameEditsConfigs2"); Path file3 = new Path("TestNameEditsConfigs3"); MiniDFSCluster cluster = null; SecondaryNameNode secondary = null; Configuration conf = null; FileSystem fileSys = null; File newNameDir = new File(base_dir, "name"); File newEditsDir = new File(base_dir, "edits"); File nameAndEdits = new File(base_dir, "name_and_edits"); File checkpointNameDir = new File(base_dir, "secondname"); File checkpointEditsDir = new File(base_dir, "secondedits"); File checkpointNameAndEdits = new File(base_dir, "second_name_and_edits"); // Start namenode with same dfs.name.dir and dfs.name.edits.dir conf = new Configuration(); conf.set("dfs.name.dir", nameAndEdits.getPath()); conf.set("dfs.name.edits.dir", nameAndEdits.getPath()); conf.set("fs.checkpoint.dir", checkpointNameAndEdits.getPath()); conf.set("fs.checkpoint.edits.dir", checkpointNameAndEdits.getPath()); replication = (short)conf.getInt("dfs.replication", 3); // Manage our own dfs directories cluster = new MiniDFSCluster(0, conf, NUM_DATA_NODES, true, false, true, null, null, null, null); cluster.waitActive(); secondary = startSecondaryNameNode(conf); fileSys = cluster.getFileSystem(); try { assertTrue(!fileSys.exists(file1)); writeFile(fileSys, file1, replication); checkFile(fileSys, file1, replication); secondary.doCheckpoint(); } finally { fileSys.close(); cluster.shutdown(); secondary.shutdown(); } // Start namenode with additional dfs.name.dir and dfs.name.edits.dir conf = new Configuration(); assertTrue(newNameDir.mkdir()); assertTrue(newEditsDir.mkdir()); conf.set("dfs.name.dir", nameAndEdits.getPath() + "," + newNameDir.getPath()); conf.set("dfs.name.edits.dir", nameAndEdits.getPath() + "," + newEditsDir.getPath()); conf.set("fs.checkpoint.dir", checkpointNameDir.getPath() + "," + checkpointNameAndEdits.getPath()); conf.set("fs.checkpoint.edits.dir", checkpointEditsDir.getPath() + "," + checkpointNameAndEdits.getPath()); replication = (short)conf.getInt("dfs.replication", 3); // Manage our own dfs directories. Do not format. cluster = new MiniDFSCluster(0, conf, NUM_DATA_NODES, false, false, true, null, null, null, null); cluster.waitActive(); secondary = startSecondaryNameNode(conf); fileSys = cluster.getFileSystem(); try { assertTrue(fileSys.exists(file1)); checkFile(fileSys, file1, replication); cleanupFile(fileSys, file1); writeFile(fileSys, file2, replication); checkFile(fileSys, file2, replication); secondary.doCheckpoint(); } finally { fileSys.close(); cluster.shutdown(); secondary.shutdown(); } checkImageAndEditsFilesExistence(nameAndEdits, true, true); checkImageAndEditsFilesExistence(newNameDir, true, false); checkImageAndEditsFilesExistence(newEditsDir, false, true); checkImageAndEditsFilesExistence(checkpointNameAndEdits, true, true); checkImageAndEditsFilesExistence(checkpointNameDir, true, false); checkImageAndEditsFilesExistence(checkpointEditsDir, false, true); // Now remove common directory both have and start namenode with // separate name and edits dirs new File(nameAndEdits, FILE_EDITS).renameTo( new File(newNameDir, FILE_EDITS)); new File(nameAndEdits, FILE_IMAGE).renameTo( new File(newEditsDir, FILE_IMAGE)); new File(checkpointNameAndEdits, FILE_EDITS).renameTo( new File(checkpointNameDir, FILE_EDITS)); new File(checkpointNameAndEdits, FILE_IMAGE).renameTo( new File(checkpointEditsDir, FILE_IMAGE)); conf = new Configuration(); conf.set("dfs.name.dir", newNameDir.getPath()); conf.set("dfs.name.edits.dir", newEditsDir.getPath()); conf.set("fs.checkpoint.dir", checkpointNameDir.getPath()); conf.set("fs.checkpoint.edits.dir", checkpointEditsDir.getPath()); replication = (short)conf.getInt("dfs.replication", 3); cluster = new MiniDFSCluster(0, conf, NUM_DATA_NODES, false, false, true, null, null, null, null); cluster.waitActive(); secondary = startSecondaryNameNode(conf); fileSys = cluster.getFileSystem(); try { assertTrue(!fileSys.exists(file1)); assertTrue(fileSys.exists(file2)); checkFile(fileSys, file2, replication); cleanupFile(fileSys, file2); writeFile(fileSys, file3, replication); checkFile(fileSys, file3, replication); secondary.doCheckpoint(); } finally { fileSys.close(); cluster.shutdown(); secondary.shutdown(); } checkImageAndEditsFilesExistence(newNameDir, true, false); checkImageAndEditsFilesExistence(newEditsDir, false, true); checkImageAndEditsFilesExistence(checkpointNameDir, true, false); checkImageAndEditsFilesExistence(checkpointEditsDir, false, true); // Add old name_and_edits dir. File system should not read image or edits // from old dir // clear the directories entirely, otherwise the startup will fail assertTrue(FileUtil.fullyDelete(nameAndEdits)); assertTrue(FileUtil.fullyDelete(checkpointNameAndEdits)); nameAndEdits.mkdirs(); checkpointNameAndEdits.mkdirs(); conf = new Configuration(); conf.set("dfs.name.dir", nameAndEdits.getPath() + "," + newNameDir.getPath()); conf.set("dfs.name.edits.dir", nameAndEdits + "," + newEditsDir.getPath()); conf.set("fs.checkpoint.dir", checkpointNameDir.getPath() + "," + checkpointNameAndEdits.getPath()); conf.set("fs.checkpoint.edits.dir", checkpointEditsDir.getPath() + "," + checkpointNameAndEdits.getPath()); replication = (short)conf.getInt("dfs.replication", 3); cluster = new MiniDFSCluster(0, conf, NUM_DATA_NODES, false, false, true, null, null, null, null); cluster.waitActive(); secondary = startSecondaryNameNode(conf); fileSys = cluster.getFileSystem(); try { assertTrue(!fileSys.exists(file1)); assertTrue(!fileSys.exists(file2)); assertTrue(fileSys.exists(file3)); checkFile(fileSys, file3, replication); secondary.doCheckpoint(); } finally { fileSys.close(); cluster.shutdown(); secondary.shutdown(); } checkImageAndEditsFilesExistence(nameAndEdits, true, true); checkImageAndEditsFilesExistence(checkpointNameAndEdits, true, true); } /** * Test various configuration options of dfs.name.dir and dfs.name.edits.dir * This test tries to simulate failure scenarios. * 1. Start cluster with shared name and edits dir * 2. Restart cluster by adding separate name and edits dirs * 3. Restart cluster by removing shared name and edits dir * 4. Restart cluster with old shared name and edits dir, but only latest * name dir. This should fail since we dont have latest edits dir * 5. Restart cluster with old shared name and edits dir, but only latest * edits dir. This should fail since we dont have latest name dir */ public void testNameEditsConfigsFailure() throws IOException { Path file1 = new Path("TestNameEditsConfigs1"); Path file2 = new Path("TestNameEditsConfigs2"); Path file3 = new Path("TestNameEditsConfigs3"); MiniDFSCluster cluster = null; Configuration conf = null; FileSystem fileSys = null; File newNameDir = new File(base_dir, "name"); File newEditsDir = new File(base_dir, "edits"); File nameAndEdits = new File(base_dir, "name_and_edits"); // Start namenode with same dfs.name.dir and dfs.name.edits.dir conf = new Configuration(); conf.set("dfs.name.dir", nameAndEdits.getPath()); conf.set("dfs.name.edits.dir", nameAndEdits.getPath()); replication = (short)conf.getInt("dfs.replication", 3); // Manage our own dfs directories cluster = new MiniDFSCluster(0, conf, NUM_DATA_NODES, true, false, true, null, null, null, null); cluster.waitActive(); fileSys = cluster.getFileSystem(); try { assertTrue(!fileSys.exists(file1)); writeFile(fileSys, file1, replication); checkFile(fileSys, file1, replication); } finally { fileSys.close(); cluster.shutdown(); } // 2 // Start namenode with additional dfs.namenode.name.dir and dfs.namenode.edits.dir conf = new Configuration(); assertTrue(newNameDir.mkdir()); assertTrue(newEditsDir.mkdir()); conf.set("dfs.name.dir", nameAndEdits.getPath() + "," + newNameDir.getPath()); conf.set("dfs.name.edits.dir", nameAndEdits.getPath() + "," + newEditsDir.getPath()); replication = (short)conf.getInt("dfs.replication", 3); // Manage our own dfs directories. Do not format. cluster = new MiniDFSCluster(0, conf, NUM_DATA_NODES, false, false, true, null, null, null, null); cluster.waitActive(); fileSys = cluster.getFileSystem(); try { assertTrue(fileSys.exists(file1)); checkFile(fileSys, file1, replication); cleanupFile(fileSys, file1); writeFile(fileSys, file2, replication); checkFile(fileSys, file2, replication); } finally { fileSys.close(); cluster.shutdown(); } // 3 // Now remove common directory both have and start namenode with // separate name and edits dirs conf = new Configuration(); conf.set("dfs.name.dir", newNameDir.getPath()); conf.set("dfs.name.edits.dir", newEditsDir.getPath()); replication = (short)conf.getInt("dfs.replication", 3); cluster = new MiniDFSCluster(0, conf, NUM_DATA_NODES, false, false, true, null, null, null, null); cluster.waitActive(); fileSys = cluster.getFileSystem(); try { assertTrue(!fileSys.exists(file1)); assertTrue(fileSys.exists(file2)); checkFile(fileSys, file2, replication); cleanupFile(fileSys, file2); writeFile(fileSys, file3, replication); checkFile(fileSys, file3, replication); } finally { fileSys.close(); // force to save the namespace cluster.getNameNode().saveNamespace(true, false); cluster.shutdown(); } // 4 // Add old shared directory for name and edits along with latest name conf = new Configuration(); conf.set("dfs.name.dir", newNameDir.getPath() + "," + nameAndEdits.getPath()); conf.set("dfs.name.edits.dir", nameAndEdits.getPath()); replication = (short)conf.getInt("dfs.replication", 3); try { cluster = new MiniDFSCluster(0, conf, NUM_DATA_NODES, false, false, true, null, null, null, null); assertTrue(false); } catch (IOException e) { // expect to fail System.out.println("cluster start failed due to missing " + "latest edits dir"); } finally { cluster = null; } // 5 // Add old shared directory for name and edits along with latest edits. // This is OK, since the latest edits will have segments leading all // the way from the image in name_and_edits. conf = new Configuration(); conf.set("dfs.name.dir", nameAndEdits.getPath()); conf.set("dfs.name.edits.dir", newEditsDir.getPath() + "," + nameAndEdits.getPath()); replication = (short)conf.getInt("dfs.replication", 3); try { cluster = new MiniDFSCluster(0, conf, NUM_DATA_NODES, false, false, true, null, null, null, null); fileSys = cluster.getFileSystem(); assertFalse(fileSys.exists(file1)); assertFalse(fileSys.exists(file2)); assertTrue(fileSys.exists(file3)); checkFile(fileSys, file3, replication); cleanupFile(fileSys, file3); writeFile(fileSys, file3, replication); checkFile(fileSys, file3, replication); } catch (IOException e) { // expect to fail System.out.println("cluster start failed due to missing latest name dir"); } finally { cluster = null; } } public void testNameDirPolicyFailures() throws Exception { // Test dfs.name.dir.policy configuration // when using uri's testNameDirPolicyFailure(1, true); testNameDirPolicyFailure(2, true); // when using simple directory names testNameDirPolicyFailure(1, false); testNameDirPolicyFailure(2, false); } public void testNonEmptyStorageDirectory() throws IOException { MiniDFSCluster cluster = null; Configuration conf = null; File nameAndEdits = new File(base_dir, "name_and_edits1"); File nameAndEdits2 = new File(base_dir, "name_and_edits2"); // Start namenode with same dfs.name.dir and dfs.name.edits.dir conf = new Configuration(); conf.set("dfs.name.dir", nameAndEdits.getPath()); conf.set("dfs.name.edits.dir", nameAndEdits.getPath()); cluster = new MiniDFSCluster(0, conf, NUM_DATA_NODES, true, false, true, null, null, null, null); cluster.shutdown(); File current = new File(nameAndEdits2, "current"); current.mkdirs(); conf.set("dfs.name.dir", nameAndEdits.getPath() + "," + nameAndEdits2.getPath()); conf.set("dfs.name.edits.dir", nameAndEdits.getPath() + "," + nameAndEdits2.getPath()); // added directories, we should fail, because the directories are non-empty try{ cluster = new MiniDFSCluster(0, conf, NUM_DATA_NODES, false, false, true, null, null, null, null); fail("We should not proceed with a non-empty directory"); } catch (Exception e) { assertTrue(e.toString().contains("is not empty")); } current.delete(); // now we should be able to start cluster = new MiniDFSCluster(0, conf, NUM_DATA_NODES, false, false, true, null, null, null, null); cluster.shutdown(); } }