/** * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.hadoop.hdfs.server.namenode.snapshot; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.PrintStream; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileChecksum; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.hdfs.AppendTestUtil; import org.apache.hadoop.hdfs.DFSConfigKeys; import org.junit.After; import org.junit.Before; import org.junit.Test; import static org.hamcrest.CoreMatchers.is; import static org.hamcrest.CoreMatchers.not; import static org.junit.Assert.assertEquals; import static org.junit.Assert.assertThat; import static org.junit.Assert.assertTrue; import static org.junit.Assert.fail; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FsShell; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hdfs.DFSTestUtil; import org.apache.hadoop.hdfs.DistributedFileSystem; import org.apache.hadoop.hdfs.MiniDFSCluster; import org.apache.hadoop.util.ToolRunner; public class TestSnapshotFileLength { private static final long SEED = 0; private static final short REPLICATION = 1; private static final int BLOCKSIZE = 1024; private static final Configuration conf = new Configuration(); private static MiniDFSCluster cluster; private static DistributedFileSystem hdfs; private final Path dir = new Path("/TestSnapshotFileLength"); private final Path sub = new Path(dir, "sub1"); private final String file1Name = "file1"; private final String snapshot1 = "snapshot1"; @Before public void setUp() throws Exception { conf.setLong(DFSConfigKeys.DFS_NAMENODE_MIN_BLOCK_SIZE_KEY, BLOCKSIZE); conf.setInt(DFSConfigKeys.DFS_BYTES_PER_CHECKSUM_KEY, BLOCKSIZE); cluster = new MiniDFSCluster.Builder(conf).numDataNodes(REPLICATION) .build(); cluster.waitActive(); hdfs = cluster.getFileSystem(); } @After public void tearDown() throws Exception { if (cluster != null) { cluster.shutdown(); cluster = null; } } /** * Test that we cannot read a file beyond its snapshot length * when accessing it via a snapshot path. * */ @Test (timeout=300000) public void testSnapshotfileLength() throws Exception { hdfs.mkdirs(sub); int bytesRead; byte[] buffer = new byte[BLOCKSIZE * 8]; int origLen = BLOCKSIZE + 1; int toAppend = BLOCKSIZE; FSDataInputStream fis = null; FileStatus fileStatus = null; // Create and write a file. Path file1 = new Path(sub, file1Name); DFSTestUtil.createFile(hdfs, file1, BLOCKSIZE, 0, BLOCKSIZE, REPLICATION, SEED); DFSTestUtil.appendFile(hdfs, file1, origLen); // Create a snapshot on the parent directory. hdfs.allowSnapshot(sub); hdfs.createSnapshot(sub, snapshot1); Path file1snap1 = SnapshotTestHelper.getSnapshotPath(sub, snapshot1, file1Name); final FileChecksum snapChksum1 = hdfs.getFileChecksum(file1snap1); assertThat("file and snapshot file checksums are not equal", hdfs.getFileChecksum(file1), is(snapChksum1)); // Append to the file. FSDataOutputStream out = hdfs.append(file1); // Nothing has been appended yet. All checksums should still be equal. // HDFS-8150:Fetching checksum for file under construction should fail try { hdfs.getFileChecksum(file1); fail("getFileChecksum should fail for files " + "with blocks under construction"); } catch (IOException ie) { assertTrue(ie.getMessage().contains( "Fail to get checksum, since file " + file1 + " is under construction.")); } assertThat("snapshot checksum (post-open for append) has changed", hdfs.getFileChecksum(file1snap1), is(snapChksum1)); try { AppendTestUtil.write(out, 0, toAppend); // Test reading from snapshot of file that is open for append byte[] dataFromSnapshot = DFSTestUtil.readFileBuffer(hdfs, file1snap1); assertThat("Wrong data size in snapshot.", dataFromSnapshot.length, is(origLen)); // Verify that checksum didn't change assertThat("snapshot checksum (post-append) has changed", hdfs.getFileChecksum(file1snap1), is(snapChksum1)); } finally { out.close(); } assertThat("file and snapshot file checksums (post-close) are equal", hdfs.getFileChecksum(file1), not(snapChksum1)); assertThat("snapshot file checksum (post-close) has changed", hdfs.getFileChecksum(file1snap1), is(snapChksum1)); // Make sure we can read the entire file via its non-snapshot path. fileStatus = hdfs.getFileStatus(file1); assertThat(fileStatus.getLen(), is((long) origLen + toAppend)); fis = hdfs.open(file1); bytesRead = fis.read(0, buffer, 0, buffer.length); assertThat(bytesRead, is(origLen + toAppend)); fis.close(); // Try to open the file via its snapshot path. fis = hdfs.open(file1snap1); fileStatus = hdfs.getFileStatus(file1snap1); assertThat(fileStatus.getLen(), is((long) origLen)); // Make sure we can only read up to the snapshot length. bytesRead = fis.read(0, buffer, 0, buffer.length); assertThat(bytesRead, is(origLen)); fis.close(); byte[] dataFromSnapshot = DFSTestUtil.readFileBuffer(hdfs, file1snap1); assertThat("Wrong data size in snapshot.", dataFromSnapshot.length, is(origLen)); } /** * Adding as part of jira HDFS-5343 * Test for checking the cat command on snapshot path it * cannot read a file beyond snapshot file length * @throws Exception */ @Test (timeout = 600000) public void testSnapshotFileLengthWithCatCommand() throws Exception { FSDataInputStream fis = null; FileStatus fileStatus = null; int bytesRead; byte[] buffer = new byte[BLOCKSIZE * 8]; hdfs.mkdirs(sub); Path file1 = new Path(sub, file1Name); DFSTestUtil.createFile(hdfs, file1, BLOCKSIZE, REPLICATION, SEED); hdfs.allowSnapshot(sub); hdfs.createSnapshot(sub, snapshot1); DFSTestUtil.appendFile(hdfs, file1, BLOCKSIZE); // Make sure we can read the entire file via its non-snapshot path. fileStatus = hdfs.getFileStatus(file1); assertEquals("Unexpected file length", BLOCKSIZE * 2, fileStatus.getLen()); fis = hdfs.open(file1); bytesRead = fis.read(buffer, 0, buffer.length); assertEquals("Unexpected # bytes read", BLOCKSIZE * 2, bytesRead); fis.close(); Path file1snap1 = SnapshotTestHelper.getSnapshotPath(sub, snapshot1, file1Name); fis = hdfs.open(file1snap1); fileStatus = hdfs.getFileStatus(file1snap1); assertEquals(fileStatus.getLen(), BLOCKSIZE); // Make sure we can only read up to the snapshot length. bytesRead = fis.read(buffer, 0, buffer.length); assertEquals("Unexpected # bytes read", BLOCKSIZE, bytesRead); fis.close(); PrintStream outBackup = System.out; PrintStream errBackup = System.err; ByteArrayOutputStream bao = new ByteArrayOutputStream(); System.setOut(new PrintStream(bao)); System.setErr(new PrintStream(bao)); // Make sure we can cat the file upto to snapshot length FsShell shell = new FsShell(); try { ToolRunner.run(conf, shell, new String[] { "-cat", "/TestSnapshotFileLength/sub1/.snapshot/snapshot1/file1" }); assertEquals("Unexpected # bytes from -cat", BLOCKSIZE, bao.size()); } finally { System.setOut(outBackup); System.setErr(errBackup); } } }