HDFS-8399. Erasure Coding: unit test the behaviour of BlockManager recovery work for the deleted blocks. Contributed by Rakesh R.
This commit is contained in:
parent
9312b168e2
commit
4de48211c6
@ -385,3 +385,6 @@
|
||||
|
||||
HDFS-8804. Erasure Coding: use DirectBufferPool in DFSStripedInputStream for
|
||||
buffer allocation. (jing9)
|
||||
|
||||
HDFS-8399. Erasure Coding: unit test the behaviour of BlockManager recovery
|
||||
work for the deleted blocks. (Rakesh R via zhz)
|
||||
|
@ -19,16 +19,24 @@
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.fs.permission.FsPermission;
|
||||
import org.apache.hadoop.fs.permission.PermissionStatus;
|
||||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||
import org.apache.hadoop.hdfs.DFSTestUtil;
|
||||
import org.apache.hadoop.hdfs.DistributedFileSystem;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.hdfs.protocol.Block;
|
||||
import org.apache.hadoop.hdfs.protocol.BlockStoragePolicy;
|
||||
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
||||
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfo;
|
||||
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStriped;
|
||||
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoUnderConstructionStriped;
|
||||
import org.apache.hadoop.hdfs.server.blockmanagement.BlockStoragePolicySuite;
|
||||
@ -199,4 +207,78 @@ public void testBlockUCStripedComputeQuotaUsage()
|
||||
// is 9(= 3 + 6). Consumed storage space should be 1024 * 9 = 9216.
|
||||
assertEquals(9216, counts.getStorageSpace());
|
||||
}
|
||||
|
||||
/**
|
||||
* Test the behavior of striped and contiguous block deletions.
|
||||
*/
|
||||
@Test(timeout = 60000)
|
||||
public void testDeleteOp() throws Exception {
|
||||
MiniDFSCluster cluster = null;
|
||||
try {
|
||||
final int len = 1024;
|
||||
final Path parentDir = new Path("/parentDir");
|
||||
final Path zone = new Path(parentDir, "zone");
|
||||
final Path zoneFile = new Path(zone, "zoneFile");
|
||||
final Path contiguousFile = new Path(parentDir, "someFile");
|
||||
final DistributedFileSystem dfs;
|
||||
final Configuration conf = new Configuration();
|
||||
final short GROUP_SIZE = HdfsConstants.NUM_DATA_BLOCKS
|
||||
+ HdfsConstants.NUM_PARITY_BLOCKS;
|
||||
conf.setInt(DFSConfigKeys.DFS_NAMENODE_MAX_XATTRS_PER_INODE_KEY, 2);
|
||||
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(GROUP_SIZE)
|
||||
.build();
|
||||
cluster.waitActive();
|
||||
|
||||
FSNamesystem fsn = cluster.getNamesystem();
|
||||
dfs = cluster.getFileSystem();
|
||||
dfs.mkdirs(zone);
|
||||
|
||||
// create erasure zone
|
||||
dfs.createErasureCodingZone(zone, null, 0);
|
||||
DFSTestUtil.createFile(dfs, zoneFile, len, (short) 1, 0xFEED);
|
||||
DFSTestUtil.createFile(dfs, contiguousFile, len, (short) 1, 0xFEED);
|
||||
final FSDirectory fsd = fsn.getFSDirectory();
|
||||
|
||||
// Case-1: Verify the behavior of striped blocks
|
||||
// Get blocks of striped file
|
||||
INode inodeStriped = fsd.getINode("/parentDir/zone/zoneFile");
|
||||
assertTrue("Failed to get INodeFile for /parentDir/zone/zoneFile",
|
||||
inodeStriped instanceof INodeFile);
|
||||
INodeFile inodeStripedFile = (INodeFile) inodeStriped;
|
||||
BlockInfo[] stripedBlks = inodeStripedFile.getBlocks();
|
||||
for (BlockInfo blockInfo : stripedBlks) {
|
||||
assertFalse("Mistakenly marked the block as deleted!",
|
||||
blockInfo.isDeleted());
|
||||
}
|
||||
|
||||
// delete erasure zone directory
|
||||
dfs.delete(zone, true);
|
||||
for (BlockInfo blockInfo : stripedBlks) {
|
||||
assertTrue("Didn't mark the block as deleted!", blockInfo.isDeleted());
|
||||
}
|
||||
|
||||
// Case-2: Verify the behavior of contiguous blocks
|
||||
// Get blocks of contiguous file
|
||||
INode inode = fsd.getINode("/parentDir/someFile");
|
||||
assertTrue("Failed to get INodeFile for /parentDir/someFile",
|
||||
inode instanceof INodeFile);
|
||||
INodeFile inodeFile = (INodeFile) inode;
|
||||
BlockInfo[] contiguousBlks = inodeFile.getBlocks();
|
||||
for (BlockInfo blockInfo : contiguousBlks) {
|
||||
assertFalse("Mistakenly marked the block as deleted!",
|
||||
blockInfo.isDeleted());
|
||||
}
|
||||
|
||||
// delete parent directory
|
||||
dfs.delete(parentDir, true);
|
||||
for (BlockInfo blockInfo : contiguousBlks) {
|
||||
assertTrue("Didn't mark the block as deleted!", blockInfo.isDeleted());
|
||||
}
|
||||
} finally {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user