diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES-HDFS-EC-7285.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES-HDFS-EC-7285.txt index 8e54c762ed..5ad084bfe2 100755 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES-HDFS-EC-7285.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES-HDFS-EC-7285.txt @@ -388,3 +388,6 @@ HDFS-8399. Erasure Coding: unit test the behaviour of BlockManager recovery work for the deleted blocks. (Rakesh R via zhz) + + HDFS-8857. Erasure Coding: Fix ArrayIndexOutOfBoundsException in + TestWriteStripedFileWithFailure. (Li Bo) diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteStripedFileWithFailure.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteStripedFileWithFailure.java index 5448773600..3679c5fcac 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteStripedFileWithFailure.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteStripedFileWithFailure.java @@ -37,7 +37,7 @@ public class TestWriteStripedFileWithFailure { public static final Log LOG = LogFactory - .getLog(TestReadStripedFileWithMissingBlocks.class); + .getLog(TestWriteStripedFileWithFailure.class); private static MiniDFSCluster cluster; private static FileSystem fs; private static Configuration conf = new HdfsConfiguration(); @@ -114,7 +114,7 @@ private void writeFileWithDNFailure(int fileLength, dataDNFailureNum); Assert.assertNotNull(dataDNFailureIndices); int[] parityDNFailureIndices = StripedFileTestUtil.randomArray(dataBlocks, - dataBlocks + parityBlocks, dataDNFailureNum); + dataBlocks + parityBlocks, parityDNFailureNum); Assert.assertNotNull(parityDNFailureIndices); int[] failedDataNodes = new int[dataDNFailureNum + parityDNFailureNum];