HDFS-14560. Allow block replication parameters to be refreshable. Contributed by Stephen O'Donnell.
This commit is contained in:
parent
54f9f75a44
commit
4f455290b1
@ -383,7 +383,7 @@ public long getTotalECBlockGroups() {
|
|||||||
final int maxCorruptFilesReturned;
|
final int maxCorruptFilesReturned;
|
||||||
|
|
||||||
final float blocksInvalidateWorkPct;
|
final float blocksInvalidateWorkPct;
|
||||||
final int blocksReplWorkMultiplier;
|
private int blocksReplWorkMultiplier;
|
||||||
|
|
||||||
// whether or not to issue block encryption keys.
|
// whether or not to issue block encryption keys.
|
||||||
final boolean encryptDataTransfer;
|
final boolean encryptDataTransfer;
|
||||||
@ -897,11 +897,78 @@ private void dumpBlockMeta(Block block, PrintWriter out) {
|
|||||||
out.println("");
|
out.println("");
|
||||||
}
|
}
|
||||||
|
|
||||||
/** @return maxReplicationStreams */
|
/** Returns the current setting for maxReplicationStreams, which is set by
|
||||||
|
* {@code DFSConfigKeys.DFS_NAMENODE_REPLICATION_MAX_STREAMS_KEY}.
|
||||||
|
*
|
||||||
|
* @return maxReplicationStreams
|
||||||
|
*/
|
||||||
public int getMaxReplicationStreams() {
|
public int getMaxReplicationStreams() {
|
||||||
return maxReplicationStreams;
|
return maxReplicationStreams;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static private void ensurePositiveInt(int val, String key) {
|
||||||
|
Preconditions.checkArgument(
|
||||||
|
(val > 0),
|
||||||
|
key + " = '" + val + "' is invalid. " +
|
||||||
|
"It should be a positive, non-zero integer value.");
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Updates the value used for maxReplicationStreams, which is set by
|
||||||
|
* {@code DFSConfigKeys.DFS_NAMENODE_REPLICATION_MAX_STREAMS_KEY} initially.
|
||||||
|
*
|
||||||
|
* @param newVal - Must be a positive non-zero integer.
|
||||||
|
*/
|
||||||
|
public void setMaxReplicationStreams(int newVal) {
|
||||||
|
ensurePositiveInt(newVal,
|
||||||
|
DFSConfigKeys.DFS_NAMENODE_REPLICATION_MAX_STREAMS_KEY);
|
||||||
|
maxReplicationStreams = newVal;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** Returns the current setting for maxReplicationStreamsHardLimit, set by
|
||||||
|
* {@code DFSConfigKeys.DFS_NAMENODE_REPLICATION_STREAMS_HARD_LIMIT_KEY}.
|
||||||
|
*
|
||||||
|
* @return maxReplicationStreamsHardLimit
|
||||||
|
*/
|
||||||
|
public int getReplicationStreamsHardLimit() {
|
||||||
|
return replicationStreamsHardLimit;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Updates the value used for replicationStreamsHardLimit, which is set by
|
||||||
|
* {@code DFSConfigKeys.DFS_NAMENODE_REPLICATION_STREAMS_HARD_LIMIT_KEY}
|
||||||
|
* initially.
|
||||||
|
*
|
||||||
|
* @param newVal - Must be a positive non-zero integer.
|
||||||
|
*/
|
||||||
|
public void setReplicationStreamsHardLimit(int newVal) {
|
||||||
|
ensurePositiveInt(newVal,
|
||||||
|
DFSConfigKeys.DFS_NAMENODE_REPLICATION_STREAMS_HARD_LIMIT_KEY);
|
||||||
|
replicationStreamsHardLimit = newVal;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** Returns the current setting for blocksReplWorkMultiplier, set by
|
||||||
|
* {@code DFSConfigKeys.
|
||||||
|
* DFS_NAMENODE_REPLICATION_WORK_MULTIPLIER_PER_ITERATION}.
|
||||||
|
*
|
||||||
|
* @return maxReplicationStreamsHardLimit
|
||||||
|
*/
|
||||||
|
public int getBlocksReplWorkMultiplier() {
|
||||||
|
return blocksReplWorkMultiplier;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Updates the value used for blocksReplWorkMultiplier, set by
|
||||||
|
* {@code DFSConfigKeys.
|
||||||
|
* DFS_NAMENODE_REPLICATION_WORK_MULTIPLIER_PER_ITERATION} initially.
|
||||||
|
* @param newVal - Must be a positive non-zero integer.
|
||||||
|
*/
|
||||||
|
public void setBlocksReplWorkMultiplier(int newVal) {
|
||||||
|
ensurePositiveInt(newVal,
|
||||||
|
DFSConfigKeys.DFS_NAMENODE_REPLICATION_WORK_MULTIPLIER_PER_ITERATION);
|
||||||
|
blocksReplWorkMultiplier = newVal;
|
||||||
|
}
|
||||||
|
|
||||||
public int getDefaultStorageNum(BlockInfo block) {
|
public int getDefaultStorageNum(BlockInfo block) {
|
||||||
switch (block.getBlockType()) {
|
switch (block.getBlockType()) {
|
||||||
case STRIPED: return ((BlockInfoStriped) block).getRealTotalBlockNum();
|
case STRIPED: return ((BlockInfoStriped) block).getRealTotalBlockNum();
|
||||||
|
@ -50,6 +50,7 @@
|
|||||||
import org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenIdentifier;
|
import org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenIdentifier;
|
||||||
import org.apache.hadoop.hdfs.server.aliasmap.InMemoryAliasMap;
|
import org.apache.hadoop.hdfs.server.aliasmap.InMemoryAliasMap;
|
||||||
import org.apache.hadoop.hdfs.server.aliasmap.InMemoryLevelDBAliasMapServer;
|
import org.apache.hadoop.hdfs.server.aliasmap.InMemoryLevelDBAliasMapServer;
|
||||||
|
import org.apache.hadoop.hdfs.server.blockmanagement.BlockManager;
|
||||||
import org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager;
|
import org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager;
|
||||||
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NamenodeRole;
|
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NamenodeRole;
|
||||||
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.RollingUpgradeStartupOption;
|
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.RollingUpgradeStartupOption;
|
||||||
@ -165,6 +166,13 @@
|
|||||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_HEARTBEAT_RECHECK_INTERVAL_DEFAULT;
|
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_HEARTBEAT_RECHECK_INTERVAL_DEFAULT;
|
||||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.FS_PROTECTED_DIRECTORIES;
|
import static org.apache.hadoop.hdfs.DFSConfigKeys.FS_PROTECTED_DIRECTORIES;
|
||||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_STORAGE_POLICY_SATISFIER_MODE_KEY;
|
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_STORAGE_POLICY_SATISFIER_MODE_KEY;
|
||||||
|
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_REPLICATION_MAX_STREAMS_KEY;
|
||||||
|
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_REPLICATION_MAX_STREAMS_DEFAULT;
|
||||||
|
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_REPLICATION_STREAMS_HARD_LIMIT_KEY;
|
||||||
|
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_REPLICATION_STREAMS_HARD_LIMIT_DEFAULT;
|
||||||
|
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_REPLICATION_WORK_MULTIPLIER_PER_ITERATION;
|
||||||
|
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_REPLICATION_WORK_MULTIPLIER_PER_ITERATION_DEFAULT;
|
||||||
|
|
||||||
import static org.apache.hadoop.util.ExitUtil.terminate;
|
import static org.apache.hadoop.util.ExitUtil.terminate;
|
||||||
import static org.apache.hadoop.util.ToolRunner.confirmPrompt;
|
import static org.apache.hadoop.util.ToolRunner.confirmPrompt;
|
||||||
import static org.apache.hadoop.fs.CommonConfigurationKeys.IPC_BACKOFF_ENABLE;
|
import static org.apache.hadoop.fs.CommonConfigurationKeys.IPC_BACKOFF_ENABLE;
|
||||||
@ -299,7 +307,10 @@ public enum OperationCategory {
|
|||||||
DFS_NAMENODE_HEARTBEAT_RECHECK_INTERVAL_KEY,
|
DFS_NAMENODE_HEARTBEAT_RECHECK_INTERVAL_KEY,
|
||||||
FS_PROTECTED_DIRECTORIES,
|
FS_PROTECTED_DIRECTORIES,
|
||||||
HADOOP_CALLER_CONTEXT_ENABLED_KEY,
|
HADOOP_CALLER_CONTEXT_ENABLED_KEY,
|
||||||
DFS_STORAGE_POLICY_SATISFIER_MODE_KEY));
|
DFS_STORAGE_POLICY_SATISFIER_MODE_KEY,
|
||||||
|
DFS_NAMENODE_REPLICATION_MAX_STREAMS_KEY,
|
||||||
|
DFS_NAMENODE_REPLICATION_STREAMS_HARD_LIMIT_KEY,
|
||||||
|
DFS_NAMENODE_REPLICATION_WORK_MULTIPLIER_PER_ITERATION));
|
||||||
|
|
||||||
private static final String USAGE = "Usage: hdfs namenode ["
|
private static final String USAGE = "Usage: hdfs namenode ["
|
||||||
+ StartupOption.BACKUP.getName() + "] | \n\t["
|
+ StartupOption.BACKUP.getName() + "] | \n\t["
|
||||||
@ -2125,12 +2136,63 @@ protected String reconfigurePropertyImpl(String property, String newVal)
|
|||||||
return reconfigureIPCBackoffEnabled(newVal);
|
return reconfigureIPCBackoffEnabled(newVal);
|
||||||
} else if (property.equals(DFS_STORAGE_POLICY_SATISFIER_MODE_KEY)) {
|
} else if (property.equals(DFS_STORAGE_POLICY_SATISFIER_MODE_KEY)) {
|
||||||
return reconfigureSPSModeEvent(newVal, property);
|
return reconfigureSPSModeEvent(newVal, property);
|
||||||
|
} else if (property.equals(DFS_NAMENODE_REPLICATION_MAX_STREAMS_KEY)
|
||||||
|
|| property.equals(DFS_NAMENODE_REPLICATION_STREAMS_HARD_LIMIT_KEY)
|
||||||
|
|| property.equals(
|
||||||
|
DFS_NAMENODE_REPLICATION_WORK_MULTIPLIER_PER_ITERATION)) {
|
||||||
|
return reconfReplicationParameters(newVal, property);
|
||||||
} else {
|
} else {
|
||||||
throw new ReconfigurationException(property, newVal, getConf().get(
|
throw new ReconfigurationException(property, newVal, getConf().get(
|
||||||
property));
|
property));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private String reconfReplicationParameters(final String newVal,
|
||||||
|
final String property) throws ReconfigurationException {
|
||||||
|
BlockManager bm = namesystem.getBlockManager();
|
||||||
|
int newSetting;
|
||||||
|
namesystem.writeLock();
|
||||||
|
try {
|
||||||
|
if (property.equals(DFS_NAMENODE_REPLICATION_MAX_STREAMS_KEY)) {
|
||||||
|
bm.setMaxReplicationStreams(
|
||||||
|
adjustNewVal(DFS_NAMENODE_REPLICATION_MAX_STREAMS_DEFAULT, newVal));
|
||||||
|
newSetting = bm.getMaxReplicationStreams();
|
||||||
|
} else if (property.equals(
|
||||||
|
DFS_NAMENODE_REPLICATION_STREAMS_HARD_LIMIT_KEY)) {
|
||||||
|
bm.setReplicationStreamsHardLimit(
|
||||||
|
adjustNewVal(DFS_NAMENODE_REPLICATION_STREAMS_HARD_LIMIT_DEFAULT,
|
||||||
|
newVal));
|
||||||
|
newSetting = bm.getReplicationStreamsHardLimit();
|
||||||
|
} else if (
|
||||||
|
property.equals(
|
||||||
|
DFS_NAMENODE_REPLICATION_WORK_MULTIPLIER_PER_ITERATION)) {
|
||||||
|
bm.setBlocksReplWorkMultiplier(
|
||||||
|
adjustNewVal(
|
||||||
|
DFS_NAMENODE_REPLICATION_WORK_MULTIPLIER_PER_ITERATION_DEFAULT,
|
||||||
|
newVal));
|
||||||
|
newSetting = bm.getBlocksReplWorkMultiplier();
|
||||||
|
} else {
|
||||||
|
throw new IllegalArgumentException("Unexpected property " +
|
||||||
|
property + "in reconfReplicationParameters");
|
||||||
|
}
|
||||||
|
LOG.info("RECONFIGURE* changed {} to {}", property, newSetting);
|
||||||
|
return String.valueOf(newSetting);
|
||||||
|
} catch (IllegalArgumentException e) {
|
||||||
|
throw new ReconfigurationException(property, newVal, getConf().get(
|
||||||
|
property), e);
|
||||||
|
} finally {
|
||||||
|
namesystem.writeUnlock();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private int adjustNewVal(int defaultVal, String newVal) {
|
||||||
|
if (newVal == null) {
|
||||||
|
return defaultVal;
|
||||||
|
} else {
|
||||||
|
return Integer.parseInt(newVal);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
private String reconfHeartbeatInterval(final DatanodeManager datanodeManager,
|
private String reconfHeartbeatInterval(final DatanodeManager datanodeManager,
|
||||||
final String property, final String newVal)
|
final String property, final String newVal)
|
||||||
throws ReconfigurationException {
|
throws ReconfigurationException {
|
||||||
|
@ -0,0 +1,143 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.hdfs.server.namenode;
|
||||||
|
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.conf.ReconfigurationException;
|
||||||
|
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||||
|
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||||
|
import org.apache.hadoop.hdfs.MiniDFSNNTopology;
|
||||||
|
import org.apache.hadoop.hdfs.server.blockmanagement.BlockManager;
|
||||||
|
import org.apache.hadoop.test.LambdaTestUtils;
|
||||||
|
import org.junit.After;
|
||||||
|
import org.junit.Before;
|
||||||
|
import org.junit.Test;
|
||||||
|
import java.io.IOException;
|
||||||
|
import static org.junit.Assert.assertEquals;
|
||||||
|
import static org.junit.Assert.assertTrue;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* This class tests the replication related parameters in the namenode can
|
||||||
|
* be refreshed dynamically, without a namenode restart.
|
||||||
|
*/
|
||||||
|
public class TestRefreshNamenodeReplicationConfig {
|
||||||
|
private MiniDFSCluster cluster = null;
|
||||||
|
private BlockManager bm;
|
||||||
|
|
||||||
|
@Before
|
||||||
|
public void setup() throws IOException {
|
||||||
|
Configuration config = new Configuration();
|
||||||
|
config.setInt(
|
||||||
|
DFSConfigKeys.DFS_NAMENODE_REPLICATION_MAX_STREAMS_KEY, 8);
|
||||||
|
config.setInt(
|
||||||
|
DFSConfigKeys.DFS_NAMENODE_REPLICATION_STREAMS_HARD_LIMIT_KEY, 10);
|
||||||
|
config.setInt(
|
||||||
|
DFSConfigKeys.DFS_NAMENODE_REPLICATION_WORK_MULTIPLIER_PER_ITERATION,
|
||||||
|
12);
|
||||||
|
|
||||||
|
cluster = new MiniDFSCluster.Builder(config)
|
||||||
|
.nnTopology(MiniDFSNNTopology.simpleSingleNN(0, 0))
|
||||||
|
.numDataNodes(0).build();
|
||||||
|
cluster.waitActive();
|
||||||
|
bm = cluster.getNameNode().getNamesystem().getBlockManager();
|
||||||
|
}
|
||||||
|
|
||||||
|
@After
|
||||||
|
public void teardown() throws IOException {
|
||||||
|
cluster.shutdown();
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Tests to ensure each of the block replication parameters can be passed
|
||||||
|
* updated successfully.
|
||||||
|
*/
|
||||||
|
@Test(timeout = 90000)
|
||||||
|
public void testParamsCanBeReconfigured() throws ReconfigurationException {
|
||||||
|
|
||||||
|
assertEquals(8, bm.getMaxReplicationStreams());
|
||||||
|
assertEquals(10, bm.getReplicationStreamsHardLimit());
|
||||||
|
assertEquals(12, bm.getBlocksReplWorkMultiplier());
|
||||||
|
|
||||||
|
cluster.getNameNode().reconfigurePropertyImpl(
|
||||||
|
DFSConfigKeys.DFS_NAMENODE_REPLICATION_MAX_STREAMS_KEY, "20");
|
||||||
|
cluster.getNameNode().reconfigurePropertyImpl(
|
||||||
|
DFSConfigKeys.DFS_NAMENODE_REPLICATION_STREAMS_HARD_LIMIT_KEY,
|
||||||
|
"22");
|
||||||
|
cluster.getNameNode().reconfigurePropertyImpl(
|
||||||
|
DFSConfigKeys.DFS_NAMENODE_REPLICATION_WORK_MULTIPLIER_PER_ITERATION,
|
||||||
|
"24");
|
||||||
|
|
||||||
|
assertEquals(20, bm.getMaxReplicationStreams());
|
||||||
|
assertEquals(22, bm.getReplicationStreamsHardLimit());
|
||||||
|
assertEquals(24, bm.getBlocksReplWorkMultiplier());
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Tests to ensure reconfiguration fails with a negative, zero or string value
|
||||||
|
* value for each parameter.
|
||||||
|
*/
|
||||||
|
@Test(timeout = 90000)
|
||||||
|
public void testReconfigureFailsWithInvalidValues() throws Exception {
|
||||||
|
String[] keys = new String[]{
|
||||||
|
DFSConfigKeys.DFS_NAMENODE_REPLICATION_MAX_STREAMS_KEY,
|
||||||
|
DFSConfigKeys.DFS_NAMENODE_REPLICATION_STREAMS_HARD_LIMIT_KEY,
|
||||||
|
DFSConfigKeys.DFS_NAMENODE_REPLICATION_WORK_MULTIPLIER_PER_ITERATION
|
||||||
|
};
|
||||||
|
|
||||||
|
// Ensure we cannot set any of the parameters negative
|
||||||
|
for (String key : keys) {
|
||||||
|
ReconfigurationException e =
|
||||||
|
LambdaTestUtils.intercept(ReconfigurationException.class,
|
||||||
|
() -> cluster.getNameNode().reconfigurePropertyImpl(key, "-20"));
|
||||||
|
assertTrue(e.getCause() instanceof IllegalArgumentException);
|
||||||
|
assertEquals(key+" = '-20' is invalid. It should be a "
|
||||||
|
+"positive, non-zero integer value.", e.getCause().getMessage());
|
||||||
|
}
|
||||||
|
// Ensure none of the values were updated from the defaults
|
||||||
|
assertEquals(8, bm.getMaxReplicationStreams());
|
||||||
|
assertEquals(10, bm.getReplicationStreamsHardLimit());
|
||||||
|
assertEquals(12, bm.getBlocksReplWorkMultiplier());
|
||||||
|
|
||||||
|
for (String key : keys) {
|
||||||
|
ReconfigurationException e =
|
||||||
|
LambdaTestUtils.intercept(ReconfigurationException.class,
|
||||||
|
() -> cluster.getNameNode().reconfigurePropertyImpl(key, "0"));
|
||||||
|
assertTrue(e.getCause() instanceof IllegalArgumentException);
|
||||||
|
assertEquals(key+" = '0' is invalid. It should be a "
|
||||||
|
+"positive, non-zero integer value.", e.getCause().getMessage());
|
||||||
|
}
|
||||||
|
|
||||||
|
// Ensure none of the values were updated from the defaults
|
||||||
|
assertEquals(8, bm.getMaxReplicationStreams());
|
||||||
|
assertEquals(10, bm.getReplicationStreamsHardLimit());
|
||||||
|
assertEquals(12, bm.getBlocksReplWorkMultiplier());
|
||||||
|
|
||||||
|
// Ensure none of the parameters can be set to a string value
|
||||||
|
for (String key : keys) {
|
||||||
|
ReconfigurationException e =
|
||||||
|
LambdaTestUtils.intercept(ReconfigurationException.class,
|
||||||
|
() -> cluster.getNameNode().reconfigurePropertyImpl(key, "str"));
|
||||||
|
assertTrue(e.getCause() instanceof NumberFormatException);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Ensure none of the values were updated from the defaults
|
||||||
|
assertEquals(8, bm.getMaxReplicationStreams());
|
||||||
|
assertEquals(10, bm.getReplicationStreamsHardLimit());
|
||||||
|
assertEquals(12, bm.getBlocksReplWorkMultiplier());
|
||||||
|
}
|
||||||
|
}
|
@ -394,7 +394,7 @@ public void testNameNodeGetReconfigurableProperties() throws IOException {
|
|||||||
final List<String> outs = Lists.newArrayList();
|
final List<String> outs = Lists.newArrayList();
|
||||||
final List<String> errs = Lists.newArrayList();
|
final List<String> errs = Lists.newArrayList();
|
||||||
getReconfigurableProperties("namenode", address, outs, errs);
|
getReconfigurableProperties("namenode", address, outs, errs);
|
||||||
assertEquals(7, outs.size());
|
assertEquals(10, outs.size());
|
||||||
assertEquals(DFS_HEARTBEAT_INTERVAL_KEY, outs.get(1));
|
assertEquals(DFS_HEARTBEAT_INTERVAL_KEY, outs.get(1));
|
||||||
assertEquals(DFS_NAMENODE_HEARTBEAT_RECHECK_INTERVAL_KEY, outs.get(2));
|
assertEquals(DFS_NAMENODE_HEARTBEAT_RECHECK_INTERVAL_KEY, outs.get(2));
|
||||||
assertEquals(errs.size(), 0);
|
assertEquals(errs.size(), 0);
|
||||||
|
Loading…
Reference in New Issue
Block a user