HDFS-14968. Add ability to log stale datanodes. Contributed by Ahmed Hussein.

This commit is contained in:
Kihwal Lee 2020-01-22 09:14:38 -06:00
parent d40d7cc4f9
commit bd03053ea2
3 changed files with 109 additions and 13 deletions

View File

@ -483,6 +483,11 @@ public class DFSConfigKeys extends CommonConfigurationKeys {
// Whether to enable datanode's stale state detection and usage for writes
public static final String DFS_NAMENODE_AVOID_STALE_DATANODE_FOR_WRITE_KEY = "dfs.namenode.avoid.write.stale.datanode";
public static final boolean DFS_NAMENODE_AVOID_STALE_DATANODE_FOR_WRITE_DEFAULT = false;
// enable and disable logging datanode staleness. Disabled by default.
public static final String DFS_NAMENODE_ENABLE_LOG_STALE_DATANODE_KEY =
"dfs.namenode.enable.log.stale.datanode";
public static final boolean DFS_NAMENODE_ENABLE_LOG_STALE_DATANODE_DEFAULT =
false;
// The default value of the time interval for marking datanodes as stale
public static final String DFS_NAMENODE_STALE_DATANODE_INTERVAL_KEY = "dfs.namenode.stale.datanode.interval";
public static final long DFS_NAMENODE_STALE_DATANODE_INTERVAL_DEFAULT = 30 * 1000; // 30s

View File

@ -18,8 +18,10 @@
package org.apache.hadoop.hdfs.server.blockmanagement;
import java.util.ArrayList;
import java.util.HashSet;
import java.util.List;
import java.util.Map;
import java.util.Set;
import java.util.concurrent.TimeUnit;
import org.apache.hadoop.conf.Configuration;
@ -43,7 +45,15 @@
*/
class HeartbeatManager implements DatanodeStatistics {
static final Logger LOG = LoggerFactory.getLogger(HeartbeatManager.class);
private static final String REPORT_DELTA_STALE_DN_HEADER =
"StaleNodes Report: [New Stale Nodes]: %d";
private static final String REPORT_STALE_DN_LINE_ENTRY = "%n\t %s";
private static final String REPORT_STALE_DN_LINE_TAIL = ", %s";
private static final String REPORT_REMOVE_DEAD_NODE_ENTRY =
"StaleNodes Report: [Remove DeadNode]: %s";
private static final String REPORT_REMOVE_STALE_NODE_ENTRY =
"StaleNodes Report: [Remove StaleNode]: %s";
private static final int REPORT_STALE_NODE_NODES_PER_LINE = 10;
/**
* Stores a subset of the datanodeMap in DatanodeManager,
* containing nodes that are considered alive.
@ -56,14 +66,19 @@ class HeartbeatManager implements DatanodeStatistics {
/** Statistics, which are synchronized by the heartbeat manager lock. */
private final DatanodeStats stats = new DatanodeStats();
/** The time period to check for expired datanodes */
/** The time period to check for expired datanodes. */
private final long heartbeatRecheckInterval;
/** Heartbeat monitor thread */
/** Heartbeat monitor thread. */
private final Daemon heartbeatThread = new Daemon(new Monitor());
private final StopWatch heartbeatStopWatch = new StopWatch();
final Namesystem namesystem;
final BlockManager blockManager;
/** Enable log for datanode staleness. */
private final boolean enableLogStaleNodes;
/** reports for stale datanodes. */
private final Set<DatanodeDescriptor> staleDataNodes = new HashSet<>();
HeartbeatManager(final Namesystem namesystem,
final BlockManager blockManager, final Configuration conf) {
@ -78,6 +93,9 @@ class HeartbeatManager implements DatanodeStatistics {
long staleInterval = conf.getLong(
DFSConfigKeys.DFS_NAMENODE_STALE_DATANODE_INTERVAL_KEY,
DFSConfigKeys.DFS_NAMENODE_STALE_DATANODE_INTERVAL_DEFAULT);// 30s
enableLogStaleNodes = conf.getBoolean(
DFSConfigKeys.DFS_NAMENODE_ENABLE_LOG_STALE_DATANODE_KEY,
DFSConfigKeys.DFS_NAMENODE_ENABLE_LOG_STALE_DATANODE_DEFAULT);
if (avoidStaleDataNodesForWrite && staleInterval < recheckInterval) {
this.heartbeatRecheckInterval = staleInterval;
@ -228,6 +246,7 @@ synchronized void removeDatanode(DatanodeDescriptor node) {
if (node.isAlive()) {
stats.subtract(node);
datanodes.remove(node);
removeNodeFromStaleList(node);
node.setAlive(false);
}
}
@ -323,6 +342,59 @@ boolean shouldAbortHeartbeatCheck(long offset) {
return elapsed + offset > heartbeatRecheckInterval;
}
/**
* Remove deadNode from StaleNodeList if it exists.
* This method assumes that it is called inside a synchronized block.
*
* @param d node descriptor to be marked as dead.
* @return true if the node was already on the stale list.
*/
private boolean removeNodeFromStaleList(DatanodeDescriptor d) {
return removeNodeFromStaleList(d, true);
}
/**
* Remove node from StaleNodeList if it exists.
* If enabled, the log will show whether the node is removed from list because
* it is dead or not.
* This method assumes that it is called inside a synchronized block.
*
* @param d node descriptor to be marked as dead.
* @param isDead
* @return true if the node was already in the stale list.
*/
private boolean removeNodeFromStaleList(DatanodeDescriptor d,
boolean isDead) {
boolean result = false;
result = staleDataNodes.remove(d);
if (enableLogStaleNodes && result) {
LOG.info(String.format(isDead ?
REPORT_REMOVE_DEAD_NODE_ENTRY : REPORT_REMOVE_STALE_NODE_ENTRY,
d));
}
return result;
}
/**
* Dump the new stale data nodes added since last heartbeat check.
*
* @param staleNodes list of datanodes added in the last heartbeat check.
*/
private void dumpStaleNodes(List<DatanodeDescriptor> staleNodes) {
// log nodes detected as stale
if (enableLogStaleNodes && (!staleNodes.isEmpty())) {
StringBuilder staleLogMSG =
new StringBuilder(String.format(REPORT_DELTA_STALE_DN_HEADER,
staleNodes.size()));
for (int ind = 0; ind < staleNodes.size(); ind++) {
String logFormat = (ind % REPORT_STALE_NODE_NODES_PER_LINE == 0) ?
REPORT_STALE_DN_LINE_ENTRY : REPORT_STALE_DN_LINE_TAIL;
staleLogMSG.append(String.format(logFormat, staleNodes.get(ind)));
}
LOG.info(staleLogMSG.toString());
}
}
/**
* Check if there are any expired heartbeats, and if so,
* whether any blocks have to be re-replicated.
@ -365,9 +437,9 @@ void heartbeatCheck() {
// locate the first failed storage that isn't on a dead node.
DatanodeStorageInfo failedStorage = null;
// check the number of stale nodes
int numOfStaleNodes = 0;
// check the number of stale storages
int numOfStaleStorages = 0;
List<DatanodeDescriptor> staleNodes = new ArrayList<>();
synchronized(this) {
for (DatanodeDescriptor d : datanodes) {
// check if an excessive GC pause has occurred
@ -377,13 +449,21 @@ void heartbeatCheck() {
if (dead == null && dm.isDatanodeDead(d)) {
stats.incrExpiredHeartbeats();
dead = d;
// remove the node from stale list to adjust the stale list size
// before setting the stale count of the DatanodeManager
removeNodeFromStaleList(d);
} else {
if (d.isStale(dm.getStaleInterval())) {
if (staleDataNodes.add(d)) {
// the node is n
staleNodes.add(d);
}
} else {
// remove the node if it is no longer stale
removeNodeFromStaleList(d, false);
}
}
if (d.isStale(dm.getStaleInterval())) {
LOG.warn(String.format("Stale datanode {}."
+ " No heartbeat received since last {} milliseconds"),
d.getName(), dm.getStaleInterval());
numOfStaleNodes++;
}
DatanodeStorageInfo[] storageInfos = d.getStorageInfos();
for(DatanodeStorageInfo storageInfo : storageInfos) {
if (storageInfo.areBlockContentsStale()) {
@ -396,18 +476,21 @@ void heartbeatCheck() {
failedStorage = storageInfo;
}
}
}
// Set the number of stale nodes in the DatanodeManager
dm.setNumStaleNodes(numOfStaleNodes);
dm.setNumStaleNodes(staleDataNodes.size());
dm.setNumStaleStorages(numOfStaleStorages);
}
// log nodes detected as stale since last heartBeat
dumpStaleNodes(staleNodes);
allAlive = dead == null && failedStorage == null;
if (!allAlive && namesystem.isInStartupSafeMode()) {
return;
}
if (dead != null) {
// acquire the fsnamesystem lock, and then remove the dead node.
namesystem.writeLock();

View File

@ -2040,6 +2040,14 @@
</description>
</property>
<property>
<name>dfs.namenode.enable.log.stale.datanode</name>
<value>false</value>
<description>
Enable and disable logging datanode staleness. Disabled by default.
</description>
</property>
<property>
<name>dfs.namenode.stale.datanode.interval</name>
<value>30000</value>