HDFS-14042. Fix NPE when PROVIDED storage is missing. Contributed by Virajith Jalaparti.
This commit is contained in:
parent
50f40e0536
commit
f3f5e7ad00
@ -2447,7 +2447,7 @@ public long getProvidedCapacity() {
|
||||
return providedStorageMap.getCapacity();
|
||||
}
|
||||
|
||||
public void updateHeartbeat(DatanodeDescriptor node, StorageReport[] reports,
|
||||
void updateHeartbeat(DatanodeDescriptor node, StorageReport[] reports,
|
||||
long cacheCapacity, long cacheUsed, int xceiverCount, int failedVolumes,
|
||||
VolumeFailureSummary volumeFailureSummary) {
|
||||
|
||||
@ -2458,6 +2458,17 @@ public void updateHeartbeat(DatanodeDescriptor node, StorageReport[] reports,
|
||||
failedVolumes, volumeFailureSummary);
|
||||
}
|
||||
|
||||
void updateHeartbeatState(DatanodeDescriptor node,
|
||||
StorageReport[] reports, long cacheCapacity, long cacheUsed,
|
||||
int xceiverCount, int failedVolumes,
|
||||
VolumeFailureSummary volumeFailureSummary) {
|
||||
for (StorageReport report: reports) {
|
||||
providedStorageMap.updateStorage(node, report.getStorage());
|
||||
}
|
||||
node.updateHeartbeatState(reports, cacheCapacity, cacheUsed, xceiverCount,
|
||||
failedVolumes, volumeFailureSummary);
|
||||
}
|
||||
|
||||
/**
|
||||
* StatefulBlockInfo is used to build the "toUC" list, which is a list of
|
||||
* updates to the information about under-construction blocks.
|
||||
|
@ -373,7 +373,7 @@ public int numBlocks() {
|
||||
/**
|
||||
* Updates stats from datanode heartbeat.
|
||||
*/
|
||||
public void updateHeartbeat(StorageReport[] reports, long cacheCapacity,
|
||||
void updateHeartbeat(StorageReport[] reports, long cacheCapacity,
|
||||
long cacheUsed, int xceiverCount, int volFailures,
|
||||
VolumeFailureSummary volumeFailureSummary) {
|
||||
updateHeartbeatState(reports, cacheCapacity, cacheUsed, xceiverCount,
|
||||
@ -384,7 +384,7 @@ public void updateHeartbeat(StorageReport[] reports, long cacheCapacity,
|
||||
/**
|
||||
* process datanode heartbeat or stats initialization.
|
||||
*/
|
||||
public void updateHeartbeatState(StorageReport[] reports, long cacheCapacity,
|
||||
void updateHeartbeatState(StorageReport[] reports, long cacheCapacity,
|
||||
long cacheUsed, int xceiverCount, int volFailures,
|
||||
VolumeFailureSummary volumeFailureSummary) {
|
||||
updateStorageStats(reports, cacheCapacity, cacheUsed, xceiverCount,
|
||||
|
@ -251,7 +251,7 @@ synchronized void updateLifeline(final DatanodeDescriptor node,
|
||||
// updateHeartbeat, because we don't want to modify the
|
||||
// heartbeatedSinceRegistration flag. Arrival of a lifeline message does
|
||||
// not count as arrival of the first heartbeat.
|
||||
node.updateHeartbeatState(reports, cacheCapacity, cacheUsed,
|
||||
blockManager.updateHeartbeatState(node, reports, cacheCapacity, cacheUsed,
|
||||
xceiverCount, failedVolumes, volumeFailureSummary);
|
||||
stats.add(node);
|
||||
}
|
||||
|
@ -18,6 +18,7 @@
|
||||
package org.apache.hadoop.hdfs.server.datanode;
|
||||
|
||||
import static java.util.concurrent.TimeUnit.SECONDS;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_DATANODE_DATA_DIR_KEY;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_DATANODE_LIFELINE_INTERVAL_SECONDS_KEY;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_HEARTBEAT_INTERVAL_KEY;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_HEARTBEAT_RECHECK_INTERVAL_KEY;
|
||||
@ -196,6 +197,10 @@ public void testSendLifelineIfHeartbeatBlocked() throws Exception {
|
||||
namesystem.getNumDeadDataNodes());
|
||||
assertEquals("Expect DataNode not marked stale due to lifeline.", 0,
|
||||
namesystem.getNumStaleDataNodes());
|
||||
// add a new volume on the next heartbeat
|
||||
cluster.getDataNodes().get(0).reconfigurePropertyImpl(
|
||||
DFS_DATANODE_DATA_DIR_KEY,
|
||||
cluster.getDataDirectory().concat("/data-new"));
|
||||
}
|
||||
|
||||
// Verify that we did in fact call the lifeline RPC.
|
||||
|
Loading…
Reference in New Issue
Block a user