YARN-6667. Handle containerId duplicate without failing the heartbeat in Federation Interceptor. (#4810)

This commit is contained in:
slfan1989 2022-09-03 01:25:26 +08:00 committed by GitHub
parent 7b239a80fe
commit 3a96de7756
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
2 changed files with 162 additions and 11 deletions

View File

@ -85,6 +85,7 @@
import org.apache.hadoop.yarn.server.federation.policies.exceptions.FederationPolicyInitializationException;
import org.apache.hadoop.yarn.server.federation.resolver.SubClusterResolver;
import org.apache.hadoop.yarn.server.federation.store.records.SubClusterId;
import org.apache.hadoop.yarn.server.federation.store.records.SubClusterInfo;
import org.apache.hadoop.yarn.server.federation.utils.FederationRegistryClient;
import org.apache.hadoop.yarn.server.federation.utils.FederationStateStoreFacade;
import org.apache.hadoop.yarn.server.uam.UnmanagedAMPoolManager;
@ -1475,6 +1476,7 @@ protected void mergeAllocateResponse(AllocateResponse homeResponse,
private void cacheAllocatedContainers(List<Container> containers,
SubClusterId subClusterId) {
for (Container container : containers) {
SubClusterId chooseSubClusterId = SubClusterId.newInstance(subClusterId.toString());
LOG.debug("Adding container {}", container);
if (this.containerIdToSubClusterIdMap.containsKey(container.getId())) {
@ -1497,22 +1499,53 @@ private void cacheAllocatedContainers(List<Container> containers,
+ " from same sub-cluster: {}, so ignoring.",
container.getId(), subClusterId);
} else {
LOG.info("Duplicate containerID found in the allocated containers. " +
"try to re-pick the sub-cluster.");
// The same container allocation from different sub-clusters,
// something is wrong.
// TODO: YARN-6667 if some subcluster RM is configured wrong, we
// should not fail the entire heartbeat.
try {
boolean existAllocatedScHealth = isSCHealth(existingSubClusterId);
boolean newAllocatedScHealth = isSCHealth(subClusterId);
if (existAllocatedScHealth) {
// If the previous RM which allocated Container is normal,
// the previous RM will be used first
LOG.info("Use Previous Allocated Container's subCluster. " +
"ContainerId: {} ApplicationId: {} From RM: {}.", this.attemptId,
container.getId(), existingSubClusterId);
chooseSubClusterId = existingSubClusterId;
} else if (newAllocatedScHealth) {
// If the previous RM which allocated Container is abnormal,
// but the RM of the newly allocated Container is normal, use the new RM
LOG.info("Use Newly Allocated Container's subCluster. " +
"ApplicationId: {} ContainerId: {} From RM: {}.", this.attemptId,
container.getId(), subClusterId);
chooseSubClusterId = subClusterId;
} else {
// There is a very small probability that an exception will be thrown.
// The RM of the previously allocated Container
// and the RM of the newly allocated Container are not normal.
throw new YarnRuntimeException(
"Duplicate containerID found in the allocated containers. This"
+ " can happen if the RM epoch is not configured properly."
+ " ContainerId: " + container.getId().toString()
+ " ApplicationId: " + this.attemptId + " From RM: "
+ subClusterId
+ " . Previous container was from sub-cluster: "
+ existingSubClusterId);
" Can't use any subCluster because an exception occurred" +
" ContainerId: " + container.getId() + " ApplicationId: " + this.attemptId +
" From RM: " + subClusterId + ". " +
" Previous Container was From subCluster: " + existingSubClusterId);
}
} catch (Exception ex) {
// An exception occurred
throw new YarnRuntimeException(
" Can't use any subCluster because an exception occurred" +
" ContainerId: " + container.getId() + " ApplicationId: " + this.attemptId +
" From RM: " + subClusterId + ". " +
" Previous Container was From subCluster: " + existingSubClusterId, ex);
}
}
}
this.containerIdToSubClusterIdMap.put(container.getId(), subClusterId);
this.containerIdToSubClusterIdMap.put(container.getId(), chooseSubClusterId);
}
}
@ -1761,4 +1794,25 @@ public static <T> boolean isNullOrEmpty(Collection<T> c) {
public static <T1, T2> boolean isNullOrEmpty(Map<T1, T2> c) {
return (c == null || c.size() == 0);
}
@VisibleForTesting
protected void cacheAllocatedContainersForSubClusterId(
List<Container> containers, SubClusterId subClusterId) {
cacheAllocatedContainers(containers, subClusterId);
}
@VisibleForTesting
protected Map<ContainerId, SubClusterId> getContainerIdToSubClusterIdMap() {
return containerIdToSubClusterIdMap;
}
private boolean isSCHealth(SubClusterId subClusterId) throws YarnException {
Set<SubClusterId> timeOutScs = getTimedOutSCs(true);
SubClusterInfo subClusterInfo = federationFacade.getSubCluster(subClusterId);
if (timeOutScs.contains(subClusterId) ||
subClusterInfo == null || subClusterInfo.getState().isUnusable()) {
return false;
}
return true;
}
}

View File

@ -36,6 +36,7 @@
import org.apache.hadoop.registry.client.api.RegistryOperations;
import org.apache.hadoop.registry.client.impl.FSRegistryOperationsService;
import org.apache.hadoop.security.UserGroupInformation;
import org.apache.hadoop.test.LambdaTestUtils;
import org.apache.hadoop.yarn.api.protocolrecords.AllocateRequest;
import org.apache.hadoop.yarn.api.protocolrecords.AllocateResponse;
import org.apache.hadoop.yarn.api.protocolrecords.FinishApplicationMasterRequest;
@ -61,6 +62,7 @@
import org.apache.hadoop.yarn.exceptions.ApplicationMasterNotRegisteredException;
import org.apache.hadoop.yarn.exceptions.InvalidApplicationMasterRequestException;
import org.apache.hadoop.yarn.exceptions.YarnException;
import org.apache.hadoop.yarn.exceptions.YarnRuntimeException;
import org.apache.hadoop.yarn.server.MockResourceManagerFacade;
import org.apache.hadoop.yarn.server.federation.policies.manager.UniformBroadcastPolicyManager;
import org.apache.hadoop.yarn.server.federation.store.impl.MemoryFederationStateStore;
@ -970,6 +972,101 @@ private PreemptionMessage createDummyPreemptionMessage(
return preemptionMessage;
}
@Test
public void testSameContainerFromDiffRM() throws IOException, InterruptedException {
UserGroupInformation ugi =
interceptor.getUGIWithToken(interceptor.getAttemptId());
ugi.doAs((PrivilegedExceptionAction<Object>) () -> {
// Register the application
RegisterApplicationMasterRequest registerReq =
Records.newRecord(RegisterApplicationMasterRequest.class);
registerReq.setHost(Integer.toString(testAppId));
registerReq.setRpcPort(0);
registerReq.setTrackingUrl("");
RegisterApplicationMasterResponse registerResponse =
interceptor.registerApplicationMaster(registerReq);
Assert.assertNotNull(registerResponse);
lastResponseId = 0;
Assert.assertEquals(0, interceptor.getUnmanagedAMPoolSize());
// Allocate the first batch of containers, with sc1 active
SubClusterId subClusterId1 = SubClusterId.newInstance("SC-1");
registerSubCluster(subClusterId1);
int numberOfContainers = 3;
List<Container> containers =
getContainersAndAssert(numberOfContainers, numberOfContainers);
Assert.assertNotNull(containers);
Assert.assertEquals(3, containers.size());
// with sc2 active
SubClusterId subClusterId2 = SubClusterId.newInstance("SC-2");
registerSubCluster(subClusterId2);
// 1.Container has been registered to SubCluster1, try to register the same Container
// to SubCluster2.
// Because SubCluster1 is in normal state at this time,
// So the SubCluster corresponding to Container should be SubCluster1
interceptor.cacheAllocatedContainersForSubClusterId(containers, subClusterId2);
Map<ContainerId, SubClusterId> cIdToSCMap = interceptor.getContainerIdToSubClusterIdMap();
for (SubClusterId subClusterId : cIdToSCMap.values()) {
Assert.assertNotNull(subClusterId);
Assert.assertEquals(subClusterId1, subClusterId);
}
// 2.Deregister SubCluster1, Register the same Containers to SubCluster2
// So the SubCluster corresponding to Container should be SubCluster2
deRegisterSubCluster(subClusterId1);
interceptor.cacheAllocatedContainersForSubClusterId(containers, subClusterId2);
Map<ContainerId, SubClusterId> cIdToSCMap2 = interceptor.getContainerIdToSubClusterIdMap();
for (SubClusterId subClusterId : cIdToSCMap2.values()) {
Assert.assertNotNull(subClusterId);
Assert.assertEquals(subClusterId2, subClusterId);
}
// 3.Deregister subClusterId2, Register the same Containers to SubCluster1
// Because both SubCluster1 and SubCluster2 are abnormal at this time,
// an exception will be thrown when registering the first Container.
deRegisterSubCluster(subClusterId2);
Container container1 = containers.get(0);
Assert.assertNotNull(container1);
String errMsg =
" Can't use any subCluster because an exception occurred" +
" ContainerId: " + container1.getId() +
" ApplicationId: " + interceptor.getAttemptId() +
" From RM: " + subClusterId1 + ". " +
" Previous Container was From subCluster: " + subClusterId2;
LambdaTestUtils.intercept(YarnRuntimeException.class, errMsg,
() -> interceptor.cacheAllocatedContainersForSubClusterId(containers, subClusterId1));
// 4. register SubCluster1, re-register the Container,
// and try to finish application
registerSubCluster(subClusterId1);
interceptor.cacheAllocatedContainersForSubClusterId(containers, subClusterId1);
releaseContainersAndAssert(containers);
// Finish the application
FinishApplicationMasterRequest finishReq =
Records.newRecord(FinishApplicationMasterRequest.class);
finishReq.setDiagnostics("");
finishReq.setTrackingUrl("");
finishReq.setFinalApplicationStatus(FinalApplicationStatus.SUCCEEDED);
FinishApplicationMasterResponse finishResponse =
interceptor.finishApplicationMaster(finishReq);
Assert.assertNotNull(finishResponse);
Assert.assertTrue(finishResponse.getIsUnregistered());
return null;
});
}
@Test
public void testBatchFinishApplicationMaster() throws IOException, InterruptedException {