YARN-6979. Add flag to notify all types of container updates to NM via NodeHeartbeatResponse. (Kartheek Muthyala via asuresh)
This commit is contained in:
parent
436c2638f9
commit
8410d862d3
@ -179,7 +179,7 @@ public Set<String> getNodeLabels() {
|
||||
}
|
||||
|
||||
@Override
|
||||
public void updateNodeHeartbeatResponseForContainersDecreasing(
|
||||
public void updateNodeHeartbeatResponseForUpdatedContainers(
|
||||
NodeHeartbeatResponse response) {
|
||||
// TODO Auto-generated method stub
|
||||
|
||||
|
@ -168,7 +168,7 @@ public Set<String> getNodeLabels() {
|
||||
}
|
||||
|
||||
@Override
|
||||
public void updateNodeHeartbeatResponseForContainersDecreasing(
|
||||
public void updateNodeHeartbeatResponseForUpdatedContainers(
|
||||
NodeHeartbeatResponse response) {
|
||||
// TODO Auto-generated method stub
|
||||
}
|
||||
|
@ -167,6 +167,10 @@ private static void addDeprecatedKeys() {
|
||||
public static final String RM_APPLICATION_MASTER_SERVICE_PROCESSORS =
|
||||
RM_PREFIX + "application-master-service.processors";
|
||||
|
||||
public static final String RM_AUTO_UPDATE_CONTAINERS =
|
||||
RM_PREFIX + "auto-update.containers";
|
||||
public static final boolean DEFAULT_RM_AUTO_UPDATE_CONTAINERS = false;
|
||||
|
||||
/** The actual bind address for the RM.*/
|
||||
public static final String RM_BIND_HOST =
|
||||
RM_PREFIX + "bind-host";
|
||||
|
@ -72,6 +72,14 @@
|
||||
<value></value>
|
||||
</property>
|
||||
|
||||
<property>
|
||||
<description>
|
||||
If set to true, then ALL container updates will be automatically sent to
|
||||
the NM in the next heartbeat</description>
|
||||
<name>yarn.resourcemanager.auto-update.containers</name>
|
||||
<value>false</value>
|
||||
</property>
|
||||
|
||||
<property>
|
||||
<description>The number of threads used to handle applications manager requests.</description>
|
||||
<name>yarn.resourcemanager.client.thread-count</name>
|
||||
|
@ -104,10 +104,10 @@ public abstract void setAreNodeLabelsAcceptedByRM(
|
||||
|
||||
public abstract void setResource(Resource resource);
|
||||
|
||||
public abstract List<Container> getContainersToDecrease();
|
||||
public abstract List<Container> getContainersToUpdate();
|
||||
|
||||
public abstract void addAllContainersToDecrease(
|
||||
Collection<Container> containersToDecrease);
|
||||
public abstract void addAllContainersToUpdate(
|
||||
Collection<Container> containersToUpdate);
|
||||
|
||||
public abstract ContainerQueuingLimit getContainerQueuingLimit();
|
||||
|
||||
|
@ -75,7 +75,7 @@ public class NodeHeartbeatResponsePBImpl extends NodeHeartbeatResponse {
|
||||
private MasterKey containerTokenMasterKey = null;
|
||||
private MasterKey nmTokenMasterKey = null;
|
||||
private ContainerQueuingLimit containerQueuingLimit = null;
|
||||
private List<Container> containersToDecrease = null;
|
||||
private List<Container> containersToUpdate = null;
|
||||
private List<SignalContainerRequest> containersToSignal = null;
|
||||
|
||||
public NodeHeartbeatResponsePBImpl() {
|
||||
@ -119,8 +119,8 @@ private void mergeLocalToBuilder() {
|
||||
if (this.systemCredentials != null) {
|
||||
addSystemCredentialsToProto();
|
||||
}
|
||||
if (this.containersToDecrease != null) {
|
||||
addContainersToDecreaseToProto();
|
||||
if (this.containersToUpdate != null) {
|
||||
addContainersToUpdateToProto();
|
||||
}
|
||||
if (this.containersToSignal != null) {
|
||||
addContainersToSignalToProto();
|
||||
@ -499,39 +499,39 @@ public void remove() {
|
||||
builder.addAllApplicationsToCleanup(iterable);
|
||||
}
|
||||
|
||||
private void initContainersToDecrease() {
|
||||
if (this.containersToDecrease != null) {
|
||||
private void initContainersToUpdate() {
|
||||
if (this.containersToUpdate != null) {
|
||||
return;
|
||||
}
|
||||
NodeHeartbeatResponseProtoOrBuilder p = viaProto ? proto : builder;
|
||||
List<ContainerProto> list = p.getContainersToDecreaseList();
|
||||
this.containersToDecrease = new ArrayList<>();
|
||||
List<ContainerProto> list = p.getContainersToUpdateList();
|
||||
this.containersToUpdate = new ArrayList<>();
|
||||
|
||||
for (ContainerProto c : list) {
|
||||
this.containersToDecrease.add(convertFromProtoFormat(c));
|
||||
this.containersToUpdate.add(convertFromProtoFormat(c));
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public List<Container> getContainersToDecrease() {
|
||||
initContainersToDecrease();
|
||||
return this.containersToDecrease;
|
||||
public List<Container> getContainersToUpdate() {
|
||||
initContainersToUpdate();
|
||||
return this.containersToUpdate;
|
||||
}
|
||||
|
||||
@Override
|
||||
public void addAllContainersToDecrease(
|
||||
final Collection<Container> containersToDecrease) {
|
||||
if (containersToDecrease == null) {
|
||||
public void addAllContainersToUpdate(
|
||||
final Collection<Container> containersToBeUpdated) {
|
||||
if (containersToBeUpdated == null) {
|
||||
return;
|
||||
}
|
||||
initContainersToDecrease();
|
||||
this.containersToDecrease.addAll(containersToDecrease);
|
||||
initContainersToUpdate();
|
||||
this.containersToUpdate.addAll(containersToBeUpdated);
|
||||
}
|
||||
|
||||
private void addContainersToDecreaseToProto() {
|
||||
private void addContainersToUpdateToProto() {
|
||||
maybeInitBuilder();
|
||||
builder.clearContainersToDecrease();
|
||||
if (this.containersToDecrease == null) {
|
||||
builder.clearContainersToUpdate();
|
||||
if (this.containersToUpdate == null) {
|
||||
return;
|
||||
}
|
||||
Iterable<ContainerProto> iterable = new
|
||||
@ -539,7 +539,7 @@ private void addContainersToDecreaseToProto() {
|
||||
@Override
|
||||
public Iterator<ContainerProto> iterator() {
|
||||
return new Iterator<ContainerProto>() {
|
||||
private Iterator<Container> iter = containersToDecrease.iterator();
|
||||
private Iterator<Container> iter = containersToUpdate.iterator();
|
||||
@Override
|
||||
public boolean hasNext() {
|
||||
return iter.hasNext();
|
||||
@ -555,7 +555,7 @@ public void remove() {
|
||||
};
|
||||
}
|
||||
};
|
||||
builder.addAllContainersToDecrease(iterable);
|
||||
builder.addAllContainersToUpdate(iterable);
|
||||
}
|
||||
|
||||
@Override
|
||||
|
@ -111,11 +111,14 @@ message NodeHeartbeatResponseProto {
|
||||
repeated ContainerIdProto containers_to_be_removed_from_nm = 9;
|
||||
repeated SystemCredentialsForAppsProto system_credentials_for_apps = 10;
|
||||
optional bool areNodeLabelsAcceptedByRM = 11 [default = false];
|
||||
// to be deprecated in favour of containers_to_update
|
||||
repeated ContainerProto containers_to_decrease = 12;
|
||||
repeated SignalContainerRequestProto containers_to_signal = 13;
|
||||
optional ResourceProto resource = 14;
|
||||
optional ContainerQueuingLimitProto container_queuing_limit = 15;
|
||||
repeated AppCollectorsMapProto app_collectors_map = 16;
|
||||
// to be used in place of containers_to_decrease
|
||||
repeated ContainerProto containers_to_update = 17;
|
||||
}
|
||||
|
||||
message ContainerQueuingLimitProto {
|
||||
|
@ -180,14 +180,14 @@ public void testNodeHeartbeatResponsePBImplWithRMAcceptLbls() {
|
||||
@Test
|
||||
public void testNodeHeartbeatResponsePBImplWithDecreasedContainers() {
|
||||
NodeHeartbeatResponsePBImpl original = new NodeHeartbeatResponsePBImpl();
|
||||
original.addAllContainersToDecrease(
|
||||
original.addAllContainersToUpdate(
|
||||
Arrays.asList(getDecreasedContainer(1, 2, 2048, 2),
|
||||
getDecreasedContainer(2, 3, 1024, 1)));
|
||||
NodeHeartbeatResponsePBImpl copy =
|
||||
new NodeHeartbeatResponsePBImpl(original.getProto());
|
||||
assertEquals(1, copy.getContainersToDecrease().get(0)
|
||||
assertEquals(1, copy.getContainersToUpdate().get(0)
|
||||
.getId().getContainerId());
|
||||
assertEquals(1024, copy.getContainersToDecrease().get(1)
|
||||
assertEquals(1024, copy.getContainersToUpdate().get(1)
|
||||
.getResource().getMemorySize());
|
||||
}
|
||||
|
||||
|
@ -1099,7 +1099,7 @@ public void run() {
|
||||
parseCredentials(systemCredentials));
|
||||
}
|
||||
List<org.apache.hadoop.yarn.api.records.Container>
|
||||
containersToDecrease = response.getContainersToDecrease();
|
||||
containersToDecrease = response.getContainersToUpdate();
|
||||
if (!containersToDecrease.isEmpty()) {
|
||||
dispatcher.getEventHandler().handle(
|
||||
new CMgrDecreaseContainersResourceEvent(
|
||||
|
@ -166,6 +166,7 @@
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.Collection;
|
||||
import java.util.EnumSet;
|
||||
import java.util.HashMap;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
@ -1241,10 +1242,19 @@ private void updateContainerInternal(ContainerId containerId,
|
||||
org.apache.hadoop.yarn.server.nodemanager.
|
||||
containermanager.container.ContainerState currentState =
|
||||
container.getContainerState();
|
||||
if (currentState != org.apache.hadoop.yarn.server.
|
||||
nodemanager.containermanager.container.ContainerState.RUNNING &&
|
||||
currentState != org.apache.hadoop.yarn.server.
|
||||
nodemanager.containermanager.container.ContainerState.SCHEDULED) {
|
||||
EnumSet<org.apache.hadoop.yarn.server.nodemanager.containermanager
|
||||
.container.ContainerState> allowedStates = EnumSet.of(
|
||||
org.apache.hadoop.yarn.server.nodemanager.containermanager.container
|
||||
.ContainerState.RUNNING,
|
||||
org.apache.hadoop.yarn.server.nodemanager.containermanager.container
|
||||
.ContainerState.SCHEDULED,
|
||||
org.apache.hadoop.yarn.server.nodemanager.containermanager.container
|
||||
.ContainerState.LOCALIZING,
|
||||
org.apache.hadoop.yarn.server.nodemanager.containermanager.container
|
||||
.ContainerState.REINITIALIZING,
|
||||
org.apache.hadoop.yarn.server.nodemanager.containermanager.container
|
||||
.ContainerState.RELAUNCHING);
|
||||
if (!allowedStates.contains(currentState)) {
|
||||
throw RPCUtil.getRemoteException("Container " + containerId.toString()
|
||||
+ " is in " + currentState.name() + " state."
|
||||
+ " Resource can only be changed when a container is in"
|
||||
@ -1279,17 +1289,12 @@ private void updateContainerInternal(ContainerId containerId,
|
||||
org.apache.hadoop.yarn.api.records.Container.newInstance(
|
||||
containerId, null, null, targetResource, null, null,
|
||||
currentExecType);
|
||||
} else {
|
||||
increasedContainer =
|
||||
org.apache.hadoop.yarn.api.records.Container.newInstance(
|
||||
containerId, null, null, currentResource, null, null,
|
||||
targetExecType);
|
||||
}
|
||||
if (context.getIncreasedContainers().putIfAbsent(containerId,
|
||||
increasedContainer) != null){
|
||||
throw RPCUtil.getRemoteException("Container " + containerId.toString()
|
||||
+ " resource is being increased -or- " +
|
||||
"is undergoing ExecutionType promoted.");
|
||||
if (context.getIncreasedContainers().putIfAbsent(containerId,
|
||||
increasedContainer) != null){
|
||||
throw RPCUtil.getRemoteException("Container " + containerId.toString()
|
||||
+ " resource is being increased -or- " +
|
||||
"is undergoing ExecutionType promoted.");
|
||||
}
|
||||
}
|
||||
}
|
||||
this.readLock.lock();
|
||||
|
@ -173,6 +173,7 @@ private void onUpdateContainer(UpdateContainerSchedulerEvent updateEvent) {
|
||||
new ChangeMonitoringContainerResourceEvent(containerId,
|
||||
updateEvent.getUpdatedToken().getResource()));
|
||||
} else {
|
||||
// Is Queued or localizing..
|
||||
updateEvent.getContainer().setContainerTokenIdentifier(
|
||||
updateEvent.getUpdatedToken());
|
||||
}
|
||||
|
@ -551,7 +551,7 @@ public NodeHeartbeatResponse nodeHeartbeat(NodeHeartbeatRequest request)
|
||||
getResponseId() + 1, NodeAction.NORMAL, null, null, null, null,
|
||||
nextHeartBeatInterval);
|
||||
rmNode.updateNodeHeartbeatResponseForCleanup(nodeHeartBeatResponse);
|
||||
rmNode.updateNodeHeartbeatResponseForContainersDecreasing(
|
||||
rmNode.updateNodeHeartbeatResponseForUpdatedContainers(
|
||||
nodeHeartBeatResponse);
|
||||
|
||||
populateKeys(request, nodeHeartBeatResponse);
|
||||
|
@ -51,8 +51,7 @@
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptEventType;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.event.RMAppAttemptContainerFinishedEvent;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNodeCleanContainerEvent;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode
|
||||
.RMNodeDecreaseContainerEvent;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNodeUpdateContainerEvent;
|
||||
import org.apache.hadoop.yarn.server.scheduler.SchedulerRequestKey;
|
||||
import org.apache.hadoop.yarn.state.InvalidStateTransitionException;
|
||||
import org.apache.hadoop.yarn.state.MultipleArcTransition;
|
||||
@ -284,7 +283,6 @@ public void setContainer(Container container) {
|
||||
@Override
|
||||
public RMContainerState getState() {
|
||||
this.readLock.lock();
|
||||
|
||||
try {
|
||||
return this.stateMachine.getCurrentState();
|
||||
} finally {
|
||||
@ -598,7 +596,7 @@ public void transition(RMContainerImpl container, RMContainerEvent event) {
|
||||
RMContainerUpdatesAcquiredEvent acquiredEvent =
|
||||
(RMContainerUpdatesAcquiredEvent) event;
|
||||
if (acquiredEvent.isIncreasedContainer()) {
|
||||
// If container is increased but not acquired by AM, we will start
|
||||
// If container is increased but not started by AM, we will start
|
||||
// containerAllocationExpirer for this container in this transition.
|
||||
container.containerAllocationExpirer.register(
|
||||
new AllocationExpirationInfo(event.getContainerId(), true));
|
||||
@ -641,7 +639,7 @@ public void transition(RMContainerImpl container, RMContainerEvent event) {
|
||||
container.lastConfirmedResource = rmContainerResource;
|
||||
container.containerAllocationExpirer.unregister(
|
||||
new AllocationExpirationInfo(event.getContainerId()));
|
||||
container.eventHandler.handle(new RMNodeDecreaseContainerEvent(
|
||||
container.eventHandler.handle(new RMNodeUpdateContainerEvent(
|
||||
container.nodeId,
|
||||
Collections.singletonList(container.getContainer())));
|
||||
} else if (Resources.fitsIn(nmContainerResource, rmContainerResource)) {
|
||||
|
@ -144,7 +144,7 @@ public interface RMNode {
|
||||
* applications to clean up for this node.
|
||||
* @param response the {@link NodeHeartbeatResponse} to update
|
||||
*/
|
||||
public void updateNodeHeartbeatResponseForCleanup(NodeHeartbeatResponse response);
|
||||
void updateNodeHeartbeatResponseForCleanup(NodeHeartbeatResponse response);
|
||||
|
||||
public NodeHeartbeatResponse getLastNodeHeartBeatResponse();
|
||||
|
||||
@ -169,9 +169,9 @@ public interface RMNode {
|
||||
public Set<String> getNodeLabels();
|
||||
|
||||
/**
|
||||
* Update containers to be decreased
|
||||
* Update containers to be updated
|
||||
*/
|
||||
public void updateNodeHeartbeatResponseForContainersDecreasing(
|
||||
void updateNodeHeartbeatResponseForUpdatedContainers(
|
||||
NodeHeartbeatResponse response);
|
||||
|
||||
public List<Container> pullNewlyIncreasedContainers();
|
||||
|
@ -42,7 +42,7 @@ public enum RMNodeEventType {
|
||||
// Source: Container
|
||||
CONTAINER_ALLOCATED,
|
||||
CLEANUP_CONTAINER,
|
||||
DECREASE_CONTAINER,
|
||||
UPDATE_CONTAINER,
|
||||
|
||||
// Source: ClientRMService
|
||||
SIGNAL_CONTAINER,
|
||||
|
@ -171,7 +171,7 @@ public class RMNodeImpl implements RMNode, EventHandler<RMNodeEvent> {
|
||||
private final List<ApplicationId> runningApplications =
|
||||
new ArrayList<ApplicationId>();
|
||||
|
||||
private final Map<ContainerId, Container> toBeDecreasedContainers =
|
||||
private final Map<ContainerId, Container> toBeUpdatedContainers =
|
||||
new HashMap<>();
|
||||
|
||||
private final Map<ContainerId, Container> nmReportedIncreasedContainers =
|
||||
@ -228,8 +228,8 @@ RMNodeEventType.RECONNECTED, new ReconnectNodeTransition())
|
||||
.addTransition(NodeState.RUNNING, NodeState.RUNNING,
|
||||
RMNodeEventType.RESOURCE_UPDATE, new UpdateNodeResourceWhenRunningTransition())
|
||||
.addTransition(NodeState.RUNNING, NodeState.RUNNING,
|
||||
RMNodeEventType.DECREASE_CONTAINER,
|
||||
new DecreaseContainersTransition())
|
||||
RMNodeEventType.UPDATE_CONTAINER,
|
||||
new UpdateContainersTransition())
|
||||
.addTransition(NodeState.RUNNING, NodeState.RUNNING,
|
||||
RMNodeEventType.SIGNAL_CONTAINER, new SignalContainerTransition())
|
||||
.addTransition(NodeState.RUNNING, NodeState.SHUTDOWN,
|
||||
@ -614,18 +614,18 @@ public void updateNodeHeartbeatResponseForCleanup(NodeHeartbeatResponse response
|
||||
};
|
||||
|
||||
@VisibleForTesting
|
||||
public Collection<Container> getToBeDecreasedContainers() {
|
||||
return toBeDecreasedContainers.values();
|
||||
public Collection<Container> getToBeUpdatedContainers() {
|
||||
return toBeUpdatedContainers.values();
|
||||
}
|
||||
|
||||
@Override
|
||||
public void updateNodeHeartbeatResponseForContainersDecreasing(
|
||||
public void updateNodeHeartbeatResponseForUpdatedContainers(
|
||||
NodeHeartbeatResponse response) {
|
||||
this.writeLock.lock();
|
||||
|
||||
try {
|
||||
response.addAllContainersToDecrease(toBeDecreasedContainers.values());
|
||||
toBeDecreasedContainers.clear();
|
||||
response.addAllContainersToUpdate(toBeUpdatedContainers.values());
|
||||
toBeUpdatedContainers.clear();
|
||||
} finally {
|
||||
this.writeLock.unlock();
|
||||
}
|
||||
@ -1031,16 +1031,19 @@ public void transition(RMNodeImpl rmNode, RMNodeEvent event) {
|
||||
RMNodeFinishedContainersPulledByAMEvent) event).getContainers());
|
||||
}
|
||||
}
|
||||
|
||||
public static class DecreaseContainersTransition
|
||||
|
||||
/**
|
||||
* Transition to Update a container.
|
||||
*/
|
||||
public static class UpdateContainersTransition
|
||||
implements SingleArcTransition<RMNodeImpl, RMNodeEvent> {
|
||||
|
||||
@Override
|
||||
public void transition(RMNodeImpl rmNode, RMNodeEvent event) {
|
||||
RMNodeDecreaseContainerEvent de = (RMNodeDecreaseContainerEvent) event;
|
||||
RMNodeUpdateContainerEvent de = (RMNodeUpdateContainerEvent) event;
|
||||
|
||||
for (Container c : de.getToBeDecreasedContainers()) {
|
||||
rmNode.toBeDecreasedContainers.put(c.getId(), c);
|
||||
for (Container c : de.getToBeUpdatedContainers()) {
|
||||
rmNode.toBeUpdatedContainers.put(c.getId(), c);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -23,17 +23,22 @@
|
||||
import org.apache.hadoop.yarn.api.records.Container;
|
||||
import org.apache.hadoop.yarn.api.records.NodeId;
|
||||
|
||||
public class RMNodeDecreaseContainerEvent extends RMNodeEvent {
|
||||
final List<Container> toBeDecreasedContainers;
|
||||
/**
|
||||
* This class is used to create update container event
|
||||
* for the containers running on a node.
|
||||
*
|
||||
*/
|
||||
public class RMNodeUpdateContainerEvent extends RMNodeEvent {
|
||||
private List<Container> toBeUpdatedContainers;
|
||||
|
||||
public RMNodeDecreaseContainerEvent(NodeId nodeId,
|
||||
List<Container> toBeDecreasedContainers) {
|
||||
super(nodeId, RMNodeEventType.DECREASE_CONTAINER);
|
||||
|
||||
this.toBeDecreasedContainers = toBeDecreasedContainers;
|
||||
public RMNodeUpdateContainerEvent(NodeId nodeId,
|
||||
List<Container> toBeUpdatedContainers) {
|
||||
super(nodeId, RMNodeEventType.UPDATE_CONTAINER);
|
||||
|
||||
this.toBeUpdatedContainers = toBeUpdatedContainers;
|
||||
}
|
||||
|
||||
public List<Container> getToBeDecreasedContainers() {
|
||||
return toBeDecreasedContainers;
|
||||
|
||||
public List<Container> getToBeUpdatedContainers() {
|
||||
return toBeUpdatedContainers;
|
||||
}
|
||||
}
|
@ -150,6 +150,10 @@ public abstract class AbstractYarnScheduler
|
||||
*/
|
||||
protected final ReentrantReadWriteLock.WriteLock writeLock;
|
||||
|
||||
// If set to true, then ALL container updates will be automatically sent to
|
||||
// the NM in the next heartbeat.
|
||||
private boolean autoUpdateContainers = false;
|
||||
|
||||
/**
|
||||
* Construct the service.
|
||||
*
|
||||
@ -178,6 +182,9 @@ public void serviceInit(Configuration conf) throws Exception {
|
||||
configuredMaximumAllocationWaitTime);
|
||||
maxClusterLevelAppPriority = getMaxPriorityFromConf(conf);
|
||||
createReleaseCache();
|
||||
autoUpdateContainers =
|
||||
conf.getBoolean(YarnConfiguration.RM_AUTO_UPDATE_CONTAINERS,
|
||||
YarnConfiguration.DEFAULT_RM_AUTO_UPDATE_CONTAINERS);
|
||||
super.serviceInit(conf);
|
||||
}
|
||||
|
||||
@ -235,6 +242,10 @@ public boolean accept(SchedulerNode node) {
|
||||
return nodeTracker.getNodes(nodeFilter);
|
||||
}
|
||||
|
||||
public boolean shouldContainersBeAutoUpdated() {
|
||||
return this.autoUpdateContainers;
|
||||
}
|
||||
|
||||
@Override
|
||||
public Resource getClusterResource() {
|
||||
return nodeTracker.getClusterCapacity();
|
||||
|
@ -74,8 +74,7 @@
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerUpdatesAcquiredEvent;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNodeCleanContainerEvent;
|
||||
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode
|
||||
.RMNodeDecreaseContainerEvent;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNodeUpdateContainerEvent;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.SchedulingMode;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.placement.SchedulingPlacementSet;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.policy.SchedulableEntity;
|
||||
@ -663,20 +662,38 @@ private Container updateContainerAndNMToken(RMContainer rmContainer,
|
||||
+ " an updated container " + container.getId(), e);
|
||||
return null;
|
||||
}
|
||||
|
||||
if (updateType == null ||
|
||||
ContainerUpdateType.PROMOTE_EXECUTION_TYPE == updateType ||
|
||||
ContainerUpdateType.DEMOTE_EXECUTION_TYPE == updateType) {
|
||||
|
||||
if (updateType == null) {
|
||||
// This is a newly allocated container
|
||||
rmContainer.handle(new RMContainerEvent(
|
||||
rmContainer.getContainerId(), RMContainerEventType.ACQUIRED));
|
||||
} else {
|
||||
rmContainer.handle(new RMContainerUpdatesAcquiredEvent(
|
||||
rmContainer.getContainerId(),
|
||||
ContainerUpdateType.INCREASE_RESOURCE == updateType));
|
||||
if (ContainerUpdateType.DECREASE_RESOURCE == updateType) {
|
||||
// Resource increase is handled as follows:
|
||||
// If the AM does not use the updated token to increase the container
|
||||
// for a configured period of time, the RM will automatically rollback
|
||||
// the update by performing a container decrease. This rollback (which
|
||||
// essentially is another resource decrease update) is notified to the
|
||||
// NM heartbeat response. If autoUpdate flag is set, then AM does not
|
||||
// need to do anything - same code path as resource decrease.
|
||||
//
|
||||
// Resource Decrease is always automatic: the AM never has to do
|
||||
// anything. It is always via NM heartbeat response.
|
||||
//
|
||||
// ExecutionType updates (both Promotion and Demotion) are either
|
||||
// always automatic (if the flag is set) or the AM has to explicitly
|
||||
// call updateContainer() on the NM. There is no expiry
|
||||
boolean autoUpdate =
|
||||
ContainerUpdateType.DECREASE_RESOURCE == updateType ||
|
||||
((AbstractYarnScheduler)rmContext.getScheduler())
|
||||
.shouldContainersBeAutoUpdated();
|
||||
if (autoUpdate) {
|
||||
this.rmContext.getDispatcher().getEventHandler().handle(
|
||||
new RMNodeDecreaseContainerEvent(rmContainer.getNodeId(),
|
||||
new RMNodeUpdateContainerEvent(rmContainer.getNodeId(),
|
||||
Collections.singletonList(rmContainer.getContainer())));
|
||||
} else {
|
||||
rmContainer.handle(new RMContainerUpdatesAcquiredEvent(
|
||||
rmContainer.getContainerId(),
|
||||
ContainerUpdateType.INCREASE_RESOURCE == updateType));
|
||||
}
|
||||
}
|
||||
return container;
|
||||
|
@ -242,7 +242,7 @@ public Set<String> getNodeLabels() {
|
||||
}
|
||||
|
||||
@Override
|
||||
public void updateNodeHeartbeatResponseForContainersDecreasing(
|
||||
public void updateNodeHeartbeatResponseForUpdatedContainers(
|
||||
NodeHeartbeatResponse response) {
|
||||
|
||||
}
|
||||
|
@ -43,11 +43,13 @@
|
||||
import org.apache.hadoop.yarn.api.records.Resource;
|
||||
import org.apache.hadoop.yarn.api.records.ResourceRequest;
|
||||
import org.apache.hadoop.yarn.api.records.UpdateContainerRequest;
|
||||
import org.apache.hadoop.yarn.api.records.UpdatedContainer;
|
||||
import org.apache.hadoop.yarn.server.api.DistributedSchedulingAMProtocolPB;
|
||||
import org.apache.hadoop.yarn.api.protocolrecords.AllocateRequest;
|
||||
import org.apache.hadoop.yarn.api.protocolrecords.AllocateResponse;
|
||||
import org.apache.hadoop.yarn.server.api.protocolrecords.DistributedSchedulingAllocateRequest;
|
||||
import org.apache.hadoop.yarn.server.api.protocolrecords.DistributedSchedulingAllocateResponse;
|
||||
import org.apache.hadoop.yarn.server.api.protocolrecords.NodeHeartbeatResponse;
|
||||
import org.apache.hadoop.yarn.server.api.protocolrecords.RegisterDistributedSchedulingAMResponse;
|
||||
import org.apache.hadoop.yarn.api.protocolrecords.FinishApplicationMasterRequest;
|
||||
import org.apache.hadoop.yarn.api.protocolrecords.FinishApplicationMasterResponse;
|
||||
@ -122,6 +124,21 @@ public void createAndStartRM() {
|
||||
rm.start();
|
||||
}
|
||||
|
||||
public void createAndStartRMWithAutoUpdateContainer() {
|
||||
CapacitySchedulerConfiguration csConf =
|
||||
new CapacitySchedulerConfiguration();
|
||||
YarnConfiguration conf = new YarnConfiguration(csConf);
|
||||
conf.setBoolean(YarnConfiguration.RM_AUTO_UPDATE_CONTAINERS, true);
|
||||
conf.setClass(YarnConfiguration.RM_SCHEDULER, CapacityScheduler.class,
|
||||
ResourceScheduler.class);
|
||||
conf.setBoolean(
|
||||
YarnConfiguration.OPPORTUNISTIC_CONTAINER_ALLOCATION_ENABLED, true);
|
||||
conf.setInt(
|
||||
YarnConfiguration.NM_CONTAINER_QUEUING_SORTING_NODES_INTERVAL_MS, 100);
|
||||
rm = new MockRM(conf);
|
||||
rm.start();
|
||||
}
|
||||
|
||||
@After
|
||||
public void stopRM() {
|
||||
if (rm != null) {
|
||||
@ -548,6 +565,157 @@ public void testContainerPromoteAfterContainerComplete() throws Exception {
|
||||
verifyMetrics(metrics, 7168, 7, 1024, 1, 1);
|
||||
}
|
||||
|
||||
@Test(timeout = 600000)
|
||||
public void testContainerAutoUpdateContainer() throws Exception {
|
||||
rm.stop();
|
||||
createAndStartRMWithAutoUpdateContainer();
|
||||
MockNM nm1 = new MockNM("h1:1234", 4096, rm.getResourceTrackerService());
|
||||
nm1.registerNode();
|
||||
|
||||
OpportunisticContainerAllocatorAMService amservice =
|
||||
(OpportunisticContainerAllocatorAMService) rm
|
||||
.getApplicationMasterService();
|
||||
RMApp app1 = rm.submitApp(1 * GB, "app", "user", null, "default");
|
||||
ApplicationAttemptId attemptId =
|
||||
app1.getCurrentAppAttempt().getAppAttemptId();
|
||||
MockAM am1 = MockRM.launchAndRegisterAM(app1, rm, nm1);
|
||||
ResourceScheduler scheduler = rm.getResourceScheduler();
|
||||
RMNode rmNode1 = rm.getRMContext().getRMNodes().get(nm1.getNodeId());
|
||||
|
||||
nm1.nodeHeartbeat(true);
|
||||
|
||||
((RMNodeImpl) rmNode1)
|
||||
.setOpportunisticContainersStatus(getOppurtunisticStatus(-1, 100));
|
||||
|
||||
OpportunisticContainerContext ctxt =
|
||||
((CapacityScheduler) scheduler).getApplicationAttempt(attemptId)
|
||||
.getOpportunisticContainerContext();
|
||||
// Send add and update node events to AM Service.
|
||||
amservice.handle(new NodeAddedSchedulerEvent(rmNode1));
|
||||
amservice.handle(new NodeUpdateSchedulerEvent(rmNode1));
|
||||
|
||||
nm1.nodeHeartbeat(true);
|
||||
Thread.sleep(1000);
|
||||
|
||||
AllocateResponse allocateResponse = am1.allocate(Arrays.asList(
|
||||
ResourceRequest.newInstance(Priority.newInstance(1), "*",
|
||||
Resources.createResource(1 * GB), 2, true, null,
|
||||
ExecutionTypeRequest
|
||||
.newInstance(ExecutionType.OPPORTUNISTIC, true))), null);
|
||||
List<Container> allocatedContainers =
|
||||
allocateResponse.getAllocatedContainers();
|
||||
Assert.assertEquals(2, allocatedContainers.size());
|
||||
Container container = allocatedContainers.get(0);
|
||||
// Start Container in NM
|
||||
nm1.nodeHeartbeat(Arrays.asList(ContainerStatus
|
||||
.newInstance(container.getId(), ExecutionType.OPPORTUNISTIC,
|
||||
ContainerState.RUNNING, "", 0)), true);
|
||||
Thread.sleep(200);
|
||||
|
||||
// Verify that container is actually running wrt the RM..
|
||||
RMContainer rmContainer = ((CapacityScheduler) scheduler)
|
||||
.getApplicationAttempt(container.getId().getApplicationAttemptId())
|
||||
.getRMContainer(container.getId());
|
||||
Assert.assertEquals(RMContainerState.RUNNING, rmContainer.getState());
|
||||
|
||||
// Send Promotion req... this should result in update error
|
||||
// Since the container doesn't exist anymore..
|
||||
allocateResponse = am1.sendContainerUpdateRequest(Arrays.asList(
|
||||
UpdateContainerRequest.newInstance(0, container.getId(),
|
||||
ContainerUpdateType.PROMOTE_EXECUTION_TYPE, null,
|
||||
ExecutionType.GUARANTEED)));
|
||||
|
||||
nm1.nodeHeartbeat(Arrays.asList(ContainerStatus
|
||||
.newInstance(container.getId(), ExecutionType.OPPORTUNISTIC,
|
||||
ContainerState.RUNNING, "", 0)), true);
|
||||
Thread.sleep(200);
|
||||
// Get the update response on next allocate
|
||||
allocateResponse = am1.allocate(new ArrayList<>(), new ArrayList<>());
|
||||
// Check the update response from YARNRM
|
||||
Assert.assertEquals(1, allocateResponse.getUpdatedContainers().size());
|
||||
UpdatedContainer uc = allocateResponse.getUpdatedContainers().get(0);
|
||||
Assert.assertEquals(container.getId(), uc.getContainer().getId());
|
||||
Assert.assertEquals(ExecutionType.GUARANTEED,
|
||||
uc.getContainer().getExecutionType());
|
||||
// Check that the container is updated in NM through NM heartbeat response
|
||||
NodeHeartbeatResponse response = nm1.nodeHeartbeat(true);
|
||||
Assert.assertEquals(1, response.getContainersToUpdate().size());
|
||||
Container containersFromNM = response.getContainersToUpdate().get(0);
|
||||
Assert.assertEquals(container.getId(), containersFromNM.getId());
|
||||
Assert.assertEquals(ExecutionType.GUARANTEED,
|
||||
containersFromNM.getExecutionType());
|
||||
|
||||
//Increase resources
|
||||
allocateResponse = am1.sendContainerUpdateRequest(Arrays.asList(
|
||||
UpdateContainerRequest.newInstance(1, container.getId(),
|
||||
ContainerUpdateType.INCREASE_RESOURCE,
|
||||
Resources.createResource(2 * GB, 1), null)));
|
||||
response = nm1.nodeHeartbeat(Arrays.asList(ContainerStatus
|
||||
.newInstance(container.getId(), ExecutionType.GUARANTEED,
|
||||
ContainerState.RUNNING, "", 0)), true);
|
||||
|
||||
Thread.sleep(200);
|
||||
if (allocateResponse.getUpdatedContainers().size() == 0) {
|
||||
allocateResponse = am1.allocate(new ArrayList<>(), new ArrayList<>());
|
||||
}
|
||||
Assert.assertEquals(1, allocateResponse.getUpdatedContainers().size());
|
||||
uc = allocateResponse.getUpdatedContainers().get(0);
|
||||
Assert.assertEquals(container.getId(), uc.getContainer().getId());
|
||||
Assert.assertEquals(Resource.newInstance(2 * GB, 1),
|
||||
uc.getContainer().getResource());
|
||||
|
||||
// Check that the container resources are increased in
|
||||
// NM through NM heartbeat response
|
||||
if (response.getContainersToUpdate().size() == 0) {
|
||||
response = nm1.nodeHeartbeat(true);
|
||||
}
|
||||
Assert.assertEquals(1, response.getContainersToUpdate().size());
|
||||
Assert.assertEquals(Resource.newInstance(2 * GB, 1),
|
||||
response.getContainersToUpdate().get(0).getResource());
|
||||
|
||||
//Decrease resources
|
||||
allocateResponse = am1.sendContainerUpdateRequest(Arrays.asList(
|
||||
UpdateContainerRequest.newInstance(2, container.getId(),
|
||||
ContainerUpdateType.DECREASE_RESOURCE,
|
||||
Resources.createResource(1 * GB, 1), null)));
|
||||
Assert.assertEquals(1, allocateResponse.getUpdatedContainers().size());
|
||||
|
||||
// Check that the container resources are decreased
|
||||
// in NM through NM heartbeat response
|
||||
response = nm1.nodeHeartbeat(true);
|
||||
Assert.assertEquals(1, response.getContainersToUpdate().size());
|
||||
Assert.assertEquals(Resource.newInstance(1 * GB, 1),
|
||||
response.getContainersToUpdate().get(0).getResource());
|
||||
|
||||
nm1.nodeHeartbeat(true);
|
||||
// DEMOTE the container
|
||||
allocateResponse = am1.sendContainerUpdateRequest(Arrays.asList(
|
||||
UpdateContainerRequest.newInstance(3, container.getId(),
|
||||
ContainerUpdateType.DEMOTE_EXECUTION_TYPE, null,
|
||||
ExecutionType.OPPORTUNISTIC)));
|
||||
|
||||
response = nm1.nodeHeartbeat(Arrays.asList(ContainerStatus
|
||||
.newInstance(container.getId(), ExecutionType.GUARANTEED,
|
||||
ContainerState.RUNNING, "", 0)), true);
|
||||
Thread.sleep(200);
|
||||
if (allocateResponse.getUpdatedContainers().size() == 0) {
|
||||
// Get the update response on next allocate
|
||||
allocateResponse = am1.allocate(new ArrayList<>(), new ArrayList<>());
|
||||
}
|
||||
// Check the update response from YARNRM
|
||||
Assert.assertEquals(1, allocateResponse.getUpdatedContainers().size());
|
||||
uc = allocateResponse.getUpdatedContainers().get(0);
|
||||
Assert.assertEquals(ExecutionType.OPPORTUNISTIC,
|
||||
uc.getContainer().getExecutionType());
|
||||
// Check that the container is updated in NM through NM heartbeat response
|
||||
if (response.getContainersToUpdate().size() == 0) {
|
||||
response = nm1.nodeHeartbeat(true);
|
||||
}
|
||||
Assert.assertEquals(1, response.getContainersToUpdate().size());
|
||||
Assert.assertEquals(ExecutionType.OPPORTUNISTIC,
|
||||
response.getContainersToUpdate().get(0).getExecutionType());
|
||||
}
|
||||
|
||||
private void verifyMetrics(QueueMetrics metrics, long availableMB,
|
||||
int availableVirtualCores, long allocatedMB,
|
||||
int allocatedVirtualCores, int allocatedContainers) {
|
||||
|
@ -51,8 +51,7 @@
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNode;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNodeImpl;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.ResourceScheduler;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler
|
||||
.SchedulerApplicationAttempt;
|
||||
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerNode;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.common.fica.FiCaSchedulerApp;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.common.fica
|
||||
@ -60,11 +59,9 @@
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.event.AppAttemptRemovedSchedulerEvent;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.event.NodeUpdateSchedulerEvent;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.placement.PlacementSet;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.placement.SimplePlacementSet;
|
||||
import org.apache.hadoop.yarn.util.resource.Resources;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
import org.junit.Ignore;
|
||||
import org.junit.Test;
|
||||
|
||||
public class TestContainerResizing {
|
||||
@ -205,7 +202,7 @@ public RMNodeLabelsManager createNodeLabelManager() {
|
||||
RMNodeImpl rmNode =
|
||||
(RMNodeImpl) rm1.getRMContext().getRMNodes().get(nm1.getNodeId());
|
||||
Collection<Container> decreasedContainers =
|
||||
rmNode.getToBeDecreasedContainers();
|
||||
rmNode.getToBeUpdatedContainers();
|
||||
boolean rmNodeReceivedDecreaseContainer = false;
|
||||
for (Container c : decreasedContainers) {
|
||||
if (c.getId().equals(containerId1)
|
||||
|
@ -319,7 +319,7 @@ public void testConsecutiveContainerIncreaseAllocationExpiration()
|
||||
verifyAvailableResourceOfSchedulerNode(rm1, nm1.getNodeId(), 16 * GB);
|
||||
// Verify NM receives the decrease message (3G)
|
||||
List<Container> containersToDecrease =
|
||||
nm1.nodeHeartbeat(true).getContainersToDecrease();
|
||||
nm1.nodeHeartbeat(true).getContainersToUpdate();
|
||||
Assert.assertEquals(1, containersToDecrease.size());
|
||||
Assert.assertEquals(
|
||||
3 * GB, containersToDecrease.get(0).getResource().getMemorySize());
|
||||
@ -435,7 +435,7 @@ public void testDecreaseAfterIncreaseWithAllocationExpiration()
|
||||
.getAllocatedResource().getMemorySize());
|
||||
// Verify NM receives 2 decrease message
|
||||
List<Container> containersToDecrease =
|
||||
nm1.nodeHeartbeat(true).getContainersToDecrease();
|
||||
nm1.nodeHeartbeat(true).getContainersToUpdate();
|
||||
Assert.assertEquals(2, containersToDecrease.size());
|
||||
// Sort the list to make sure containerId3 is the first
|
||||
Collections.sort(containersToDecrease);
|
||||
|
Loading…
Reference in New Issue
Block a user