HDDS-461. Container remains in CLOSING state in SCM forever. Contributed by Shashikant Banerjee.

This commit is contained in:
Nanda kumar 2018-09-19 21:43:44 +05:30
parent b3c5221f30
commit 61a4b07bda
8 changed files with 175 additions and 161 deletions

View File

@ -16,6 +16,7 @@
*/ */
package org.apache.hadoop.ozone.container.common.statemachine; package org.apache.hadoop.ozone.container.common.statemachine;
import com.google.common.base.Preconditions;
import com.google.protobuf.GeneratedMessage; import com.google.protobuf.GeneratedMessage;
import java.util.Map; import java.util.Map;
import java.util.concurrent.ConcurrentHashMap; import java.util.concurrent.ConcurrentHashMap;
@ -43,6 +44,7 @@
import java.util.LinkedList; import java.util.LinkedList;
import java.util.List; import java.util.List;
import java.util.Queue; import java.util.Queue;
import java.util.ArrayList;
import java.util.concurrent.ExecutionException; import java.util.concurrent.ExecutionException;
import java.util.concurrent.ExecutorService; import java.util.concurrent.ExecutorService;
import java.util.concurrent.TimeUnit; import java.util.concurrent.TimeUnit;
@ -50,7 +52,6 @@
import java.util.concurrent.atomic.AtomicLong; import java.util.concurrent.atomic.AtomicLong;
import java.util.concurrent.locks.Lock; import java.util.concurrent.locks.Lock;
import java.util.concurrent.locks.ReentrantLock; import java.util.concurrent.locks.ReentrantLock;
import java.util.stream.Collectors;
import static org.apache.hadoop.ozone.OzoneConsts.INVALID_PORT; import static org.apache.hadoop.ozone.OzoneConsts.INVALID_PORT;
@ -206,9 +207,18 @@ public List<GeneratedMessage> getAllAvailableReports() {
* @return List<reports> * @return List<reports>
*/ */
public List<GeneratedMessage> getReports(int maxLimit) { public List<GeneratedMessage> getReports(int maxLimit) {
List<GeneratedMessage> reportList = new ArrayList<>();
synchronized (reports) { synchronized (reports) {
return reports.parallelStream().limit(maxLimit) if (!reports.isEmpty()) {
.collect(Collectors.toList()); int size = reports.size();
int limit = size > maxLimit ? maxLimit : size;
for (int count = 0; count < limit; count++) {
GeneratedMessage report = reports.poll();
Preconditions.checkNotNull(report);
reportList.add(report);
}
}
return reportList;
} }
} }
@ -254,9 +264,20 @@ public List<ContainerAction> getAllPendingContainerActions() {
* @return List<ContainerAction> * @return List<ContainerAction>
*/ */
public List<ContainerAction> getPendingContainerAction(int maxLimit) { public List<ContainerAction> getPendingContainerAction(int maxLimit) {
List<ContainerAction> containerActionList = new ArrayList<>();
synchronized (containerActions) { synchronized (containerActions) {
return containerActions.parallelStream().limit(maxLimit) if (!containerActions.isEmpty()) {
.collect(Collectors.toList()); int size = containerActions.size();
int limit = size > maxLimit ? maxLimit : size;
for (int count = 0; count < limit; count++) {
// we need to remove the action from the containerAction queue
// as well
ContainerAction action = containerActions.poll();
Preconditions.checkNotNull(action);
containerActionList.add(action);
}
}
return containerActionList;
} }
} }
@ -295,9 +316,16 @@ public void addPipelineActionIfAbsent(PipelineAction pipelineAction) {
* @return List<ContainerAction> * @return List<ContainerAction>
*/ */
public List<PipelineAction> getPendingPipelineAction(int maxLimit) { public List<PipelineAction> getPendingPipelineAction(int maxLimit) {
List<PipelineAction> pipelineActionList = new ArrayList<>();
synchronized (pipelineActions) { synchronized (pipelineActions) {
return pipelineActions.parallelStream().limit(maxLimit) if (!pipelineActions.isEmpty()) {
.collect(Collectors.toList()); int size = pipelineActions.size();
int limit = size > maxLimit ? maxLimit : size;
for (int count = 0; count < limit; count++) {
pipelineActionList.add(pipelineActions.poll());
}
}
return pipelineActionList;
} }
} }

View File

@ -28,6 +28,7 @@
import org.apache.hadoop.ozone.container.ozoneimpl.OzoneContainer; import org.apache.hadoop.ozone.container.ozoneimpl.OzoneContainer;
import org.apache.hadoop.ozone.protocol.commands.SCMCommand; import org.apache.hadoop.ozone.protocol.commands.SCMCommand;
import org.apache.hadoop.util.Time; import org.apache.hadoop.util.Time;
import org.apache.ratis.protocol.NotLeaderException;
import org.slf4j.Logger; import org.slf4j.Logger;
import org.slf4j.LoggerFactory; import org.slf4j.LoggerFactory;
@ -89,9 +90,21 @@ public void handle(SCMCommand command, OzoneContainer container,
// submit the close container request for the XceiverServer to handle // submit the close container request for the XceiverServer to handle
container.submitContainerRequest( container.submitContainerRequest(
request.build(), replicationType, pipelineID); request.build(), replicationType, pipelineID);
cmdExecuted = true;
} catch (Exception e) { } catch (Exception e) {
LOG.error("Can't close container " + containerID, e); if (e instanceof NotLeaderException) {
// If the particular datanode is not the Ratis leader, the close
// container command will not be executed by the follower but will be
// executed by Ratis stateMachine transactions via leader to follower.
// There can also be case where the datanode is in candidate state.
// In these situations, NotLeaderException is thrown. Remove the status
// from cmdStatus Map here so that it will be retried only by SCM if the
// leader could not not close the container after a certain time.
context.removeCommandStatus(containerID);
LOG.info(e.getLocalizedMessage());
} else {
LOG.error("Can't close container " + containerID, e);
cmdExecuted = false;
}
} finally { } finally {
updateCommandStatus(context, command, cmdExecuted, LOG); updateCommandStatus(context, command, cmdExecuted, LOG);
long endTime = Time.monotonicNow(); long endTime = Time.monotonicNow();

View File

@ -21,12 +21,15 @@
import com.google.common.base.Preconditions; import com.google.common.base.Preconditions;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hdds.protocol.DatanodeDetails; import org.apache.hadoop.hdds.protocol.DatanodeDetails;
import org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos;
import org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos import org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos
.ContainerCommandRequestProto; .ContainerCommandRequestProto;
import org.apache.hadoop.hdds.protocol.proto.HddsProtos; import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
import org.apache.hadoop.hdds.protocol.proto import org.apache.hadoop.hdds.protocol.proto
.StorageContainerDatanodeProtocolProtos.PipelineReport; .StorageContainerDatanodeProtocolProtos.PipelineReport;
import org.apache.hadoop.hdds.scm.container.common.helpers.PipelineID; import org.apache.hadoop.hdds.scm.container.common.helpers.PipelineID;
import org.apache.hadoop.hdds.scm.container.common.helpers.
StorageContainerException;
import org.apache.hadoop.ozone.OzoneConfigKeys; import org.apache.hadoop.ozone.OzoneConfigKeys;
import org.apache.hadoop.ozone.container.common.interfaces.ContainerDispatcher; import org.apache.hadoop.ozone.container.common.interfaces.ContainerDispatcher;
@ -128,8 +131,13 @@ public void stop() {
@Override @Override
public void submitRequest(ContainerCommandRequestProto request, public void submitRequest(ContainerCommandRequestProto request,
HddsProtos.PipelineID pipelineID) { HddsProtos.PipelineID pipelineID) throws IOException {
storageContainer.dispatch(request); ContainerProtos.ContainerCommandResponseProto response =
storageContainer.dispatch(request);
if (response.getResult() != ContainerProtos.Result.SUCCESS) {
throw new StorageContainerException(response.getMessage(),
response.getResult());
}
} }
@Override @Override

View File

@ -74,7 +74,6 @@
import java.util.Objects; import java.util.Objects;
import java.util.UUID; import java.util.UUID;
import java.util.concurrent.ArrayBlockingQueue; import java.util.concurrent.ArrayBlockingQueue;
import java.util.concurrent.CompletableFuture;
import java.util.concurrent.ThreadPoolExecutor; import java.util.concurrent.ThreadPoolExecutor;
import java.util.concurrent.TimeUnit; import java.util.concurrent.TimeUnit;
import java.util.concurrent.atomic.AtomicLong; import java.util.concurrent.atomic.AtomicLong;
@ -319,39 +318,35 @@ public RaftServer getServer() {
return server; return server;
} }
private void processReply(RaftClientReply reply) { private void processReply(RaftClientReply reply) throws IOException {
// NotLeader exception is thrown only when the raft server to which the // NotLeader exception is thrown only when the raft server to which the
// request is submitted is not the leader. The request will be rejected // request is submitted is not the leader. The request will be rejected
// and will eventually be executed once the request comnes via the leader // and will eventually be executed once the request comes via the leader
// node. // node.
NotLeaderException notLeaderException = reply.getNotLeaderException(); NotLeaderException notLeaderException = reply.getNotLeaderException();
if (notLeaderException != null) { if (notLeaderException != null) {
LOG.info(reply.getNotLeaderException().getLocalizedMessage()); throw notLeaderException;
} }
StateMachineException stateMachineException = StateMachineException stateMachineException =
reply.getStateMachineException(); reply.getStateMachineException();
if (stateMachineException != null) { if (stateMachineException != null) {
// In case the request could not be completed, StateMachine Exception throw stateMachineException;
// will be thrown. For now, Just log the message.
// If the container could not be closed, SCM will come to know
// via containerReports. CloseContainer should be re tried via SCM.
LOG.error(stateMachineException.getLocalizedMessage());
} }
} }
@Override @Override
public void submitRequest( public void submitRequest(ContainerCommandRequestProto request,
ContainerCommandRequestProto request, HddsProtos.PipelineID pipelineID) HddsProtos.PipelineID pipelineID) throws IOException {
throws IOException { RaftClientReply reply;
// ReplicationLevel.MAJORITY ensures the transactions corresponding to
// the request here are applied on all the raft servers.
RaftClientRequest raftClientRequest = RaftClientRequest raftClientRequest =
createRaftClientRequest(request, pipelineID, createRaftClientRequest(request, pipelineID,
RaftClientRequest.writeRequestType(replicationLevel)); RaftClientRequest.writeRequestType(replicationLevel));
CompletableFuture<RaftClientReply> reply = try {
server.submitClientRequestAsync(raftClientRequest); reply = server.submitClientRequestAsync(raftClientRequest).get();
reply.thenAccept(this::processReply); } catch (Exception e) {
throw new IOException(e.getMessage(), e);
}
processReply(reply);
} }
private RaftClientRequest createRaftClientRequest( private RaftClientRequest createRaftClientRequest(

View File

@ -23,7 +23,8 @@
import org.apache.hadoop.hdds.conf.OzoneConfiguration; import org.apache.hadoop.hdds.conf.OzoneConfiguration;
import org.apache.hadoop.hdds.protocol.DatanodeDetails; import org.apache.hadoop.hdds.protocol.DatanodeDetails;
import org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos; import org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos;
import org.apache.hadoop.hdds.protocol.proto.HddsProtos; import org.apache.hadoop.hdds.protocol.proto.HddsProtos.PipelineID;
import org.apache.hadoop.hdds.protocol.proto.HddsProtos.ReplicationType;
import org.apache.hadoop.hdds.protocol.proto.StorageContainerDatanodeProtocolProtos; import org.apache.hadoop.hdds.protocol.proto.StorageContainerDatanodeProtocolProtos;
import org.apache.hadoop.hdds.protocol.proto import org.apache.hadoop.hdds.protocol.proto
.StorageContainerDatanodeProtocolProtos.PipelineReportsProto; .StorageContainerDatanodeProtocolProtos.PipelineReportsProto;
@ -46,12 +47,14 @@
import java.io.*; import java.io.*;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.HashMap;
import java.util.Iterator; import java.util.Iterator;
import java.util.Map;
import static org.apache.hadoop.ozone.OzoneConsts.INVALID_PORT; import static org.apache.hadoop.ozone.OzoneConsts.INVALID_PORT;
/** /**
* Ozone main class sets up the network server and initializes the container * Ozone main class sets up the network servers and initializes the container
* layer. * layer.
*/ */
public class OzoneContainer { public class OzoneContainer {
@ -64,7 +67,7 @@ public class OzoneContainer {
private final OzoneConfiguration config; private final OzoneConfiguration config;
private final VolumeSet volumeSet; private final VolumeSet volumeSet;
private final ContainerSet containerSet; private final ContainerSet containerSet;
private final XceiverServerSpi[] server; private final Map<ReplicationType, XceiverServerSpi> servers;
/** /**
* Construct OzoneContainer object. * Construct OzoneContainer object.
@ -82,14 +85,13 @@ public OzoneContainer(DatanodeDetails datanodeDetails, OzoneConfiguration
buildContainerSet(); buildContainerSet();
hddsDispatcher = new HddsDispatcher(config, containerSet, volumeSet, hddsDispatcher = new HddsDispatcher(config, containerSet, volumeSet,
context); context);
server = new XceiverServerSpi[]{ servers = new HashMap<>();
new XceiverServerGrpc(datanodeDetails, this.config, this servers.put(ReplicationType.STAND_ALONE,
.hddsDispatcher, createReplicationService()), new XceiverServerGrpc(datanodeDetails, config, hddsDispatcher,
XceiverServerRatis.newXceiverServerRatis(datanodeDetails, this createReplicationService()));
.config, hddsDispatcher, context) servers.put(ReplicationType.RATIS, XceiverServerRatis
}; .newXceiverServerRatis(datanodeDetails, config, hddsDispatcher,
context));
} }
private GrpcReplicationService createReplicationService() { private GrpcReplicationService createReplicationService() {
@ -133,7 +135,7 @@ public void buildContainerSet() {
*/ */
public void start() throws IOException { public void start() throws IOException {
LOG.info("Attempting to start container services."); LOG.info("Attempting to start container services.");
for (XceiverServerSpi serverinstance : server) { for (XceiverServerSpi serverinstance : servers.values()) {
serverinstance.start(); serverinstance.start();
} }
hddsDispatcher.init(); hddsDispatcher.init();
@ -145,7 +147,7 @@ public void start() throws IOException {
public void stop() { public void stop() {
//TODO: at end of container IO integration work. //TODO: at end of container IO integration work.
LOG.info("Attempting to stop container services."); LOG.info("Attempting to stop container services.");
for(XceiverServerSpi serverinstance: server) { for(XceiverServerSpi serverinstance: servers.values()) {
serverinstance.stop(); serverinstance.stop();
} }
hddsDispatcher.shutdown(); hddsDispatcher.shutdown();
@ -169,7 +171,7 @@ public ContainerSet getContainerSet() {
public PipelineReportsProto getPipelineReport() { public PipelineReportsProto getPipelineReport() {
PipelineReportsProto.Builder pipelineReportsProto = PipelineReportsProto.Builder pipelineReportsProto =
PipelineReportsProto.newBuilder(); PipelineReportsProto.newBuilder();
for (XceiverServerSpi serverInstance : server) { for (XceiverServerSpi serverInstance : servers.values()) {
pipelineReportsProto pipelineReportsProto
.addAllPipelineReport(serverInstance.getPipelineReport()); .addAllPipelineReport(serverInstance.getPipelineReport());
} }
@ -181,82 +183,38 @@ public PipelineReportsProto getPipelineReport() {
* @param request * @param request
* @param replicationType * @param replicationType
* @param pipelineID * @param pipelineID
* @throws IOException
*/ */
public void submitContainerRequest( public void submitContainerRequest(
ContainerProtos.ContainerCommandRequestProto request, ContainerProtos.ContainerCommandRequestProto request,
HddsProtos.ReplicationType replicationType, ReplicationType replicationType,
HddsProtos.PipelineID pipelineID) throws IOException { PipelineID pipelineID) throws IOException {
XceiverServerSpi serverInstance; if (containerSet.getContainer(request.getContainerID())
long containerId = getContainerIdForCmd(request); .getContainerData().isClosed()) {
if (replicationType == HddsProtos.ReplicationType.RATIS) { LOG.debug("Container {} is already closed", request.getContainerID());
serverInstance = getRatisSerer(); // It might happen that the where the first attempt of closing the
Preconditions.checkNotNull(serverInstance); // container failed with NOT_LEADER_EXCEPTION. In such cases, SCM will
serverInstance.submitRequest(request, pipelineID); // retry to check the container got really closed via Ratis.
LOG.info("submitting {} request over RATIS server for container {}", // In such cases of the retry attempt, if the container is already closed
request.getCmdType(), containerId); // via Ratis, we should just return.
} else {
serverInstance = getStandaAloneSerer();
Preconditions.checkNotNull(serverInstance);
getStandaAloneSerer().submitRequest(request, pipelineID);
LOG.info(
"submitting {} request over STAND_ALONE server for container {}",
request.getCmdType(), containerId);
} }
LOG.info("submitting {} request over {} server for container {}",
request.getCmdType(), replicationType, request.getContainerID());
Preconditions.checkState(servers.containsKey(replicationType));
servers.get(replicationType).submitRequest(request, pipelineID);
} }
private long getContainerIdForCmd( private int getPortByType(ReplicationType replicationType) {
ContainerProtos.ContainerCommandRequestProto request) return servers.containsKey(replicationType) ?
throws IllegalArgumentException { servers.get(replicationType).getIPCPort() : INVALID_PORT;
ContainerProtos.Type type = request.getCmdType();
switch (type) {
case CloseContainer:
return request.getContainerID();
// Right now, we handle only closeContainer via queuing it over the
// over the XceiVerServer. For all other commands we throw Illegal
// argument exception here. Will need to extend the switch cases
// in case we want add another commands here.
default:
throw new IllegalArgumentException("Cmd " + request.getCmdType()
+ " not supported over HearBeat Response");
}
}
private XceiverServerSpi getRatisSerer() {
for (XceiverServerSpi serverInstance : server) {
if (serverInstance instanceof XceiverServerRatis) {
return serverInstance;
}
}
return null;
}
private XceiverServerSpi getStandaAloneSerer() {
for (XceiverServerSpi serverInstance : server) {
if (!(serverInstance instanceof XceiverServerRatis)) {
return serverInstance;
}
}
return null;
}
private int getPortbyType(HddsProtos.ReplicationType replicationType) {
for (XceiverServerSpi serverinstance : server) {
if (serverinstance.getServerType() == replicationType) {
return serverinstance.getIPCPort();
}
}
return INVALID_PORT;
} }
/** /**
* Returns the container server IPC port. * Returns the container servers IPC port.
* *
* @return Container server IPC port. * @return Container servers IPC port.
*/ */
public int getContainerServerPort() { public int getContainerServerPort() {
return getPortbyType(HddsProtos.ReplicationType.STAND_ALONE); return getPortByType(ReplicationType.STAND_ALONE);
} }
/** /**
@ -265,7 +223,7 @@ public int getContainerServerPort() {
* @return Ratis port. * @return Ratis port.
*/ */
public int getRatisContainerServerPort() { public int getRatisContainerServerPort() {
return getPortbyType(HddsProtos.ReplicationType.RATIS); return getPortByType(ReplicationType.RATIS);
} }
/** /**

View File

@ -30,25 +30,13 @@
public class CloseContainerCommand public class CloseContainerCommand
extends SCMCommand<CloseContainerCommandProto> { extends SCMCommand<CloseContainerCommandProto> {
private long containerID;
private HddsProtos.ReplicationType replicationType; private HddsProtos.ReplicationType replicationType;
private PipelineID pipelineID; private PipelineID pipelineID;
public CloseContainerCommand(long containerID, public CloseContainerCommand(long containerID,
HddsProtos.ReplicationType replicationType, HddsProtos.ReplicationType replicationType,
PipelineID pipelineID) { PipelineID pipelineID) {
super(); super(containerID);
this.containerID = containerID;
this.replicationType = replicationType;
this.pipelineID = pipelineID;
}
// Should be called only for protobuf conversion
private CloseContainerCommand(long containerID,
HddsProtos.ReplicationType replicationType,
PipelineID pipelineID, long id) {
super(id);
this.containerID = containerID;
this.replicationType = replicationType; this.replicationType = replicationType;
this.pipelineID = pipelineID; this.pipelineID = pipelineID;
} }
@ -75,7 +63,7 @@ public byte[] getProtoBufMessage() {
public CloseContainerCommandProto getProto() { public CloseContainerCommandProto getProto() {
return CloseContainerCommandProto.newBuilder() return CloseContainerCommandProto.newBuilder()
.setContainerID(containerID) .setContainerID(getId())
.setCmdId(getId()) .setCmdId(getId())
.setReplicationType(replicationType) .setReplicationType(replicationType)
.setPipelineID(pipelineID.getProtobuf()) .setPipelineID(pipelineID.getProtobuf())
@ -85,13 +73,12 @@ public CloseContainerCommandProto getProto() {
public static CloseContainerCommand getFromProtobuf( public static CloseContainerCommand getFromProtobuf(
CloseContainerCommandProto closeContainerProto) { CloseContainerCommandProto closeContainerProto) {
Preconditions.checkNotNull(closeContainerProto); Preconditions.checkNotNull(closeContainerProto);
return new CloseContainerCommand(closeContainerProto.getContainerID(), return new CloseContainerCommand(closeContainerProto.getCmdId(),
closeContainerProto.getReplicationType(), closeContainerProto.getReplicationType(),
PipelineID.getFromProtobuf(closeContainerProto.getPipelineID()), PipelineID.getFromProtobuf(closeContainerProto.getPipelineID()));
closeContainerProto.getCmdId());
} }
public long getContainerID() { public long getContainerID() {
return containerID; return getId();
} }
} }

View File

@ -17,10 +17,10 @@
package org.apache.hadoop.hdds.scm.container; package org.apache.hadoop.hdds.scm.container;
import java.io.IOException; import java.io.IOException;
import org.apache.hadoop.hdds.protocol.DatanodeDetails;
import org.apache.hadoop.hdds.protocol.proto.HddsProtos; import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
import org.apache.hadoop.hdds.scm.container.common.helpers.ContainerInfo; import org.apache.hadoop.hdds.scm.container.common.helpers.ContainerInfo;
import org.apache.hadoop.hdds.scm.container.common.helpers.ContainerWithPipeline; import org.apache.hadoop.hdds.scm.container.common.helpers.ContainerWithPipeline;
import org.apache.hadoop.hdds.scm.container.common.helpers.Pipeline;
import org.apache.hadoop.hdds.server.events.EventHandler; import org.apache.hadoop.hdds.server.events.EventHandler;
import org.apache.hadoop.hdds.server.events.EventPublisher; import org.apache.hadoop.hdds.server.events.EventPublisher;
import org.apache.hadoop.hdds.server.events.IdentifiableEventPayload; import org.apache.hadoop.hdds.server.events.IdentifiableEventPayload;
@ -57,7 +57,7 @@ public void onMessage(ContainerID containerID, EventPublisher publisher) {
LOG.info("Close container Event triggered for container : {}", LOG.info("Close container Event triggered for container : {}",
containerID.getId()); containerID.getId());
ContainerWithPipeline containerWithPipeline = null; ContainerWithPipeline containerWithPipeline;
ContainerInfo info; ContainerInfo info;
try { try {
containerWithPipeline = containerWithPipeline =
@ -74,42 +74,66 @@ public void onMessage(ContainerID containerID, EventPublisher publisher) {
return; return;
} }
if (info.getState() == HddsProtos.LifeCycleState.OPEN) { HddsProtos.LifeCycleState state = info.getState();
for (DatanodeDetails datanode : try {
containerWithPipeline.getPipeline().getMachines()) { switch (state) {
CommandForDatanode closeContainerCommand = new CommandForDatanode<>( case ALLOCATED:
datanode.getUuid(), // We cannot close a container in ALLOCATED state, moving the
new CloseContainerCommand(containerID.getId(), // container to CREATING state, this should eventually
info.getReplicationType(), info.getPipelineID())); // timeout and the container will be moved to DELETING state.
publisher.fireEvent(DATANODE_COMMAND, closeContainerCommand); LOG.debug("Closing container {} in {} state", containerID, state);
publisher.fireEvent(CLOSE_CONTAINER_RETRYABLE_REQ, new
CloseContainerRetryableReq(containerID));
}
try {
// Finalize event will make sure the state of the container transitions
// from OPEN to CLOSING in containerStateManager.
containerManager.updateContainerState(containerID.getId(),
HddsProtos.LifeCycleEvent.FINALIZE);
} catch (IOException ex) {
LOG.error("Failed to update the container state to FINALIZE for"
+ "container : {}" + containerID, ex);
}
} else if (info.getState() == HddsProtos.LifeCycleState.ALLOCATED) {
try {
// Create event will make sure the state of the container transitions
// from OPEN to CREATING in containerStateManager, this will move
// the container out of active allocation path.
containerManager.updateContainerState(containerID.getId(), containerManager.updateContainerState(containerID.getId(),
HddsProtos.LifeCycleEvent.CREATE); HddsProtos.LifeCycleEvent.CREATE);
} catch (IOException ex) { break;
LOG.error("Failed to update the container state to CREATE for" case CREATING:
+ "container:{}" + containerID, ex); // We cannot close a container in CREATING state, it will eventually
// timeout and moved to DELETING state.
LOG.debug("Closing container {} in {} state", containerID, state);
break;
case OPEN:
containerManager.updateContainerState(containerID.getId(),
HddsProtos.LifeCycleEvent.FINALIZE);
fireCloseContainerEvents(containerWithPipeline, info, publisher);
break;
case CLOSING:
fireCloseContainerEvents(containerWithPipeline, info, publisher);
break;
case CLOSED:
case DELETING:
case DELETED:
LOG.info(
"container with id : {} is in {} state and need not be closed.",
containerID.getId(), info.getState());
break;
default:
throw new IOException(
"Invalid container state for container " + containerID);
} }
} else { } catch (IOException ex) {
LOG.info("container with id : {} is in {} state and need not be closed.", LOG.error("Failed to update the container state for" + "container : {}"
containerID.getId(), info.getState()); + containerID, ex);
} }
}
private void fireCloseContainerEvents(
ContainerWithPipeline containerWithPipeline, ContainerInfo info,
EventPublisher publisher) {
ContainerID containerID = info.containerID();
// fire events.
CloseContainerCommand closeContainerCommand =
new CloseContainerCommand(containerID.getId(),
info.getReplicationType(), info.getPipelineID());
Pipeline pipeline = containerWithPipeline.getPipeline();
pipeline.getMachines().stream().map(
datanode -> new CommandForDatanode<>(datanode.getUuid(),
closeContainerCommand)).forEach((command) -> {
publisher.fireEvent(DATANODE_COMMAND, command);
});
publisher.fireEvent(CLOSE_CONTAINER_RETRYABLE_REQ,
new CloseContainerRetryableReq(containerID));
LOG.trace("Issuing {} on Pipeline {} for container", closeContainerCommand,
pipeline, containerID);
} }
/** /**

View File

@ -132,7 +132,6 @@ public void testIfCloseContainerCommandHandlerIsInvoked() throws Exception {
// Make sure the closeContainerCommandHandler is Invoked // Make sure the closeContainerCommandHandler is Invoked
Assert.assertTrue( Assert.assertTrue(
closeContainerHandler.getInvocationCount() > lastInvocationCount); closeContainerHandler.getInvocationCount() > lastInvocationCount);
} }
@Test @Test
@ -190,6 +189,7 @@ public void testCloseContainerViaStandAlone()
Assert.assertFalse((logCapturer.getOutput().contains( Assert.assertFalse((logCapturer.getOutput().contains(
"submitting CloseContainer request over RATIS server for container " "submitting CloseContainer request over RATIS server for container "
+ containerID))); + containerID)));
logCapturer.stopCapturing();
} }
@Test @Test
@ -239,13 +239,14 @@ public void testCloseContainerViaRatis() throws IOException,
Assert.assertTrue(isContainerClosed(cluster, Assert.assertTrue(isContainerClosed(cluster,
containerID, datanodeDetails)); containerID, datanodeDetails));
} }
// Make sure it was really closed via Ratis not STAND_ALONE server
Assert.assertFalse(logCapturer.getOutput().contains( Assert.assertFalse(logCapturer.getOutput().contains(
"submitting CloseContainer request over STAND_ALONE " "submitting CloseContainer request over STAND_ALONE "
+ "server for container " + containerID)); + "server for container " + containerID));
// Make sure it was really closed via StandAlone not Ratis server
Assert.assertTrue((logCapturer.getOutput().contains( Assert.assertTrue((logCapturer.getOutput().contains(
"submitting CloseContainer request over RATIS server for container " "submitting CloseContainer request over RATIS server for container "
+ containerID))); + containerID)));
logCapturer.stopCapturing();
} }
private Boolean isContainerClosed(MiniOzoneCluster cluster, long containerID, private Boolean isContainerClosed(MiniOzoneCluster cluster, long containerID,