HDFS-11699. Ozone:SCM: Add support for close containers in SCM. Contributed by Anu Engineer.
This commit is contained in:
parent
b4a3cf1476
commit
1094af072c
@ -27,9 +27,11 @@
|
|||||||
import org.apache.hadoop.ozone.protocol.proto.OzoneProtos;
|
import org.apache.hadoop.ozone.protocol.proto.OzoneProtos;
|
||||||
import org.apache.hadoop.ozone.protocol.proto.OzoneProtos.ReplicationFactor;
|
import org.apache.hadoop.ozone.protocol.proto.OzoneProtos.ReplicationFactor;
|
||||||
import org.apache.hadoop.ozone.protocol.proto.OzoneProtos.ReplicationType;
|
import org.apache.hadoop.ozone.protocol.proto.OzoneProtos.ReplicationType;
|
||||||
import org.apache.hadoop.ozone.protocol.proto.StorageContainerDatanodeProtocolProtos;
|
import org.apache.hadoop.ozone.protocol.proto
|
||||||
|
.StorageContainerDatanodeProtocolProtos;
|
||||||
import org.apache.hadoop.ozone.protocol.proto
|
import org.apache.hadoop.ozone.protocol.proto
|
||||||
.StorageContainerDatanodeProtocolProtos.ContainerReportsRequestProto;
|
.StorageContainerDatanodeProtocolProtos.ContainerReportsRequestProto;
|
||||||
|
import org.apache.hadoop.ozone.scm.container.closer.ContainerCloser;
|
||||||
import org.apache.hadoop.ozone.scm.container.replication.ContainerSupervisor;
|
import org.apache.hadoop.ozone.scm.container.replication.ContainerSupervisor;
|
||||||
import org.apache.hadoop.ozone.scm.exceptions.SCMException;
|
import org.apache.hadoop.ozone.scm.exceptions.SCMException;
|
||||||
import org.apache.hadoop.ozone.scm.node.NodeManager;
|
import org.apache.hadoop.ozone.scm.node.NodeManager;
|
||||||
@ -57,6 +59,9 @@
|
|||||||
import static org.apache.hadoop.ozone.OzoneConsts.SCM_CONTAINER_DB;
|
import static org.apache.hadoop.ozone.OzoneConsts.SCM_CONTAINER_DB;
|
||||||
import static org.apache.hadoop.ozone.scm.exceptions.SCMException.ResultCodes
|
import static org.apache.hadoop.ozone.scm.exceptions.SCMException.ResultCodes
|
||||||
.FAILED_TO_CHANGE_CONTAINER_STATE;
|
.FAILED_TO_CHANGE_CONTAINER_STATE;
|
||||||
|
import static org.apache.hadoop.scm.ScmConfigKeys
|
||||||
|
.OZONE_SCM_CONTAINER_SIZE_DEFAULT;
|
||||||
|
import static org.apache.hadoop.scm.ScmConfigKeys.OZONE_SCM_CONTAINER_SIZE_GB;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Mapping class contains the mapping from a name to a pipeline mapping. This
|
* Mapping class contains the mapping from a name to a pipeline mapping. This
|
||||||
@ -77,6 +82,8 @@ public class ContainerMapping implements Mapping {
|
|||||||
private final LeaseManager<ContainerInfo> containerLeaseManager;
|
private final LeaseManager<ContainerInfo> containerLeaseManager;
|
||||||
private final ContainerSupervisor containerSupervisor;
|
private final ContainerSupervisor containerSupervisor;
|
||||||
private final float containerCloseThreshold;
|
private final float containerCloseThreshold;
|
||||||
|
private final ContainerCloser closer;
|
||||||
|
private final long size;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Constructs a mapping class that creates mapping between container names
|
* Constructs a mapping class that creates mapping between container names
|
||||||
@ -98,6 +105,7 @@ public ContainerMapping(
|
|||||||
cacheSizeMB) throws IOException {
|
cacheSizeMB) throws IOException {
|
||||||
this.nodeManager = nodeManager;
|
this.nodeManager = nodeManager;
|
||||||
this.cacheSize = cacheSizeMB;
|
this.cacheSize = cacheSizeMB;
|
||||||
|
this.closer = new ContainerCloser(nodeManager, conf);
|
||||||
|
|
||||||
File metaDir = OzoneUtils.getOzoneMetaDirPath(conf);
|
File metaDir = OzoneUtils.getOzoneMetaDirPath(conf);
|
||||||
|
|
||||||
@ -113,6 +121,10 @@ public ContainerMapping(
|
|||||||
this.lock = new ReentrantLock();
|
this.lock = new ReentrantLock();
|
||||||
|
|
||||||
this.pipelineSelector = new PipelineSelector(nodeManager, conf);
|
this.pipelineSelector = new PipelineSelector(nodeManager, conf);
|
||||||
|
|
||||||
|
// To be replaced with code getStorageSize once it is committed.
|
||||||
|
size = conf.getLong(OZONE_SCM_CONTAINER_SIZE_GB,
|
||||||
|
OZONE_SCM_CONTAINER_SIZE_DEFAULT) * 1024 * 1024 * 1024;
|
||||||
this.containerStateManager =
|
this.containerStateManager =
|
||||||
new ContainerStateManager(conf, this);
|
new ContainerStateManager(conf, this);
|
||||||
this.containerSupervisor =
|
this.containerSupervisor =
|
||||||
@ -342,6 +354,7 @@ public OzoneProtos.LifeCycleState updateContainerState(
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Returns the container State Manager.
|
* Returns the container State Manager.
|
||||||
|
*
|
||||||
* @return ContainerStateManager
|
* @return ContainerStateManager
|
||||||
*/
|
*/
|
||||||
@Override
|
@Override
|
||||||
@ -351,6 +364,18 @@ public ContainerStateManager getStateManager() {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Process container report from Datanode.
|
* Process container report from Datanode.
|
||||||
|
* <p>
|
||||||
|
* Processing follows a very simple logic for time being.
|
||||||
|
* <p>
|
||||||
|
* 1. Datanodes report the current State -- denoted by the datanodeState
|
||||||
|
* <p>
|
||||||
|
* 2. We are the older SCM state from the Database -- denoted by
|
||||||
|
* the knownState.
|
||||||
|
* <p>
|
||||||
|
* 3. We copy the usage etc. from currentState to newState and log that
|
||||||
|
* newState to the DB. This allows us SCM to bootup again and read the
|
||||||
|
* state of the world from the DB, and then reconcile the state from
|
||||||
|
* container reports, when they arrive.
|
||||||
*
|
*
|
||||||
* @param reports Container report
|
* @param reports Container report
|
||||||
*/
|
*/
|
||||||
@ -360,63 +385,37 @@ public void processContainerReports(ContainerReportsRequestProto reports)
|
|||||||
List<StorageContainerDatanodeProtocolProtos.ContainerInfo>
|
List<StorageContainerDatanodeProtocolProtos.ContainerInfo>
|
||||||
containerInfos = reports.getReportsList();
|
containerInfos = reports.getReportsList();
|
||||||
containerSupervisor.handleContainerReport(reports);
|
containerSupervisor.handleContainerReport(reports);
|
||||||
for (StorageContainerDatanodeProtocolProtos.ContainerInfo containerInfo :
|
for (StorageContainerDatanodeProtocolProtos.ContainerInfo datanodeState :
|
||||||
containerInfos) {
|
containerInfos) {
|
||||||
byte[] dbKey = containerInfo.getContainerNameBytes().toByteArray();
|
byte[] dbKey = datanodeState.getContainerNameBytes().toByteArray();
|
||||||
lock.lock();
|
lock.lock();
|
||||||
try {
|
try {
|
||||||
byte[] containerBytes = containerStore.get(dbKey);
|
byte[] containerBytes = containerStore.get(dbKey);
|
||||||
if (containerBytes != null) {
|
if (containerBytes != null) {
|
||||||
OzoneProtos.SCMContainerInfo oldInfo =
|
OzoneProtos.SCMContainerInfo knownState =
|
||||||
OzoneProtos.SCMContainerInfo.PARSER.parseFrom(containerBytes);
|
OzoneProtos.SCMContainerInfo.PARSER.parseFrom(containerBytes);
|
||||||
|
|
||||||
OzoneProtos.SCMContainerInfo.Builder builder =
|
OzoneProtos.SCMContainerInfo newState =
|
||||||
OzoneProtos.SCMContainerInfo.newBuilder();
|
reconcileState(datanodeState, knownState);
|
||||||
builder.setContainerName(oldInfo.getContainerName());
|
|
||||||
builder.setPipeline(oldInfo.getPipeline());
|
// FIX ME: This can be optimized, we write twice to memory, where a
|
||||||
// If used size is greater than allocated size, we will be updating
|
// single write would work well.
|
||||||
// allocated size with used size. This update is done as a fallback
|
//
|
||||||
// mechanism in case SCM crashes without properly updating allocated
|
// We need to write this to DB again since the closed only write
|
||||||
// size. Correct allocated value will be updated by
|
// the updated State.
|
||||||
// ContainerStateManager during SCM shutdown.
|
containerStore.put(dbKey, newState.toByteArray());
|
||||||
long usedSize = containerInfo.getUsed();
|
|
||||||
long allocated = oldInfo.getAllocatedBytes() > usedSize ?
|
// If the container is closed, then state is already written to SCM
|
||||||
oldInfo.getAllocatedBytes() : usedSize;
|
// DB.TODO: So can we can write only once to DB.
|
||||||
builder.setAllocatedBytes(allocated);
|
if (closeContainerIfNeeded(newState)) {
|
||||||
builder.setUsedBytes(containerInfo.getUsed());
|
LOG.info("Closing the Container: {}", newState.getContainerName());
|
||||||
builder.setNumberOfKeys(containerInfo.getKeyCount());
|
|
||||||
builder.setState(oldInfo.getState());
|
|
||||||
builder.setStateEnterTime(oldInfo.getStateEnterTime());
|
|
||||||
builder.setContainerID(oldInfo.getContainerID());
|
|
||||||
if (oldInfo.getOwner() != null) {
|
|
||||||
builder.setOwner(oldInfo.getOwner());
|
|
||||||
}
|
|
||||||
OzoneProtos.SCMContainerInfo newContainerInfo = builder.build();
|
|
||||||
containerStore.put(dbKey, newContainerInfo.toByteArray());
|
|
||||||
float containerUsedPercentage = 1.0f *
|
|
||||||
containerInfo.getUsed() / containerInfo.getSize();
|
|
||||||
// TODO: Handling of containers which are already in close queue.
|
|
||||||
if (containerUsedPercentage >= containerCloseThreshold) {
|
|
||||||
// TODO: The container has to be moved to close container queue.
|
|
||||||
// For now, we are just updating the container state to CLOSING.
|
|
||||||
// Close container implementation can decide on how to maintain
|
|
||||||
// list of containers to be closed, this is the place where we
|
|
||||||
// have to add the containers to that list.
|
|
||||||
OzoneProtos.LifeCycleState state = updateContainerState(
|
|
||||||
ContainerInfo.fromProtobuf(newContainerInfo).getContainerName(),
|
|
||||||
OzoneProtos.LifeCycleEvent.FINALIZE);
|
|
||||||
if (state != OzoneProtos.LifeCycleState.CLOSING) {
|
|
||||||
LOG.error("Failed to close container {}, reason : Not able to " +
|
|
||||||
"update container state, current container state: {}.",
|
|
||||||
containerInfo.getContainerName(), state);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
// Container not found in our container db.
|
// Container not found in our container db.
|
||||||
LOG.error("Error while processing container report from datanode :" +
|
LOG.error("Error while processing container report from datanode :" +
|
||||||
" {}, for container: {}, reason: container doesn't exist in" +
|
" {}, for container: {}, reason: container doesn't exist in" +
|
||||||
"container database.", reports.getDatanodeID(),
|
"container database.", reports.getDatanodeID(),
|
||||||
containerInfo.getContainerName());
|
datanodeState.getContainerName());
|
||||||
}
|
}
|
||||||
} finally {
|
} finally {
|
||||||
lock.unlock();
|
lock.unlock();
|
||||||
@ -424,11 +423,110 @@ public void processContainerReports(ContainerReportsRequestProto reports)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Reconciles the state from Datanode with the state in SCM.
|
||||||
|
*
|
||||||
|
* @param datanodeState - State from the Datanode.
|
||||||
|
* @param knownState - State inside SCM.
|
||||||
|
* @return new SCM State for this container.
|
||||||
|
*/
|
||||||
|
private OzoneProtos.SCMContainerInfo reconcileState(
|
||||||
|
StorageContainerDatanodeProtocolProtos.ContainerInfo datanodeState,
|
||||||
|
OzoneProtos.SCMContainerInfo knownState) {
|
||||||
|
OzoneProtos.SCMContainerInfo.Builder builder =
|
||||||
|
OzoneProtos.SCMContainerInfo.newBuilder();
|
||||||
|
builder.setContainerName(knownState.getContainerName());
|
||||||
|
builder.setPipeline(knownState.getPipeline());
|
||||||
|
// If used size is greater than allocated size, we will be updating
|
||||||
|
// allocated size with used size. This update is done as a fallback
|
||||||
|
// mechanism in case SCM crashes without properly updating allocated
|
||||||
|
// size. Correct allocated value will be updated by
|
||||||
|
// ContainerStateManager during SCM shutdown.
|
||||||
|
long usedSize = datanodeState.getUsed();
|
||||||
|
long allocated = knownState.getAllocatedBytes() > usedSize ?
|
||||||
|
knownState.getAllocatedBytes() : usedSize;
|
||||||
|
builder.setAllocatedBytes(allocated);
|
||||||
|
builder.setUsedBytes(usedSize);
|
||||||
|
builder.setNumberOfKeys(datanodeState.getKeyCount());
|
||||||
|
builder.setState(knownState.getState());
|
||||||
|
builder.setStateEnterTime(knownState.getStateEnterTime());
|
||||||
|
builder.setContainerID(knownState.getContainerID());
|
||||||
|
if (knownState.getOwner() != null) {
|
||||||
|
builder.setOwner(knownState.getOwner());
|
||||||
|
}
|
||||||
|
return builder.build();
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Queues the close container command, to datanode and writes the new state
|
||||||
|
* to container DB.
|
||||||
|
* <p>
|
||||||
|
* TODO : Remove this 2 ContainerInfo definitions. It is brain dead to have
|
||||||
|
* one protobuf in one file and another definition in another file.
|
||||||
|
*
|
||||||
|
* @param newState - This is the state we maintain in SCM.
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
private boolean closeContainerIfNeeded(OzoneProtos.SCMContainerInfo newState)
|
||||||
|
throws IOException {
|
||||||
|
float containerUsedPercentage = 1.0f *
|
||||||
|
newState.getUsedBytes() / this.size;
|
||||||
|
|
||||||
|
ContainerInfo scmInfo = getContainer(newState.getContainerName());
|
||||||
|
if (containerUsedPercentage >= containerCloseThreshold
|
||||||
|
&& !isClosed(scmInfo)) {
|
||||||
|
// We will call closer till get to the closed state.
|
||||||
|
// That is SCM will make this call repeatedly until we reach the closed
|
||||||
|
// state.
|
||||||
|
closer.close(newState);
|
||||||
|
|
||||||
|
if (shouldClose(scmInfo)) {
|
||||||
|
// This event moves the Container from Open to Closing State, this is
|
||||||
|
// a state inside SCM. This is the desired state that SCM wants this
|
||||||
|
// container to reach. We will know that a container has reached the
|
||||||
|
// closed state from container reports. This state change should be
|
||||||
|
// invoked once and only once.
|
||||||
|
OzoneProtos.LifeCycleState state = updateContainerState(
|
||||||
|
scmInfo.getContainerName(),
|
||||||
|
OzoneProtos.LifeCycleEvent.FINALIZE);
|
||||||
|
if (state != OzoneProtos.LifeCycleState.CLOSING) {
|
||||||
|
LOG.error("Failed to close container {}, reason : Not able " +
|
||||||
|
"to " +
|
||||||
|
"update container state, current container state: {}.",
|
||||||
|
newState.getContainerName(), state);
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* In Container is in closed state, if it is in closed, Deleting or Deleted
|
||||||
|
* State.
|
||||||
|
*
|
||||||
|
* @param info - ContainerInfo.
|
||||||
|
* @return true if is in open state, false otherwise
|
||||||
|
*/
|
||||||
|
private boolean shouldClose(ContainerInfo info) {
|
||||||
|
return info.getState() == OzoneProtos.LifeCycleState.OPEN;
|
||||||
|
}
|
||||||
|
|
||||||
|
private boolean isClosed(ContainerInfo info) {
|
||||||
|
return info.getState() == OzoneProtos.LifeCycleState.CLOSED;
|
||||||
|
}
|
||||||
|
|
||||||
|
@VisibleForTesting
|
||||||
|
public ContainerCloser getCloser() {
|
||||||
|
return closer;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Closes this stream and releases any system resources associated with it.
|
* Closes this stream and releases any system resources associated with it.
|
||||||
* If the stream is
|
* If the stream is
|
||||||
* already closed then invoking this method has no effect.
|
* already closed then invoking this method has no effect.
|
||||||
*
|
* <p>
|
||||||
* <p>As noted in {@link AutoCloseable#close()}, cases where the close may
|
* <p>As noted in {@link AutoCloseable#close()}, cases where the close may
|
||||||
* fail require careful
|
* fail require careful
|
||||||
* attention. It is strongly advised to relinquish the underlying resources
|
* attention. It is strongly advised to relinquish the underlying resources
|
||||||
@ -457,7 +555,7 @@ public void close() throws IOException {
|
|||||||
* containerStateManager, when closing ContainerMapping, we need to update
|
* containerStateManager, when closing ContainerMapping, we need to update
|
||||||
* this in the container store.
|
* this in the container store.
|
||||||
*
|
*
|
||||||
* @throws IOException on failure.
|
* @throws IOException on failure.
|
||||||
*/
|
*/
|
||||||
@VisibleForTesting
|
@VisibleForTesting
|
||||||
public void flushContainerInfo() throws IOException {
|
public void flushContainerInfo() throws IOException {
|
||||||
|
@ -0,0 +1,192 @@
|
|||||||
|
/*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||||
|
* contributor license agreements. See the NOTICE file distributed with this
|
||||||
|
* work for additional information regarding copyright ownership. The ASF
|
||||||
|
* licenses this file to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance with the License.
|
||||||
|
* You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||||
|
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||||
|
* License for the specific language governing permissions and limitations under
|
||||||
|
* the License.
|
||||||
|
*
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.ozone.scm.container.closer;
|
||||||
|
|
||||||
|
import com.google.common.annotations.VisibleForTesting;
|
||||||
|
import com.google.common.base.Preconditions;
|
||||||
|
import com.google.common.util.concurrent.ThreadFactoryBuilder;
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.hdfs.protocol.DatanodeID;
|
||||||
|
import org.apache.hadoop.hdfs.protocol.proto.HdfsProtos;
|
||||||
|
import org.apache.hadoop.ozone.protocol.commands.CloseContainerCommand;
|
||||||
|
import org.apache.hadoop.ozone.protocol.proto.OzoneProtos;
|
||||||
|
import org.apache.hadoop.ozone.scm.node.NodeManager;
|
||||||
|
import org.apache.hadoop.util.Time;
|
||||||
|
import org.slf4j.Logger;
|
||||||
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
|
import java.util.Map;
|
||||||
|
import java.util.concurrent.ConcurrentHashMap;
|
||||||
|
import java.util.concurrent.TimeUnit;
|
||||||
|
import java.util.concurrent.atomic.AtomicBoolean;
|
||||||
|
import java.util.concurrent.atomic.AtomicInteger;
|
||||||
|
|
||||||
|
import static org.apache.hadoop.ozone.OzoneConfigKeys
|
||||||
|
.OZONE_CONTAINER_REPORT_INTERVAL;
|
||||||
|
import static org.apache.hadoop.ozone.OzoneConfigKeys
|
||||||
|
.OZONE_CONTAINER_REPORT_INTERVAL_DEFAULT;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A class that manages closing of containers. This allows transition from a
|
||||||
|
* open but full container to a closed container, to which no data is written.
|
||||||
|
*/
|
||||||
|
public class ContainerCloser {
|
||||||
|
private static final Logger LOG =
|
||||||
|
LoggerFactory.getLogger(ContainerCloser.class);
|
||||||
|
private static final long MULTIPLIER = 3L;
|
||||||
|
private static final int CLEANUP_WATER_MARK = 1000;
|
||||||
|
private final NodeManager nodeManager;
|
||||||
|
private final Map<String, Long> commandIssued;
|
||||||
|
private final Configuration configuration;
|
||||||
|
private final AtomicInteger mapCount;
|
||||||
|
private final long reportInterval;
|
||||||
|
private final AtomicInteger threadRunCount;
|
||||||
|
private final AtomicBoolean isRunning;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Constructs the ContainerCloser class.
|
||||||
|
*
|
||||||
|
* @param nodeManager - NodeManager
|
||||||
|
* @param conf - Configuration
|
||||||
|
*/
|
||||||
|
public ContainerCloser(NodeManager nodeManager, Configuration conf) {
|
||||||
|
Preconditions.checkNotNull(nodeManager);
|
||||||
|
Preconditions.checkNotNull(conf);
|
||||||
|
this.nodeManager = nodeManager;
|
||||||
|
this.configuration = conf;
|
||||||
|
this.commandIssued = new ConcurrentHashMap<>();
|
||||||
|
this.mapCount = new AtomicInteger(0);
|
||||||
|
this.threadRunCount = new AtomicInteger(0);
|
||||||
|
this.isRunning = new AtomicBoolean(false);
|
||||||
|
this.reportInterval = this.configuration.getTimeDuration(
|
||||||
|
OZONE_CONTAINER_REPORT_INTERVAL,
|
||||||
|
OZONE_CONTAINER_REPORT_INTERVAL_DEFAULT, TimeUnit.SECONDS);
|
||||||
|
Preconditions.checkState(this.reportInterval > 0,
|
||||||
|
"report interval has to be greater than 0");
|
||||||
|
}
|
||||||
|
|
||||||
|
@VisibleForTesting
|
||||||
|
public static int getCleanupWaterMark() {
|
||||||
|
return CLEANUP_WATER_MARK;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sends a Container Close command to the data nodes where this container
|
||||||
|
* lives.
|
||||||
|
*
|
||||||
|
* @param info - ContainerInfo.
|
||||||
|
*/
|
||||||
|
public void close(OzoneProtos.SCMContainerInfo info) {
|
||||||
|
|
||||||
|
if (commandIssued.containsKey(info.getContainerName())) {
|
||||||
|
// We check if we issued a close command in last 3 * reportInterval secs.
|
||||||
|
long commandQueueTime = commandIssued.get(info.getContainerName());
|
||||||
|
long currentTime = TimeUnit.MILLISECONDS.toSeconds(Time.monotonicNow());
|
||||||
|
if (currentTime > commandQueueTime + (MULTIPLIER * reportInterval)) {
|
||||||
|
commandIssued.remove(info.getContainerName());
|
||||||
|
mapCount.decrementAndGet();
|
||||||
|
} else {
|
||||||
|
// Ignore this request, since we just issued a close command. We
|
||||||
|
// should wait instead of sending a command to datanode again.
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// if we reached here, it means that we have not issued a command to the
|
||||||
|
// data node in last (3 times report interval). We are presuming that is
|
||||||
|
// enough time to close the container. Let us go ahead and queue a close
|
||||||
|
// to all the datanodes that participate in the container.
|
||||||
|
//
|
||||||
|
// Three important things to note here:
|
||||||
|
//
|
||||||
|
// 1. It is ok to send this command multiple times to a datanode. Close
|
||||||
|
// container is an idempotent command, if the container is already closed
|
||||||
|
// then we have no issues.
|
||||||
|
//
|
||||||
|
// 2. The container close command is issued to all datanodes. But
|
||||||
|
// depending on the pipeline type, some of the datanodes might ignore it.
|
||||||
|
//
|
||||||
|
// 3. SCM will see that datanode is closed from container reports, but it
|
||||||
|
// is possible that datanodes might get close commands since
|
||||||
|
// this queue can be emptied by a datanode after a close report is send
|
||||||
|
// to SCM. In that case also, data node will ignore this command.
|
||||||
|
|
||||||
|
OzoneProtos.Pipeline pipeline = info.getPipeline();
|
||||||
|
for (HdfsProtos.DatanodeIDProto datanodeID :
|
||||||
|
pipeline.getPipelineChannel().getMembersList()) {
|
||||||
|
nodeManager.addDatanodeCommand(DatanodeID.getFromProtoBuf(datanodeID),
|
||||||
|
new CloseContainerCommand(info.getContainerName()));
|
||||||
|
}
|
||||||
|
if (!commandIssued.containsKey(info.getContainerName())) {
|
||||||
|
commandIssued.put(info.getContainerName(),
|
||||||
|
TimeUnit.MILLISECONDS.toSeconds(Time.monotonicNow()));
|
||||||
|
mapCount.incrementAndGet();
|
||||||
|
}
|
||||||
|
// run the hash map cleaner thread if needed, non-blocking call.
|
||||||
|
runCleanerThreadIfNeeded();
|
||||||
|
}
|
||||||
|
|
||||||
|
private void runCleanerThreadIfNeeded() {
|
||||||
|
// Let us check if we should run a cleaner thread, not using map.size
|
||||||
|
// since it runs a loop in the case of the concurrentMap.
|
||||||
|
if (mapCount.get() > CLEANUP_WATER_MARK &&
|
||||||
|
isRunning.compareAndSet(false, true)) {
|
||||||
|
Runnable entryCleaner = () -> {
|
||||||
|
LOG.debug("Starting close container Hash map cleaner.");
|
||||||
|
try {
|
||||||
|
for (Map.Entry<String, Long> entry : commandIssued.entrySet()) {
|
||||||
|
long commandQueueTime = entry.getValue();
|
||||||
|
if (commandQueueTime + (MULTIPLIER * reportInterval) >
|
||||||
|
TimeUnit.MILLISECONDS.toSeconds(Time.monotonicNow())) {
|
||||||
|
|
||||||
|
// It is possible for this remove to fail due to race conditions.
|
||||||
|
// No big deal we will cleanup next time.
|
||||||
|
commandIssued.remove(entry.getKey());
|
||||||
|
mapCount.decrementAndGet();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
isRunning.compareAndSet(true, false);
|
||||||
|
LOG.debug("Finished running, close container Hash map cleaner.");
|
||||||
|
} catch (Exception ex) {
|
||||||
|
LOG.error("Unable to finish cleaning the closed containers map.", ex);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
// Launch the cleaner thread when we need instead of having a daemon
|
||||||
|
// thread that is sleeping all the time. We need to set the Daemon to
|
||||||
|
// true to avoid blocking clean exits.
|
||||||
|
Thread cleanerThread = new ThreadFactoryBuilder()
|
||||||
|
.setDaemon(true)
|
||||||
|
.setNameFormat("Closed Container Cleaner Thread - %d")
|
||||||
|
.build().newThread(entryCleaner);
|
||||||
|
threadRunCount.incrementAndGet();
|
||||||
|
cleanerThread.start();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@VisibleForTesting
|
||||||
|
public int getThreadRunCount() {
|
||||||
|
return threadRunCount.get();
|
||||||
|
}
|
||||||
|
|
||||||
|
@VisibleForTesting
|
||||||
|
public int getCloseCount() {
|
||||||
|
return mapCount.get();
|
||||||
|
}
|
||||||
|
}
|
@ -0,0 +1,23 @@
|
|||||||
|
/*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||||
|
* contributor license agreements. See the NOTICE file distributed with this
|
||||||
|
* work for additional information regarding copyright ownership. The ASF
|
||||||
|
* licenses this file to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance with the License.
|
||||||
|
* You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||||
|
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||||
|
* License for the specific language governing permissions and limitations under
|
||||||
|
* the License.
|
||||||
|
*
|
||||||
|
*/
|
||||||
|
|
||||||
|
/**
|
||||||
|
* This package has class that close a container. That is move a container from
|
||||||
|
* open state to close state.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.ozone.scm.container.closer;
|
@ -148,5 +148,5 @@ public interface NodeManager extends StorageContainerNodeProtocol,
|
|||||||
* @param id
|
* @param id
|
||||||
* @param command
|
* @param command
|
||||||
*/
|
*/
|
||||||
default void addDatanodeCommand(DatanodeID id, SCMCommand command) {}
|
void addDatanodeCommand(DatanodeID id, SCMCommand command);
|
||||||
}
|
}
|
||||||
|
@ -1106,6 +1106,14 @@
|
|||||||
for more info.
|
for more info.
|
||||||
</description>
|
</description>
|
||||||
</property>
|
</property>
|
||||||
|
<property>
|
||||||
|
<name>ozone.scm.max.nodepool.processing.threads</name>
|
||||||
|
<value>1</value>
|
||||||
|
<tag>OZONE, MANAGEMENT, PERFORMANCE</tag>
|
||||||
|
<description>
|
||||||
|
Number of node pools to process in parallel.
|
||||||
|
</description>
|
||||||
|
</property>
|
||||||
<property>
|
<property>
|
||||||
<name>ozone.scm.names</name>
|
<name>ozone.scm.names</name>
|
||||||
<value/>
|
<value/>
|
||||||
|
@ -37,6 +37,7 @@
|
|||||||
import org.apache.hadoop.ozone.scm.node.NodeManager;
|
import org.apache.hadoop.ozone.scm.node.NodeManager;
|
||||||
import org.apache.hadoop.ozone.scm.node.NodePoolManager;
|
import org.apache.hadoop.ozone.scm.node.NodePoolManager;
|
||||||
import org.mockito.Mockito;
|
import org.mockito.Mockito;
|
||||||
|
import org.assertj.core.util.Preconditions;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.HashMap;
|
import java.util.HashMap;
|
||||||
@ -73,6 +74,7 @@ public class MockNodeManager implements NodeManager {
|
|||||||
private final Map<String, SCMNodeStat> nodeMetricMap;
|
private final Map<String, SCMNodeStat> nodeMetricMap;
|
||||||
private final SCMNodeStat aggregateStat;
|
private final SCMNodeStat aggregateStat;
|
||||||
private boolean chillmode;
|
private boolean chillmode;
|
||||||
|
private final Map<DatanodeID, List<SCMCommand>> commandMap;
|
||||||
|
|
||||||
public MockNodeManager(boolean initializeFakeNodes, int nodeCount) {
|
public MockNodeManager(boolean initializeFakeNodes, int nodeCount) {
|
||||||
this.healthyNodes = new LinkedList<>();
|
this.healthyNodes = new LinkedList<>();
|
||||||
@ -87,6 +89,7 @@ public MockNodeManager(boolean initializeFakeNodes, int nodeCount) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
chillmode = false;
|
chillmode = false;
|
||||||
|
this.commandMap = new HashMap<>();
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -297,6 +300,31 @@ public OzoneProtos.NodeState getNodeState(DatanodeID id) {
|
|||||||
return null;
|
return null;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void addDatanodeCommand(DatanodeID id, SCMCommand command) {
|
||||||
|
if(commandMap.containsKey(id)) {
|
||||||
|
List<SCMCommand> commandList = commandMap.get(id);
|
||||||
|
Preconditions.checkNotNull(commandList);
|
||||||
|
commandList.add(command);
|
||||||
|
} else {
|
||||||
|
List<SCMCommand> commandList = new LinkedList<>();
|
||||||
|
commandList.add(command);
|
||||||
|
commandMap.put(id, commandList);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Returns the number of commands that is queued to this node manager.
|
||||||
|
public int getCommandCount(DatanodeID id) {
|
||||||
|
List<SCMCommand> list = commandMap.get(id);
|
||||||
|
return (list == null) ? 0 : list.size();
|
||||||
|
}
|
||||||
|
|
||||||
|
public void clearCommandQueue(DatanodeID id) {
|
||||||
|
if(commandMap.containsKey(id)) {
|
||||||
|
commandMap.put(id, new LinkedList<>());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Closes this stream and releases any system resources associated with it. If
|
* Closes this stream and releases any system resources associated with it. If
|
||||||
* the stream is already closed then invoking this method has no effect.
|
* the stream is already closed then invoking this method has no effect.
|
||||||
|
@ -0,0 +1,221 @@
|
|||||||
|
/*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||||
|
* contributor license agreements. See the NOTICE file distributed with this
|
||||||
|
* work for additional information regarding copyright ownership. The ASF
|
||||||
|
* licenses this file to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance with the License.
|
||||||
|
* You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||||
|
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||||
|
* License for the specific language governing permissions and limitations under
|
||||||
|
* the License.
|
||||||
|
*
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.ozone.scm.container.closer;
|
||||||
|
|
||||||
|
import org.apache.commons.lang.RandomStringUtils;
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.fs.FileUtil;
|
||||||
|
import org.apache.hadoop.hdfs.DFSTestUtil;
|
||||||
|
import org.apache.hadoop.hdfs.protocol.DatanodeID;
|
||||||
|
import org.apache.hadoop.ozone.OzoneConfigKeys;
|
||||||
|
import org.apache.hadoop.ozone.container.common.SCMTestUtils;
|
||||||
|
import org.apache.hadoop.ozone.protocol.proto.OzoneProtos;
|
||||||
|
import org.apache.hadoop.ozone.protocol.proto.StorageContainerDatanodeProtocolProtos;
|
||||||
|
import org.apache.hadoop.ozone.protocol.proto.StorageContainerDatanodeProtocolProtos.ContainerReportsRequestProto;
|
||||||
|
import org.apache.hadoop.ozone.scm.container.ContainerMapping;
|
||||||
|
import org.apache.hadoop.ozone.scm.container.MockNodeManager;
|
||||||
|
import org.apache.hadoop.ozone.scm.container.TestContainerMapping;
|
||||||
|
import org.apache.hadoop.scm.container.common.helpers.ContainerInfo;
|
||||||
|
import org.apache.hadoop.test.GenericTestUtils;
|
||||||
|
import org.junit.AfterClass;
|
||||||
|
import org.junit.Assert;
|
||||||
|
import org.junit.BeforeClass;
|
||||||
|
import org.junit.Test;
|
||||||
|
|
||||||
|
import java.io.File;
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.util.concurrent.TimeUnit;
|
||||||
|
|
||||||
|
import static org.apache.hadoop.ozone.OzoneConfigKeys.OZONE_CONTAINER_REPORT_INTERVAL;
|
||||||
|
import static org.apache.hadoop.ozone.protocol.proto.OzoneProtos.LifeCycleEvent.CREATE;
|
||||||
|
import static org.apache.hadoop.ozone.protocol.proto.OzoneProtos.LifeCycleEvent.CREATED;
|
||||||
|
import static org.apache.hadoop.scm.ScmConfigKeys.OZONE_SCM_CONTAINER_SIZE_DEFAULT;
|
||||||
|
import static org.apache.hadoop.scm.ScmConfigKeys.OZONE_SCM_CONTAINER_SIZE_GB;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Test class for Closing Container.
|
||||||
|
*/
|
||||||
|
public class TestContainerCloser {
|
||||||
|
|
||||||
|
private static final long GIGABYTE = 1024L * 1024L * 1024L;
|
||||||
|
private static Configuration configuration;
|
||||||
|
private static MockNodeManager nodeManager;
|
||||||
|
private static ContainerMapping mapping;
|
||||||
|
private static long size;
|
||||||
|
private static File testDir;
|
||||||
|
|
||||||
|
@BeforeClass
|
||||||
|
public static void setUp() throws Exception {
|
||||||
|
configuration = SCMTestUtils.getConf();
|
||||||
|
size = configuration.getLong(OZONE_SCM_CONTAINER_SIZE_GB,
|
||||||
|
OZONE_SCM_CONTAINER_SIZE_DEFAULT) * 1024 * 1024 * 1024;
|
||||||
|
configuration.setTimeDuration(OZONE_CONTAINER_REPORT_INTERVAL,
|
||||||
|
1, TimeUnit.SECONDS);
|
||||||
|
testDir = GenericTestUtils
|
||||||
|
.getTestDir(TestContainerMapping.class.getSimpleName());
|
||||||
|
configuration.set(OzoneConfigKeys.OZONE_METADATA_DIRS,
|
||||||
|
testDir.getAbsolutePath());
|
||||||
|
nodeManager = new MockNodeManager(true, 10);
|
||||||
|
mapping = new ContainerMapping(configuration, nodeManager, 128);
|
||||||
|
}
|
||||||
|
|
||||||
|
@AfterClass
|
||||||
|
public static void tearDown() throws Exception {
|
||||||
|
if (mapping != null) {
|
||||||
|
mapping.close();
|
||||||
|
}
|
||||||
|
FileUtil.fullyDelete(testDir);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testClose() throws IOException {
|
||||||
|
String containerName = "container-" + RandomStringUtils.randomNumeric(5);
|
||||||
|
|
||||||
|
ContainerInfo info = mapping.allocateContainer(
|
||||||
|
OzoneProtos.ReplicationType.STAND_ALONE,
|
||||||
|
OzoneProtos.ReplicationFactor.ONE, containerName, "ozone");
|
||||||
|
|
||||||
|
//Execute these state transitions so that we can close the container.
|
||||||
|
mapping.updateContainerState(containerName, CREATE);
|
||||||
|
mapping.updateContainerState(containerName, CREATED);
|
||||||
|
long currentCount = mapping.getCloser().getCloseCount();
|
||||||
|
long runCount = mapping.getCloser().getThreadRunCount();
|
||||||
|
|
||||||
|
DatanodeID datanodeID = info.getPipeline().getLeader();
|
||||||
|
// Send a container report with used set to 1 GB. This should not close.
|
||||||
|
sendContainerReport(info, 1 * GIGABYTE);
|
||||||
|
|
||||||
|
// with only one container the cleaner thread should not run.
|
||||||
|
Assert.assertEquals(0, mapping.getCloser().getThreadRunCount());
|
||||||
|
|
||||||
|
// With only 1 GB, the container should not be queued for closing.
|
||||||
|
Assert.assertEquals(0, mapping.getCloser().getCloseCount());
|
||||||
|
|
||||||
|
// Assert that the Close command was not queued for this Datanode.
|
||||||
|
Assert.assertEquals(0, nodeManager.getCommandCount(datanodeID));
|
||||||
|
|
||||||
|
long newUsed = (long) (size * 0.91f);
|
||||||
|
sendContainerReport(info, newUsed);
|
||||||
|
|
||||||
|
// with only one container the cleaner thread should not run.
|
||||||
|
Assert.assertEquals(runCount, mapping.getCloser().getThreadRunCount());
|
||||||
|
|
||||||
|
// and close count will be one.
|
||||||
|
Assert.assertEquals(1,
|
||||||
|
mapping.getCloser().getCloseCount() - currentCount);
|
||||||
|
|
||||||
|
// Assert that the Close command was Queued for this Datanode.
|
||||||
|
Assert.assertEquals(1, nodeManager.getCommandCount(datanodeID));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testRepeatedClose() throws IOException,
|
||||||
|
InterruptedException {
|
||||||
|
// This test asserts that if we queue more than one report then the
|
||||||
|
// second report is discarded by the system if it lands in the 3 * report
|
||||||
|
// frequency window.
|
||||||
|
|
||||||
|
configuration.setTimeDuration(OZONE_CONTAINER_REPORT_INTERVAL, 1,
|
||||||
|
TimeUnit.SECONDS);
|
||||||
|
String containerName = "container-" + RandomStringUtils.randomNumeric(5);
|
||||||
|
|
||||||
|
ContainerInfo info = mapping.allocateContainer(
|
||||||
|
OzoneProtos.ReplicationType.STAND_ALONE,
|
||||||
|
OzoneProtos.ReplicationFactor.ONE, containerName, "ozone");
|
||||||
|
|
||||||
|
//Execute these state transitions so that we can close the container.
|
||||||
|
mapping.updateContainerState(containerName, CREATE);
|
||||||
|
|
||||||
|
long currentCount = mapping.getCloser().getCloseCount();
|
||||||
|
long runCount = mapping.getCloser().getThreadRunCount();
|
||||||
|
|
||||||
|
|
||||||
|
DatanodeID datanodeID = info.getPipeline().getLeader();
|
||||||
|
|
||||||
|
// Send this command twice and assert we have only one command in the queue.
|
||||||
|
sendContainerReport(info, 5 * GIGABYTE);
|
||||||
|
sendContainerReport(info, 5 * GIGABYTE);
|
||||||
|
|
||||||
|
// Assert that the Close command was Queued for this Datanode.
|
||||||
|
Assert.assertEquals(1,
|
||||||
|
nodeManager.getCommandCount(datanodeID));
|
||||||
|
// And close count will be one.
|
||||||
|
Assert.assertEquals(1,
|
||||||
|
mapping.getCloser().getCloseCount() - currentCount);
|
||||||
|
Thread.sleep(TimeUnit.SECONDS.toMillis(4));
|
||||||
|
|
||||||
|
//send another close and the system will queue this to the command queue.
|
||||||
|
sendContainerReport(info, 5 * GIGABYTE);
|
||||||
|
Assert.assertEquals(2,
|
||||||
|
nodeManager.getCommandCount(datanodeID));
|
||||||
|
// but the close count will still be one, since from the point of view of
|
||||||
|
// closer we are closing only one container even if we have send multiple
|
||||||
|
// close commands to the datanode.
|
||||||
|
Assert.assertEquals(1, mapping.getCloser().getCloseCount()
|
||||||
|
- currentCount);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testCleanupThreadRuns() throws IOException,
|
||||||
|
InterruptedException {
|
||||||
|
// This test asserts that clean up thread runs once we have closed a
|
||||||
|
// number above cleanup water mark.
|
||||||
|
|
||||||
|
long runCount = mapping.getCloser().getThreadRunCount();
|
||||||
|
|
||||||
|
for (int x = 0; x < ContainerCloser.getCleanupWaterMark() + 10; x++) {
|
||||||
|
String containerName = "container-" + RandomStringUtils.randomNumeric(7);
|
||||||
|
ContainerInfo info = mapping.allocateContainer(
|
||||||
|
OzoneProtos.ReplicationType.STAND_ALONE,
|
||||||
|
OzoneProtos.ReplicationFactor.ONE, containerName, "ozone");
|
||||||
|
mapping.updateContainerState(containerName, CREATE);
|
||||||
|
mapping.updateContainerState(containerName, CREATED);
|
||||||
|
sendContainerReport(info, 5 * GIGABYTE);
|
||||||
|
}
|
||||||
|
|
||||||
|
Thread.sleep(TimeUnit.SECONDS.toMillis(1));
|
||||||
|
|
||||||
|
// Assert that cleanup thread ran at least once.
|
||||||
|
Assert.assertTrue(mapping.getCloser().getThreadRunCount() - runCount > 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
private void sendContainerReport(ContainerInfo info, long used) throws
|
||||||
|
IOException {
|
||||||
|
ContainerReportsRequestProto.Builder
|
||||||
|
reports = ContainerReportsRequestProto.newBuilder();
|
||||||
|
reports.setType(ContainerReportsRequestProto.reportType.fullReport);
|
||||||
|
|
||||||
|
StorageContainerDatanodeProtocolProtos.ContainerInfo.Builder ciBuilder =
|
||||||
|
StorageContainerDatanodeProtocolProtos.ContainerInfo.newBuilder();
|
||||||
|
ciBuilder.setContainerName(info.getContainerName())
|
||||||
|
.setFinalhash("e16cc9d6024365750ed8dbd194ea46d2")
|
||||||
|
.setSize(size)
|
||||||
|
.setUsed(used)
|
||||||
|
.setKeyCount(100000000L)
|
||||||
|
.setReadCount(100000000L)
|
||||||
|
.setWriteCount(100000000L)
|
||||||
|
.setReadBytes(2000000000L)
|
||||||
|
.setWriteBytes(2000000000L)
|
||||||
|
.setContainerID(1L);
|
||||||
|
reports.setDatanodeID(
|
||||||
|
DFSTestUtil.getLocalDatanodeID().getProtoBufMessage());
|
||||||
|
reports.addReports(ciBuilder);
|
||||||
|
mapping.processContainerReports(reports.build());
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue
Block a user