Revert "HDDS-705. OS3Exception resource name should be the actual resource name."
This reverts commit 977c6f6470
.
There was spurious edit in this commit.
This commit is contained in:
parent
e50334513c
commit
c696419f3e
@ -22,7 +22,6 @@
|
||||
import com.google.common.base.Preconditions;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.hdds.protocol.DatanodeDetails;
|
||||
import org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos;
|
||||
import org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos.ContainerCommandRequestProto;
|
||||
import org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos.ContainerCommandResponseProto;
|
||||
import org.apache.hadoop.hdds.protocol.datanode.proto.XceiverClientProtocolServiceGrpc;
|
||||
@ -41,9 +40,6 @@
|
||||
import java.io.IOException;
|
||||
import java.util.concurrent.CompletableFuture;
|
||||
import java.util.concurrent.ExecutionException;
|
||||
import java.util.UUID;
|
||||
import java.util.Map;
|
||||
import java.util.HashMap;
|
||||
import java.util.concurrent.Semaphore;
|
||||
import java.util.concurrent.TimeUnit;
|
||||
|
||||
@ -54,9 +50,9 @@ public class XceiverClientGrpc extends XceiverClientSpi {
|
||||
static final Logger LOG = LoggerFactory.getLogger(XceiverClientGrpc.class);
|
||||
private final Pipeline pipeline;
|
||||
private final Configuration config;
|
||||
private Map<UUID, XceiverClientProtocolServiceStub> asyncStubs;
|
||||
private XceiverClientProtocolServiceStub asyncStub;
|
||||
private XceiverClientMetrics metrics;
|
||||
private Map<UUID, ManagedChannel> channels;
|
||||
private ManagedChannel channel;
|
||||
private final Semaphore semaphore;
|
||||
private boolean closed = false;
|
||||
|
||||
@ -76,55 +72,40 @@ public XceiverClientGrpc(Pipeline pipeline, Configuration config) {
|
||||
this.semaphore =
|
||||
new Semaphore(HddsClientUtils.getMaxOutstandingRequests(config));
|
||||
this.metrics = XceiverClientManager.getXceiverClientMetrics();
|
||||
this.channels = new HashMap<>();
|
||||
this.asyncStubs = new HashMap<>();
|
||||
}
|
||||
|
||||
@Override
|
||||
public void connect() throws Exception {
|
||||
|
||||
// leader by default is the 1st datanode in the datanode list of pipleline
|
||||
DatanodeDetails leader = this.pipeline.getLeader();
|
||||
// just make a connection to the 1st datanode at the beginning
|
||||
connectToDatanode(leader);
|
||||
}
|
||||
|
||||
private void connectToDatanode(DatanodeDetails dn) {
|
||||
// read port from the data node, on failure use default configured
|
||||
// port.
|
||||
int port = dn.getPort(DatanodeDetails.Port.Name.STANDALONE).getValue();
|
||||
int port = leader.getPort(DatanodeDetails.Port.Name.STANDALONE).getValue();
|
||||
if (port == 0) {
|
||||
port = config.getInt(OzoneConfigKeys.DFS_CONTAINER_IPC_PORT,
|
||||
OzoneConfigKeys.DFS_CONTAINER_IPC_PORT_DEFAULT);
|
||||
}
|
||||
LOG.debug("Connecting to server Port : " + dn.getIpAddress());
|
||||
ManagedChannel channel =
|
||||
NettyChannelBuilder.forAddress(dn.getIpAddress(), port).usePlaintext()
|
||||
LOG.debug("Connecting to server Port : " + leader.getIpAddress());
|
||||
channel = NettyChannelBuilder.forAddress(leader.getIpAddress(), port)
|
||||
.usePlaintext()
|
||||
.maxInboundMessageSize(OzoneConfigKeys.DFS_CONTAINER_CHUNK_MAX_SIZE)
|
||||
.build();
|
||||
XceiverClientProtocolServiceStub asyncStub =
|
||||
XceiverClientProtocolServiceGrpc.newStub(channel);
|
||||
asyncStubs.put(dn.getUuid(), asyncStub);
|
||||
channels.put(dn.getUuid(), channel);
|
||||
asyncStub = XceiverClientProtocolServiceGrpc.newStub(channel);
|
||||
}
|
||||
|
||||
/**
|
||||
* Returns if the xceiver client connects to all servers in the pipeline.
|
||||
* Returns if the xceiver client connects to a server.
|
||||
*
|
||||
* @return True if the connection is alive, false otherwise.
|
||||
*/
|
||||
@VisibleForTesting
|
||||
public boolean isConnected(DatanodeDetails details) {
|
||||
return isConnected(channels.get(details.getUuid()));
|
||||
}
|
||||
|
||||
private boolean isConnected(ManagedChannel channel) {
|
||||
return channel != null && !channel.isTerminated() && !channel.isShutdown();
|
||||
public boolean isConnected() {
|
||||
return !channel.isTerminated() && !channel.isShutdown();
|
||||
}
|
||||
|
||||
@Override
|
||||
public void close() {
|
||||
closed = true;
|
||||
for (ManagedChannel channel : channels.values()) {
|
||||
channel.shutdownNow();
|
||||
try {
|
||||
channel.awaitTermination(60, TimeUnit.MINUTES);
|
||||
@ -133,58 +114,12 @@ public void close() {
|
||||
e);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public Pipeline getPipeline() {
|
||||
return pipeline;
|
||||
}
|
||||
|
||||
@Override
|
||||
public ContainerCommandResponseProto sendCommand(
|
||||
ContainerCommandRequestProto request) throws IOException {
|
||||
int size = pipeline.getMachines().size();
|
||||
ContainerCommandResponseProto responseProto = null;
|
||||
int dnIndex = 0;
|
||||
|
||||
// In case of an exception or an error, we will try to read from the
|
||||
// datanodes in the pipeline in a round robin fashion.
|
||||
|
||||
// TODO: cache the correct leader info in here, so that any subsequent calls
|
||||
// should first go to leader
|
||||
for (DatanodeDetails dn : pipeline.getMachines()) {
|
||||
try {
|
||||
|
||||
// In case the command gets retried on a 2nd datanode,
|
||||
// sendCommandAsyncCall will create a new channel and async stub
|
||||
// in case these don't exist for the specific datanode.
|
||||
responseProto =
|
||||
sendCommandAsync(request, dn).get();
|
||||
dnIndex++;
|
||||
if (responseProto.getResult() == ContainerProtos.Result.SUCCESS
|
||||
|| dnIndex == size) {
|
||||
return responseProto;
|
||||
}
|
||||
} catch (ExecutionException | InterruptedException e) {
|
||||
if (dnIndex < size) {
|
||||
LOG.warn(
|
||||
"Failed to execute command " + request + " on datanode " + dn
|
||||
.getUuidString() +". Retrying", e);
|
||||
} else {
|
||||
throw new IOException("Failed to execute command " + request, e);
|
||||
}
|
||||
}
|
||||
}
|
||||
return responseProto;
|
||||
}
|
||||
|
||||
// TODO: for a true async API, once the waitable future while executing
|
||||
// the command on one channel fails, it should be retried asynchronously
|
||||
// on the future Task for all the remaining datanodes.
|
||||
|
||||
// Note: this Async api is not used currently used in any active I/O path.
|
||||
// In case it gets used, the asynchronous retry logic needs to be plugged
|
||||
// in here.
|
||||
/**
|
||||
* Sends a given command to server gets a waitable future back.
|
||||
*
|
||||
@ -193,25 +128,15 @@ public ContainerCommandResponseProto sendCommand(
|
||||
* @throws IOException
|
||||
*/
|
||||
@Override
|
||||
public CompletableFuture<ContainerCommandResponseProto> sendCommandAsync(
|
||||
ContainerCommandRequestProto request)
|
||||
throws IOException, ExecutionException, InterruptedException {
|
||||
return sendCommandAsync(request, pipeline.getLeader());
|
||||
}
|
||||
|
||||
private CompletableFuture<ContainerCommandResponseProto> sendCommandAsync(
|
||||
ContainerCommandRequestProto request, DatanodeDetails dn)
|
||||
public CompletableFuture<ContainerCommandResponseProto>
|
||||
sendCommandAsync(ContainerCommandRequestProto request)
|
||||
throws IOException, ExecutionException, InterruptedException {
|
||||
if(closed){
|
||||
throw new IOException("This channel is not connected.");
|
||||
}
|
||||
|
||||
UUID dnId = dn.getUuid();
|
||||
ManagedChannel channel = channels.get(dnId);
|
||||
// If the channel doesn't exist for this specific datanode or the channel
|
||||
// is closed, just reconnect
|
||||
if (!isConnected(channel)) {
|
||||
reconnect(dn);
|
||||
if(channel == null || !isConnected()) {
|
||||
reconnect();
|
||||
}
|
||||
|
||||
final CompletableFuture<ContainerCommandResponseProto> replyFuture =
|
||||
@ -220,11 +145,8 @@ private CompletableFuture<ContainerCommandResponseProto> sendCommandAsync(
|
||||
long requestTime = Time.monotonicNowNanos();
|
||||
metrics.incrPendingContainerOpsMetrics(request.getCmdType());
|
||||
// create a new grpc stream for each non-async call.
|
||||
|
||||
// TODO: for async calls, we should reuse StreamObserver resources.
|
||||
final StreamObserver<ContainerCommandRequestProto> requestObserver =
|
||||
asyncStubs.get(dnId)
|
||||
.send(new StreamObserver<ContainerCommandResponseProto>() {
|
||||
asyncStub.send(new StreamObserver<ContainerCommandResponseProto>() {
|
||||
@Override
|
||||
public void onNext(ContainerCommandResponseProto value) {
|
||||
replyFuture.complete(value);
|
||||
@ -233,7 +155,6 @@ public void onNext(ContainerCommandResponseProto value) {
|
||||
Time.monotonicNowNanos() - requestTime);
|
||||
semaphore.release();
|
||||
}
|
||||
|
||||
@Override
|
||||
public void onError(Throwable t) {
|
||||
replyFuture.completeExceptionally(t);
|
||||
@ -246,8 +167,9 @@ public void onError(Throwable t) {
|
||||
@Override
|
||||
public void onCompleted() {
|
||||
if (!replyFuture.isDone()) {
|
||||
replyFuture.completeExceptionally(new IOException(
|
||||
"Stream completed but no reply for request " + request));
|
||||
replyFuture.completeExceptionally(
|
||||
new IOException("Stream completed but no reply for request "
|
||||
+ request));
|
||||
}
|
||||
}
|
||||
});
|
||||
@ -256,18 +178,15 @@ public void onCompleted() {
|
||||
return replyFuture;
|
||||
}
|
||||
|
||||
private void reconnect(DatanodeDetails dn)
|
||||
throws IOException {
|
||||
ManagedChannel channel;
|
||||
private void reconnect() throws IOException {
|
||||
try {
|
||||
connectToDatanode(dn);
|
||||
channel = channels.get(dn.getUuid());
|
||||
connect();
|
||||
} catch (Exception e) {
|
||||
LOG.error("Error while connecting: ", e);
|
||||
throw new IOException(e);
|
||||
}
|
||||
|
||||
if (channel == null || !isConnected(channel)) {
|
||||
if (channel == null || !isConnected()) {
|
||||
throw new IOException("This channel is not connected.");
|
||||
}
|
||||
}
|
||||
|
@ -27,6 +27,7 @@
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.hdds.scm.container.common.helpers.Pipeline;
|
||||
import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
|
||||
import org.apache.hadoop.hdds.scm.container.common.helpers.PipelineID;
|
||||
|
||||
import java.io.Closeable;
|
||||
import java.io.IOException;
|
||||
@ -58,7 +59,7 @@ public class XceiverClientManager implements Closeable {
|
||||
|
||||
//TODO : change this to SCM configuration class
|
||||
private final Configuration conf;
|
||||
private final Cache<String, XceiverClientSpi> clientCache;
|
||||
private final Cache<PipelineID, XceiverClientSpi> clientCache;
|
||||
private final boolean useRatis;
|
||||
|
||||
private static XceiverClientMetrics metrics;
|
||||
@ -82,10 +83,10 @@ public XceiverClientManager(Configuration conf) {
|
||||
.expireAfterAccess(staleThresholdMs, TimeUnit.MILLISECONDS)
|
||||
.maximumSize(maxSize)
|
||||
.removalListener(
|
||||
new RemovalListener<String, XceiverClientSpi>() {
|
||||
new RemovalListener<PipelineID, XceiverClientSpi>() {
|
||||
@Override
|
||||
public void onRemoval(
|
||||
RemovalNotification<String, XceiverClientSpi>
|
||||
RemovalNotification<PipelineID, XceiverClientSpi>
|
||||
removalNotification) {
|
||||
synchronized (clientCache) {
|
||||
// Mark the entry as evicted
|
||||
@ -97,7 +98,7 @@ public void onRemoval(
|
||||
}
|
||||
|
||||
@VisibleForTesting
|
||||
public Cache<String, XceiverClientSpi> getClientCache() {
|
||||
public Cache<PipelineID, XceiverClientSpi> getClientCache() {
|
||||
return clientCache;
|
||||
}
|
||||
|
||||
@ -139,14 +140,13 @@ public void releaseClient(XceiverClientSpi client) {
|
||||
|
||||
private XceiverClientSpi getClient(Pipeline pipeline)
|
||||
throws IOException {
|
||||
HddsProtos.ReplicationType type = pipeline.getType();
|
||||
try {
|
||||
return clientCache.get(pipeline.getId().getId().toString() + type,
|
||||
return clientCache.get(pipeline.getId(),
|
||||
new Callable<XceiverClientSpi>() {
|
||||
@Override
|
||||
public XceiverClientSpi call() throws Exception {
|
||||
XceiverClientSpi client = null;
|
||||
switch (type) {
|
||||
switch (pipeline.getType()) {
|
||||
case RATIS:
|
||||
client = XceiverClientRatis.newXceiverClientRatis(pipeline, conf);
|
||||
break;
|
||||
|
@ -18,11 +18,9 @@
|
||||
|
||||
package org.apache.hadoop.hdds.scm;
|
||||
|
||||
import com.google.common.annotations.VisibleForTesting;
|
||||
import org.apache.hadoop.hdds.HddsUtils;
|
||||
import org.apache.hadoop.hdds.scm.container.common.helpers.PipelineID;
|
||||
import org.apache.hadoop.io.MultipleIOException;
|
||||
import org.apache.ratis.proto.RaftProtos;
|
||||
import org.apache.ratis.retry.RetryPolicy;
|
||||
import org.apache.ratis.thirdparty.com.google.protobuf
|
||||
.InvalidProtocolBufferException;
|
||||
@ -54,7 +52,6 @@
|
||||
import java.util.Objects;
|
||||
import java.util.concurrent.CompletableFuture;
|
||||
import java.util.concurrent.CompletionException;
|
||||
import java.util.concurrent.TimeUnit;
|
||||
import java.util.concurrent.atomic.AtomicReference;
|
||||
|
||||
/**
|
||||
@ -212,11 +209,6 @@ private CompletableFuture<RaftClientReply> sendRequestAsync(
|
||||
getClient().sendAsync(() -> byteString);
|
||||
}
|
||||
|
||||
@VisibleForTesting
|
||||
public void watchForCommit(long index, long timeout) throws Exception {
|
||||
getClient().sendWatchAsync(index, RaftProtos.ReplicationLevel.ALL_COMMITTED)
|
||||
.get(timeout, TimeUnit.MILLISECONDS);
|
||||
}
|
||||
/**
|
||||
* Sends a given command to server gets a waitable future back.
|
||||
*
|
||||
|
@ -299,10 +299,6 @@ public String toString() {
|
||||
return b.toString();
|
||||
}
|
||||
|
||||
public void setType(HddsProtos.ReplicationType type) {
|
||||
this.type = type;
|
||||
}
|
||||
|
||||
/**
|
||||
* Returns a JSON string of this object.
|
||||
*
|
||||
|
@ -81,17 +81,14 @@ private ContainerProtocolCalls() {
|
||||
* @param xceiverClient client to perform call
|
||||
* @param datanodeBlockID blockID to identify container
|
||||
* @param traceID container protocol call args
|
||||
* @param blockCommitSequenceId latest commit Id of the block
|
||||
* @return container protocol get block response
|
||||
* @throws IOException if there is an I/O error while performing the call
|
||||
*/
|
||||
public static GetBlockResponseProto getBlock(XceiverClientSpi xceiverClient,
|
||||
DatanodeBlockID datanodeBlockID, String traceID,
|
||||
long blockCommitSequenceId) throws IOException {
|
||||
DatanodeBlockID datanodeBlockID, String traceID) throws IOException {
|
||||
GetBlockRequestProto.Builder readBlockRequest = GetBlockRequestProto
|
||||
.newBuilder()
|
||||
.setBlockID(datanodeBlockID)
|
||||
.setBlockCommitSequenceId(blockCommitSequenceId);
|
||||
.setBlockID(datanodeBlockID);
|
||||
String id = xceiverClient.getPipeline().getLeader().getUuidString();
|
||||
|
||||
ContainerCommandRequestProto request = ContainerCommandRequestProto
|
||||
@ -391,9 +388,7 @@ public static GetSmallFileResponseProto readSmallFile(XceiverClientSpi client,
|
||||
BlockID blockID, String traceID) throws IOException {
|
||||
GetBlockRequestProto.Builder getBlock = GetBlockRequestProto
|
||||
.newBuilder()
|
||||
.setBlockID(blockID.getDatanodeBlockIDProtobuf())
|
||||
// by default, set the bcsId to be 0
|
||||
.setBlockCommitSequenceId(0);
|
||||
.setBlockID(blockID.getDatanodeBlockIDProtobuf());
|
||||
ContainerProtos.GetSmallFileRequestProto getSmallFileRequest =
|
||||
GetSmallFileRequestProto
|
||||
.newBuilder().setBlock(getBlock)
|
||||
|
@ -140,8 +140,6 @@ enum Result {
|
||||
UNKNOWN_CONTAINER_TYPE = 34;
|
||||
BLOCK_NOT_COMMITTED = 35;
|
||||
CONTAINER_UNHEALTHY = 36;
|
||||
UNKNOWN_BCSID = 37;
|
||||
BCSID_MISMATCH = 38;
|
||||
}
|
||||
|
||||
/**
|
||||
@ -317,7 +315,6 @@ message PutBlockResponseProto {
|
||||
|
||||
message GetBlockRequestProto {
|
||||
required DatanodeBlockID blockID = 1;
|
||||
required uint64 blockCommitSequenceId = 2;
|
||||
}
|
||||
|
||||
message GetBlockResponseProto {
|
||||
|
@ -483,8 +483,7 @@ ContainerCommandResponseProto handleGetBlock(
|
||||
try {
|
||||
BlockID blockID = BlockID.getFromProtobuf(
|
||||
request.getGetBlock().getBlockID());
|
||||
responseData = blockManager.getBlock(kvContainer, blockID,
|
||||
request.getGetBlock().getBlockCommitSequenceId());
|
||||
responseData = blockManager.getBlock(kvContainer, blockID);
|
||||
long numBytes = responseData.getProtoBufMessage().toByteArray().length;
|
||||
metrics.incContainerBytesStats(Type.GetBlock, numBytes);
|
||||
|
||||
@ -756,8 +755,7 @@ ContainerCommandResponseProto handleGetSmallFile(
|
||||
try {
|
||||
BlockID blockID = BlockID.getFromProtobuf(getSmallFileReq.getBlock()
|
||||
.getBlockID());
|
||||
BlockData responseData = blockManager.getBlock(kvContainer, blockID,
|
||||
getSmallFileReq.getBlock().getBlockCommitSequenceId());
|
||||
BlockData responseData = blockManager.getBlock(kvContainer, blockID);
|
||||
|
||||
ContainerProtos.ChunkInfo chunkInfo = null;
|
||||
ByteString dataBuf = ByteString.EMPTY;
|
||||
|
@ -45,8 +45,7 @@
|
||||
import java.util.Map;
|
||||
|
||||
import static org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos.Result.NO_SUCH_BLOCK;
|
||||
import static org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos.Result.UNKNOWN_BCSID;
|
||||
import static org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos.Result.BCSID_MISMATCH;
|
||||
|
||||
/**
|
||||
* This class is for performing block related operations on the KeyValue
|
||||
* Container.
|
||||
@ -69,12 +68,6 @@ public BlockManagerImpl(Configuration conf) {
|
||||
this.config = conf;
|
||||
}
|
||||
|
||||
private long getBlockCommitSequenceId(MetadataStore db)
|
||||
throws IOException {
|
||||
byte[] bscId = db.get(blockCommitSequenceIdKey);
|
||||
return bscId == null ? 0 : Longs.fromByteArray(bscId);
|
||||
}
|
||||
|
||||
/**
|
||||
* Puts or overwrites a block.
|
||||
*
|
||||
@ -98,19 +91,21 @@ public long putBlock(Container container, BlockData data) throws IOException {
|
||||
Preconditions.checkNotNull(db, "DB cannot be null here");
|
||||
|
||||
long blockCommitSequenceId = data.getBlockCommitSequenceId();
|
||||
long blockCommitSequenceIdValue = getBlockCommitSequenceId(db);
|
||||
byte[] blockCommitSequenceIdValue = db.get(blockCommitSequenceIdKey);
|
||||
|
||||
// default blockCommitSequenceId for any block is 0. It the putBlock
|
||||
// request is not coming via Ratis(for test scenarios), it will be 0.
|
||||
// In such cases, we should overwrite the block as well
|
||||
if (blockCommitSequenceId != 0) {
|
||||
if (blockCommitSequenceId <= blockCommitSequenceIdValue) {
|
||||
if (blockCommitSequenceIdValue != null && blockCommitSequenceId != 0) {
|
||||
if (blockCommitSequenceId <= Longs
|
||||
.fromByteArray(blockCommitSequenceIdValue)) {
|
||||
// Since the blockCommitSequenceId stored in the db is greater than
|
||||
// equal to blockCommitSequenceId to be updated, it means the putBlock
|
||||
// transaction is reapplied in the ContainerStateMachine on restart.
|
||||
// It also implies that the given block must already exist in the db.
|
||||
// just log and return
|
||||
LOG.warn("blockCommitSequenceId " + blockCommitSequenceIdValue
|
||||
LOG.warn("blockCommitSequenceId " + Longs
|
||||
.fromByteArray(blockCommitSequenceIdValue)
|
||||
+ " in the Container Db is greater than" + " the supplied value "
|
||||
+ blockCommitSequenceId + " .Ignoring it");
|
||||
return data.getSize();
|
||||
@ -134,12 +129,10 @@ public long putBlock(Container container, BlockData data) throws IOException {
|
||||
*
|
||||
* @param container - Container from which block need to be fetched.
|
||||
* @param blockID - BlockID of the block.
|
||||
* @param bcsId latest commit Id of the block
|
||||
* @return Key Data.
|
||||
* @throws IOException
|
||||
*/
|
||||
@Override
|
||||
public BlockData getBlock(Container container, BlockID blockID, long bcsId)
|
||||
public BlockData getBlock(Container container, BlockID blockID)
|
||||
throws IOException {
|
||||
Preconditions.checkNotNull(blockID,
|
||||
"BlockID cannot be null in GetBlock request");
|
||||
@ -152,14 +145,6 @@ public BlockData getBlock(Container container, BlockID blockID, long bcsId)
|
||||
// This is a post condition that acts as a hint to the user.
|
||||
// Should never fail.
|
||||
Preconditions.checkNotNull(db, "DB cannot be null here");
|
||||
|
||||
long containerBCSId = getBlockCommitSequenceId(db);
|
||||
if (containerBCSId < bcsId) {
|
||||
throw new StorageContainerException(
|
||||
"Unable to find the block with bcsID " + bcsId + " .Container "
|
||||
+ container.getContainerData().getContainerID() + " bcsId is "
|
||||
+ containerBCSId + ".", UNKNOWN_BCSID);
|
||||
}
|
||||
byte[] kData = db.get(Longs.toByteArray(blockID.getLocalID()));
|
||||
if (kData == null) {
|
||||
throw new StorageContainerException("Unable to find the block.",
|
||||
@ -167,12 +152,6 @@ public BlockData getBlock(Container container, BlockID blockID, long bcsId)
|
||||
}
|
||||
ContainerProtos.BlockData blockData =
|
||||
ContainerProtos.BlockData.parseFrom(kData);
|
||||
long id = blockData.getBlockCommitSequenceId();
|
||||
if (id != bcsId) {
|
||||
throw new StorageContainerException(
|
||||
"bcsId " + bcsId + " mismatches with existing block Id "
|
||||
+ id + " for block " + blockID + ".", BCSID_MISMATCH);
|
||||
}
|
||||
return BlockData.getFromProtoBuf(blockData);
|
||||
}
|
||||
|
||||
|
@ -45,12 +45,10 @@ public interface BlockManager {
|
||||
*
|
||||
* @param container - Container from which block need to be get.
|
||||
* @param blockID - BlockID of the Block.
|
||||
* @param bcsId latest commit id of the block
|
||||
* @return Block Data.
|
||||
* @throws IOException
|
||||
*/
|
||||
BlockData getBlock(Container container, BlockID blockID, long bcsId)
|
||||
throws IOException;
|
||||
BlockData getBlock(Container container, BlockID blockID) throws IOException;
|
||||
|
||||
/**
|
||||
* Deletes an existing block.
|
||||
|
@ -113,7 +113,7 @@ public void testPutAndGetBlock() throws Exception {
|
||||
assertEquals(1, keyValueContainer.getContainerData().getKeyCount());
|
||||
//Get Block
|
||||
BlockData fromGetBlockData = blockManager.getBlock(keyValueContainer,
|
||||
blockData.getBlockID(), 0);
|
||||
blockData.getBlockID());
|
||||
|
||||
assertEquals(blockData.getContainerID(), fromGetBlockData.getContainerID());
|
||||
assertEquals(blockData.getLocalID(), fromGetBlockData.getLocalID());
|
||||
@ -139,7 +139,7 @@ public void testDeleteBlock() throws Exception {
|
||||
assertEquals(0,
|
||||
keyValueContainer.getContainerData().getKeyCount());
|
||||
try {
|
||||
blockManager.getBlock(keyValueContainer, blockID, 0);
|
||||
blockManager.getBlock(keyValueContainer, blockID);
|
||||
fail("testDeleteBlock");
|
||||
} catch (StorageContainerException ex) {
|
||||
GenericTestUtils.assertExceptionContains(
|
||||
@ -197,7 +197,7 @@ public void testGetNoSuchBlock() throws Exception {
|
||||
keyValueContainer.getContainerData().getKeyCount());
|
||||
try {
|
||||
//Since the block has been deleted, we should not be able to find it
|
||||
blockManager.getBlock(keyValueContainer, blockID, 0);
|
||||
blockManager.getBlock(keyValueContainer, blockID);
|
||||
fail("testGetNoSuchBlock failed");
|
||||
} catch (StorageContainerException ex) {
|
||||
GenericTestUtils.assertExceptionContains(
|
||||
|
@ -22,9 +22,7 @@
|
||||
import org.apache.hadoop.fs.Seekable;
|
||||
import org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos;
|
||||
import org.apache.hadoop.hdds.client.BlockID;
|
||||
import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
|
||||
import org.apache.hadoop.hdds.scm.container.common.helpers.ContainerWithPipeline;
|
||||
import org.apache.hadoop.hdds.scm.container.common.helpers.Pipeline;
|
||||
import org.apache.hadoop.ozone.om.helpers.OmKeyInfo;
|
||||
import org.apache.hadoop.ozone.om.helpers.OmKeyLocationInfo;
|
||||
import org.apache.hadoop.hdds.scm.XceiverClientManager;
|
||||
@ -278,13 +276,8 @@ public static LengthInputStream getFromOmKeyInfo(
|
||||
long containerID = blockID.getContainerID();
|
||||
ContainerWithPipeline containerWithPipeline =
|
||||
storageContainerLocationClient.getContainerWithPipeline(containerID);
|
||||
Pipeline pipeline = containerWithPipeline.getPipeline();
|
||||
|
||||
// irrespective of the container state, we will always read via Standalone
|
||||
// protocol.
|
||||
pipeline.setType(HddsProtos.ReplicationType.STAND_ALONE);
|
||||
XceiverClientSpi xceiverClient = xceiverClientManager
|
||||
.acquireClient(pipeline);
|
||||
.acquireClient(containerWithPipeline.getPipeline());
|
||||
boolean success = false;
|
||||
containerKey = omKeyLocationInfo.getLocalID();
|
||||
try {
|
||||
@ -294,8 +287,7 @@ public static LengthInputStream getFromOmKeyInfo(
|
||||
ContainerProtos.DatanodeBlockID datanodeBlockID = blockID
|
||||
.getDatanodeBlockIDProtobuf();
|
||||
ContainerProtos.GetBlockResponseProto response = ContainerProtocolCalls
|
||||
.getBlock(xceiverClient, datanodeBlockID, requestId,
|
||||
omKeyLocationInfo.getBlockCommitSequenceId());
|
||||
.getBlock(xceiverClient, datanodeBlockID, requestId);
|
||||
List<ContainerProtos.ChunkInfo> chunks =
|
||||
response.getBlockData().getChunksList();
|
||||
for (ContainerProtos.ChunkInfo chunk : chunks) {
|
||||
|
@ -116,10 +116,6 @@ public void addStream(OutputStream outputStream, long length) {
|
||||
public List<ChunkOutputStreamEntry> getStreamEntries() {
|
||||
return streamEntries;
|
||||
}
|
||||
@VisibleForTesting
|
||||
public XceiverClientManager getXceiverClientManager() {
|
||||
return xceiverClientManager;
|
||||
}
|
||||
|
||||
public List<OmKeyLocationInfo> getLocationInfoList() throws IOException {
|
||||
List<OmKeyLocationInfo> locationInfoList = new ArrayList<>();
|
||||
|
@ -102,7 +102,7 @@ public void testStartMultipleDatanodes() throws Exception {
|
||||
// Verify client is able to connect to the container
|
||||
try (XceiverClientGrpc client = new XceiverClientGrpc(pipeline, conf)){
|
||||
client.connect();
|
||||
assertTrue(client.isConnected(pipeline.getLeader()));
|
||||
assertTrue(client.isConnected());
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -25,11 +25,6 @@
|
||||
import org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos;
|
||||
import org.apache.hadoop.hdds.scm.container.ContainerID;
|
||||
import org.apache.hadoop.hdds.scm.container.ContainerInfo;
|
||||
import org.apache.hadoop.hdds.scm.XceiverClientManager;
|
||||
import org.apache.hadoop.hdds.scm.XceiverClientRatis;
|
||||
import org.apache.hadoop.hdds.scm.XceiverClientSpi;
|
||||
import org.apache.hadoop.hdds.scm.container.common.helpers.ContainerWithPipeline;
|
||||
|
||||
import org.apache.hadoop.hdds.scm.container.common.helpers.Pipeline;
|
||||
import org.apache.hadoop.ozone.*;
|
||||
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
|
||||
@ -37,7 +32,6 @@
|
||||
import org.apache.hadoop.hdds.client.OzoneQuota;
|
||||
import org.apache.hadoop.hdds.client.ReplicationFactor;
|
||||
import org.apache.hadoop.hdds.client.ReplicationType;
|
||||
import org.apache.hadoop.ozone.client.io.ChunkGroupOutputStream;
|
||||
import org.apache.hadoop.ozone.client.io.OzoneInputStream;
|
||||
import org.apache.hadoop.ozone.client.io.OzoneOutputStream;
|
||||
import org.apache.hadoop.ozone.container.common.helpers.BlockData;
|
||||
@ -603,106 +597,6 @@ public void testPutKeyRatisThreeNodes()
|
||||
}
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testPutKeyAndGetKeyThreeNodes()
|
||||
throws Exception {
|
||||
String volumeName = UUID.randomUUID().toString();
|
||||
String bucketName = UUID.randomUUID().toString();
|
||||
long currentTime = Time.now();
|
||||
|
||||
String value = "sample value";
|
||||
store.createVolume(volumeName);
|
||||
OzoneVolume volume = store.getVolume(volumeName);
|
||||
volume.createBucket(bucketName);
|
||||
OzoneBucket bucket = volume.getBucket(bucketName);
|
||||
|
||||
String keyName = UUID.randomUUID().toString();
|
||||
|
||||
OzoneOutputStream out = bucket
|
||||
.createKey(keyName, value.getBytes().length, ReplicationType.RATIS,
|
||||
ReplicationFactor.THREE);
|
||||
ChunkGroupOutputStream groupOutputStream =
|
||||
(ChunkGroupOutputStream) out.getOutputStream();
|
||||
XceiverClientManager manager = groupOutputStream.getXceiverClientManager();
|
||||
out.write(value.getBytes());
|
||||
out.close();
|
||||
// First, confirm the key info from the client matches the info in OM.
|
||||
OmKeyArgs.Builder builder = new OmKeyArgs.Builder();
|
||||
builder.setVolumeName(volumeName).setBucketName(bucketName)
|
||||
.setKeyName(keyName);
|
||||
OmKeyLocationInfo keyInfo = ozoneManager.lookupKey(builder.build()).
|
||||
getKeyLocationVersions().get(0).getBlocksLatestVersionOnly().get(0);
|
||||
long containerID = keyInfo.getContainerID();
|
||||
long localID = keyInfo.getLocalID();
|
||||
OzoneKeyDetails keyDetails = (OzoneKeyDetails) bucket.getKey(keyName);
|
||||
Assert.assertEquals(keyName, keyDetails.getName());
|
||||
|
||||
List<OzoneKeyLocation> keyLocations = keyDetails.getOzoneKeyLocations();
|
||||
Assert.assertEquals(1, keyLocations.size());
|
||||
Assert.assertEquals(containerID, keyLocations.get(0).getContainerID());
|
||||
Assert.assertEquals(localID, keyLocations.get(0).getLocalID());
|
||||
|
||||
// Make sure that the data size matched.
|
||||
Assert
|
||||
.assertEquals(value.getBytes().length, keyLocations.get(0).getLength());
|
||||
|
||||
ContainerWithPipeline container =
|
||||
cluster.getStorageContainerManager().getContainerManager()
|
||||
.getContainerWithPipeline(new ContainerID(containerID));
|
||||
Pipeline pipeline = container.getPipeline();
|
||||
List<DatanodeDetails> datanodes = pipeline.getMachines();
|
||||
|
||||
DatanodeDetails datanodeDetails = datanodes.get(0);
|
||||
Assert.assertNotNull(datanodeDetails);
|
||||
|
||||
XceiverClientSpi clientSpi = manager.acquireClient(pipeline);
|
||||
Assert.assertTrue(clientSpi instanceof XceiverClientRatis);
|
||||
XceiverClientRatis ratisClient = (XceiverClientRatis)clientSpi;
|
||||
|
||||
ratisClient.watchForCommit(keyInfo.getBlockCommitSequenceId(), 5000);
|
||||
// shutdown the datanode
|
||||
cluster.shutdownHddsDatanode(datanodeDetails);
|
||||
|
||||
Assert.assertTrue(container.getContainerInfo().getState()
|
||||
== HddsProtos.LifeCycleState.OPEN);
|
||||
// try to read, this shouls be successful
|
||||
readKey(bucket, keyName, value);
|
||||
|
||||
Assert.assertTrue(container.getContainerInfo().getState()
|
||||
== HddsProtos.LifeCycleState.OPEN);
|
||||
// shutdown the second datanode
|
||||
datanodeDetails = datanodes.get(1);
|
||||
cluster.shutdownHddsDatanode(datanodeDetails);
|
||||
Assert.assertTrue(container.getContainerInfo().getState()
|
||||
== HddsProtos.LifeCycleState.OPEN);
|
||||
|
||||
// the container is open and with loss of 2 nodes we still should be able
|
||||
// to read via Standalone protocol
|
||||
// try to read
|
||||
readKey(bucket, keyName, value);
|
||||
|
||||
// shutdown the 3rd datanode
|
||||
datanodeDetails = datanodes.get(2);
|
||||
cluster.shutdownHddsDatanode(datanodeDetails);
|
||||
try {
|
||||
// try to read
|
||||
readKey(bucket, keyName, value);
|
||||
Assert.fail("Expected exception not thrown");
|
||||
} catch (Exception e) {
|
||||
}
|
||||
manager.releaseClient(clientSpi);
|
||||
}
|
||||
|
||||
private void readKey(OzoneBucket bucket, String keyName, String data)
|
||||
throws IOException {
|
||||
OzoneKey key = bucket.getKey(keyName);
|
||||
Assert.assertEquals(keyName, key.getName());
|
||||
OzoneInputStream is = bucket.readKey(keyName);
|
||||
byte[] fileContent = new byte[data.getBytes().length];
|
||||
is.read(fileContent);
|
||||
is.close();
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testGetKeyDetails() throws IOException, OzoneException {
|
||||
String volumeName = UUID.randomUUID().toString();
|
||||
|
@ -468,7 +468,6 @@ public static ContainerCommandRequestProto getBlockRequest(
|
||||
ContainerProtos.GetBlockRequestProto.Builder getRequest =
|
||||
ContainerProtos.GetBlockRequestProto.newBuilder();
|
||||
getRequest.setBlockID(blockID);
|
||||
getRequest.setBlockCommitSequenceId(0);
|
||||
|
||||
ContainerCommandRequestProto.Builder request =
|
||||
ContainerCommandRequestProto.newBuilder();
|
||||
|
@ -151,7 +151,7 @@ public void testContainerReplication() throws Exception {
|
||||
.getHandler(ContainerType.KeyValueContainer);
|
||||
|
||||
BlockData key = handler.getBlockManager()
|
||||
.getBlock(container, BlockID.getFromProtobuf(blockID), 0);
|
||||
.getBlock(container, BlockID.getFromProtobuf(blockID));
|
||||
|
||||
Assert.assertNotNull(key);
|
||||
Assert.assertEquals(1, key.getChunks().size());
|
||||
|
@ -256,6 +256,6 @@ public void testCloseContainer() throws Exception {
|
||||
openContainerBlockMap.getBlockDataMap(testContainerID));
|
||||
// Make sure the key got committed
|
||||
Assert.assertNotNull(handler.getBlockManager()
|
||||
.getBlock(container, blockID, 0));
|
||||
.getBlock(container, blockID));
|
||||
}
|
||||
}
|
@ -556,7 +556,7 @@ public void testPutBlock() throws IOException, NoSuchAlgorithmException {
|
||||
blockData.setChunks(chunkList);
|
||||
blockManager.putBlock(container, blockData);
|
||||
BlockData readBlockData = blockManager.
|
||||
getBlock(container, blockData.getBlockID(), 0);
|
||||
getBlock(container, blockData.getBlockID());
|
||||
ChunkInfo readChunk =
|
||||
ChunkInfo.getFromProtoBuf(readBlockData.getChunks().get(0));
|
||||
Assert.assertEquals(info.getChecksum(), readChunk.getChecksum());
|
||||
@ -608,7 +608,7 @@ public void testPutBlockWithLotsOfChunks() throws IOException,
|
||||
blockData.setChunks(chunkProtoList);
|
||||
blockManager.putBlock(container, blockData);
|
||||
BlockData readBlockData = blockManager.
|
||||
getBlock(container, blockData.getBlockID(), 0);
|
||||
getBlock(container, blockData.getBlockID());
|
||||
ChunkInfo lastChunk = chunkList.get(chunkList.size() - 1);
|
||||
ChunkInfo readChunk =
|
||||
ChunkInfo.getFromProtoBuf(readBlockData.getChunks().get(readBlockData
|
||||
@ -636,7 +636,7 @@ public void testDeleteBlock() throws IOException, NoSuchAlgorithmException {
|
||||
blockManager.deleteBlock(container, blockID);
|
||||
exception.expect(StorageContainerException.class);
|
||||
exception.expectMessage("Unable to find the block.");
|
||||
blockManager.getBlock(container, blockData.getBlockID(), 0);
|
||||
blockManager.getBlock(container, blockData.getBlockID());
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -20,6 +20,7 @@
|
||||
import com.google.common.cache.Cache;
|
||||
import org.apache.commons.lang3.RandomStringUtils;
|
||||
import org.apache.hadoop.hdds.scm.container.common.helpers.ContainerWithPipeline;
|
||||
import org.apache.hadoop.hdds.scm.container.common.helpers.PipelineID;
|
||||
import org.apache.hadoop.io.IOUtils;
|
||||
import org.apache.hadoop.ozone.MiniOzoneCluster;
|
||||
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
|
||||
@ -106,7 +107,7 @@ public void testFreeByReference() throws IOException {
|
||||
OzoneConfiguration conf = new OzoneConfiguration();
|
||||
conf.setInt(SCM_CONTAINER_CLIENT_MAX_SIZE_KEY, 1);
|
||||
XceiverClientManager clientManager = new XceiverClientManager(conf);
|
||||
Cache<String, XceiverClientSpi> cache =
|
||||
Cache<PipelineID, XceiverClientSpi> cache =
|
||||
clientManager.getClientCache();
|
||||
|
||||
ContainerWithPipeline container1 =
|
||||
@ -129,9 +130,8 @@ public void testFreeByReference() throws IOException {
|
||||
Assert.assertNotEquals(client1, client2);
|
||||
|
||||
// least recent container (i.e containerName1) is evicted
|
||||
XceiverClientSpi nonExistent1 = cache.getIfPresent(
|
||||
container1.getContainerInfo().getPipelineID().getId().toString()
|
||||
+ container1.getContainerInfo().getReplicationType());
|
||||
XceiverClientSpi nonExistent1 = cache
|
||||
.getIfPresent(container1.getContainerInfo().getPipelineID());
|
||||
Assert.assertEquals(null, nonExistent1);
|
||||
// However container call should succeed because of refcount on the client.
|
||||
String traceID1 = "trace" + RandomStringUtils.randomNumeric(4);
|
||||
@ -160,7 +160,7 @@ public void testFreeByEviction() throws IOException {
|
||||
OzoneConfiguration conf = new OzoneConfiguration();
|
||||
conf.setInt(SCM_CONTAINER_CLIENT_MAX_SIZE_KEY, 1);
|
||||
XceiverClientManager clientManager = new XceiverClientManager(conf);
|
||||
Cache<String, XceiverClientSpi> cache =
|
||||
Cache<PipelineID, XceiverClientSpi> cache =
|
||||
clientManager.getClientCache();
|
||||
|
||||
ContainerWithPipeline container1 =
|
||||
@ -183,9 +183,8 @@ public void testFreeByEviction() throws IOException {
|
||||
Assert.assertNotEquals(client1, client2);
|
||||
|
||||
// now client 1 should be evicted
|
||||
XceiverClientSpi nonExistent = cache.getIfPresent(
|
||||
container1.getContainerInfo().getPipelineID().getId().toString()
|
||||
+ container1.getContainerInfo().getReplicationType());
|
||||
XceiverClientSpi nonExistent = cache
|
||||
.getIfPresent(container1.getContainerInfo().getPipelineID());
|
||||
Assert.assertEquals(null, nonExistent);
|
||||
|
||||
// Any container operation should now fail
|
||||
|
@ -699,8 +699,8 @@ public void testDeleteKey() throws Exception {
|
||||
.KeyValueContainer);
|
||||
KeyValueContainer container = (KeyValueContainer) cm.getContainerSet()
|
||||
.getContainer(location.getBlockID().getContainerID());
|
||||
BlockData blockInfo = keyValueHandler.getBlockManager()
|
||||
.getBlock(container, location.getBlockID(), 0);
|
||||
BlockData blockInfo = keyValueHandler
|
||||
.getBlockManager().getBlock(container, location.getBlockID());
|
||||
KeyValueContainerData containerData =
|
||||
(KeyValueContainerData) container.getContainerData();
|
||||
File dataDir = new File(containerData.getChunksPath());
|
||||
|
@ -199,11 +199,11 @@ public Response delete(@PathParam("bucket") String bucketName)
|
||||
} catch (IOException ex) {
|
||||
if (ex.getMessage().contains("BUCKET_NOT_EMPTY")) {
|
||||
OS3Exception os3Exception = S3ErrorTable.newError(S3ErrorTable
|
||||
.BUCKET_NOT_EMPTY, bucketName);
|
||||
.BUCKET_NOT_EMPTY, S3ErrorTable.Resource.BUCKET);
|
||||
throw os3Exception;
|
||||
} else if (ex.getMessage().contains("BUCKET_NOT_FOUND")) {
|
||||
OS3Exception os3Exception = S3ErrorTable.newError(S3ErrorTable
|
||||
.NO_SUCH_BUCKET, bucketName);
|
||||
.NO_SUCH_BUCKET, S3ErrorTable.Resource.BUCKET);
|
||||
throw os3Exception;
|
||||
} else {
|
||||
throw ex;
|
||||
|
@ -28,6 +28,7 @@
|
||||
import org.apache.hadoop.ozone.client.OzoneVolume;
|
||||
import org.apache.hadoop.ozone.s3.exception.OS3Exception;
|
||||
import org.apache.hadoop.ozone.s3.exception.S3ErrorTable;
|
||||
import org.apache.hadoop.ozone.s3.exception.S3ErrorTable.Resource;
|
||||
import org.apache.hadoop.ozone.s3.header.AuthorizationHeaderV2;
|
||||
import org.apache.hadoop.ozone.s3.header.AuthorizationHeaderV4;
|
||||
|
||||
@ -60,7 +61,7 @@ protected OzoneBucket getBucket(OzoneVolume volume, String bucketName)
|
||||
LOG.error("Error occurred is {}", ex);
|
||||
if (ex.getMessage().contains("NOT_FOUND")) {
|
||||
OS3Exception oex =
|
||||
S3ErrorTable.newError(S3ErrorTable.NO_SUCH_BUCKET, bucketName);
|
||||
S3ErrorTable.newError(S3ErrorTable.NO_SUCH_BUCKET, Resource.BUCKET);
|
||||
throw oex;
|
||||
} else {
|
||||
throw ex;
|
||||
@ -79,7 +80,7 @@ protected OzoneBucket getBucket(String bucketName)
|
||||
LOG.error("Error occurred is {}", ex);
|
||||
if (ex.getMessage().contains("NOT_FOUND")) {
|
||||
OS3Exception oex =
|
||||
S3ErrorTable.newError(S3ErrorTable.NO_SUCH_BUCKET, bucketName);
|
||||
S3ErrorTable.newError(S3ErrorTable.NO_SUCH_BUCKET, Resource.BUCKET);
|
||||
throw oex;
|
||||
} else {
|
||||
throw ex;
|
||||
@ -186,7 +187,7 @@ public String parseUsername(
|
||||
|
||||
if (auth == null) {
|
||||
throw S3ErrorTable
|
||||
.newError(S3ErrorTable.MALFORMED_HEADER, auth);
|
||||
.newError(S3ErrorTable.MALFORMED_HEADER, Resource.HEADER);
|
||||
}
|
||||
|
||||
String userName;
|
||||
|
@ -150,7 +150,7 @@ public Response get(
|
||||
} catch (IOException ex) {
|
||||
if (ex.getMessage().contains("NOT_FOUND")) {
|
||||
OS3Exception os3Exception = S3ErrorTable.newError(S3ErrorTable
|
||||
.NO_SUCH_KEY, keyPath);
|
||||
.NO_SUCH_OBJECT, S3ErrorTable.Resource.OBJECT);
|
||||
throw os3Exception;
|
||||
} else {
|
||||
throw ex;
|
||||
@ -176,8 +176,9 @@ public Response head(
|
||||
} catch (IOException ex) {
|
||||
LOG.error("Exception occurred in HeadObject", ex);
|
||||
if (ex.getMessage().contains("KEY_NOT_FOUND")) {
|
||||
// Just return 404 with no content
|
||||
return Response.status(Status.NOT_FOUND).build();
|
||||
OS3Exception os3Exception = S3ErrorTable.newError(S3ErrorTable
|
||||
.NO_SUCH_OBJECT, S3ErrorTable.Resource.OBJECT);
|
||||
throw os3Exception;
|
||||
} else {
|
||||
throw ex;
|
||||
}
|
||||
@ -214,7 +215,7 @@ public Response delete(
|
||||
} catch (IOException ex) {
|
||||
if (ex.getMessage().contains("BUCKET_NOT_FOUND")) {
|
||||
throw S3ErrorTable.newError(S3ErrorTable
|
||||
.NO_SUCH_BUCKET, bucketName);
|
||||
.NO_SUCH_BUCKET, S3ErrorTable.Resource.BUCKET);
|
||||
} else if (!ex.getMessage().contains("NOT_FOUND")) {
|
||||
throw ex;
|
||||
}
|
||||
|
@ -45,23 +45,52 @@ private S3ErrorTable() {
|
||||
"BucketNotEmpty", "The bucket you tried to delete is not empty.",
|
||||
HTTP_CONFLICT);
|
||||
|
||||
public static final OS3Exception NO_SUCH_OBJECT = new OS3Exception(
|
||||
"NoSuchObject", "The specified object does not exist", HTTP_NOT_FOUND);
|
||||
|
||||
|
||||
public static final OS3Exception MALFORMED_HEADER = new OS3Exception(
|
||||
"AuthorizationHeaderMalformed", "The authorization header you provided " +
|
||||
"is invalid.", HTTP_NOT_FOUND);
|
||||
|
||||
public static final OS3Exception NO_SUCH_KEY = new OS3Exception(
|
||||
"NoSuchObject", "The specified key does not exist", HTTP_NOT_FOUND);
|
||||
|
||||
/**
|
||||
* Create a new instance of Error.
|
||||
* @param e Error Template
|
||||
* @param resource Resource associated with this exception
|
||||
* @return creates a new instance of error based on the template
|
||||
*/
|
||||
public static OS3Exception newError(OS3Exception e, String resource) {
|
||||
public static OS3Exception newError(OS3Exception e, Resource resource) {
|
||||
OS3Exception err = new OS3Exception(e.getCode(), e.getErrorMessage(),
|
||||
e.getHttpCode());
|
||||
err.setResource(resource);
|
||||
err.setResource(resource.getResource());
|
||||
return err;
|
||||
}
|
||||
|
||||
/**
|
||||
* Resources, which can be defined in OS3Exception.
|
||||
*/
|
||||
public enum Resource {
|
||||
BUCKET("Bucket"),
|
||||
OBJECT("Object"),
|
||||
HEADER("header"),
|
||||
VOLUME("Volume");
|
||||
|
||||
private final String resource;
|
||||
|
||||
/**
|
||||
* Constructs resource.
|
||||
* @param value
|
||||
*/
|
||||
Resource(String value) {
|
||||
this.resource = value;
|
||||
}
|
||||
|
||||
/**
|
||||
* Get resource.
|
||||
* @return string
|
||||
*/
|
||||
public String getResource() {
|
||||
return this.resource;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -52,24 +52,28 @@ public AuthorizationHeaderV2(String auth) throws OS3Exception {
|
||||
public void parseHeader() throws OS3Exception {
|
||||
String[] split = authHeader.split(" ");
|
||||
if (split.length != 2) {
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, authHeader);
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, S3ErrorTable
|
||||
.Resource.HEADER);
|
||||
}
|
||||
|
||||
identifier = split[0];
|
||||
if (!IDENTIFIER.equals(identifier)) {
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, authHeader);
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, S3ErrorTable
|
||||
.Resource.HEADER);
|
||||
}
|
||||
|
||||
String[] remainingSplit = split[1].split(":");
|
||||
|
||||
if (remainingSplit.length != 2) {
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, authHeader);
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, S3ErrorTable
|
||||
.Resource.HEADER);
|
||||
}
|
||||
|
||||
accessKeyID = remainingSplit[0];
|
||||
signature = remainingSplit[1];
|
||||
if (isBlank(accessKeyID) || isBlank(signature)) {
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, authHeader);
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, S3ErrorTable
|
||||
.Resource.HEADER);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -64,7 +64,8 @@ public void parseAuthHeader() throws OS3Exception {
|
||||
String[] split = authHeader.split(" ");
|
||||
|
||||
if (split.length != 4) {
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, authHeader);
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, S3ErrorTable
|
||||
.Resource.HEADER);
|
||||
}
|
||||
|
||||
algorithm = split[0];
|
||||
@ -77,21 +78,24 @@ public void parseAuthHeader() throws OS3Exception {
|
||||
credential = credential.substring(CREDENTIAL.length(), credential
|
||||
.length() - 1);
|
||||
} else {
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, authHeader);
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, S3ErrorTable
|
||||
.Resource.HEADER);
|
||||
}
|
||||
|
||||
if (signedHeaders.startsWith(SIGNEDHEADERS)) {
|
||||
signedHeaders = signedHeaders.substring(SIGNEDHEADERS.length(),
|
||||
signedHeaders.length() - 1);
|
||||
} else {
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, authHeader);
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, S3ErrorTable
|
||||
.Resource.HEADER);
|
||||
}
|
||||
|
||||
if (signature.startsWith(SIGNATURE)) {
|
||||
signature = signature.substring(SIGNATURE.length(), signature
|
||||
.length());
|
||||
} else {
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, authHeader);
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, S3ErrorTable
|
||||
.Resource.HEADER);
|
||||
}
|
||||
|
||||
// Parse credential. Other parts of header are not validated yet. When
|
||||
|
@ -63,7 +63,8 @@ public void parseCredential() throws OS3Exception {
|
||||
awsService = split[3];
|
||||
awsRequest = split[4];
|
||||
} else {
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, credential);
|
||||
throw S3ErrorTable.newError(S3ErrorTable.MALFORMED_HEADER, S3ErrorTable
|
||||
.Resource.HEADER);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -32,7 +32,7 @@ public void testOS3Exception() {
|
||||
OS3Exception ex = new OS3Exception("AccessDenied", "Access Denied",
|
||||
403);
|
||||
String requestId = OzoneUtils.getRequestID();
|
||||
ex = S3ErrorTable.newError(ex, "bucket");
|
||||
ex = S3ErrorTable.newError(ex, S3ErrorTable.Resource.BUCKET);
|
||||
ex.setRequestId(requestId);
|
||||
String val = ex.toXml();
|
||||
String formatString = "<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n" +
|
||||
|
Loading…
Reference in New Issue
Block a user