HDFS-7678. Erasure coding: DFSInputStream with decode functionality (pread). Contributed by Zhe Zhang.
This commit is contained in:
parent
6bacaa9a52
commit
8d3030f064
@ -195,3 +195,6 @@
|
||||
|
||||
HDFS-8355. Erasure Coding: Refactor BlockInfo and BlockInfoUnderConstruction.
|
||||
(Tsz Wo Nicholas Sze via jing9)
|
||||
|
||||
HDFS-7678. Erasure coding: DFSInputStream with decode functionality (pread).
|
||||
(Zhe Zhang)
|
||||
|
@ -21,15 +21,27 @@
|
||||
import org.apache.hadoop.fs.ChecksumException;
|
||||
import org.apache.hadoop.fs.ReadOption;
|
||||
import org.apache.hadoop.fs.StorageType;
|
||||
import org.apache.hadoop.hdfs.protocol.*;
|
||||
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
||||
import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
|
||||
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
||||
import org.apache.hadoop.hdfs.protocol.LocatedBlock;
|
||||
import org.apache.hadoop.hdfs.protocol.LocatedStripedBlock;
|
||||
import org.apache.hadoop.hdfs.protocol.datatransfer.InvalidEncryptionKeyException;
|
||||
import org.apache.hadoop.hdfs.util.StripedBlockUtil;
|
||||
import org.apache.hadoop.io.ByteBufferPool;
|
||||
|
||||
import static org.apache.hadoop.hdfs.util.StripedBlockUtil.ReadPortion;
|
||||
import static org.apache.hadoop.hdfs.util.StripedBlockUtil.planReadPortions;
|
||||
import static org.apache.hadoop.hdfs.util.StripedBlockUtil.divideByteRangeIntoStripes;
|
||||
import static org.apache.hadoop.hdfs.util.StripedBlockUtil.initDecodeInputs;
|
||||
import static org.apache.hadoop.hdfs.util.StripedBlockUtil.decodeAndFillBuffer;
|
||||
import static org.apache.hadoop.hdfs.util.StripedBlockUtil.getNextCompletedStripedRead;
|
||||
import static org.apache.hadoop.hdfs.util.StripedBlockUtil.ReadPortion;
|
||||
import static org.apache.hadoop.hdfs.util.StripedBlockUtil.AlignedStripe;
|
||||
import static org.apache.hadoop.hdfs.util.StripedBlockUtil.StripingChunk;
|
||||
import static org.apache.hadoop.hdfs.util.StripedBlockUtil.StripingChunkReadResult;
|
||||
|
||||
import org.apache.hadoop.io.erasurecode.ECSchema;
|
||||
|
||||
import org.apache.hadoop.net.NetUtils;
|
||||
import org.apache.htrace.Span;
|
||||
import org.apache.htrace.Trace;
|
||||
@ -37,10 +49,12 @@
|
||||
|
||||
import java.io.EOFException;
|
||||
import java.io.IOException;
|
||||
import java.io.InterruptedIOException;
|
||||
import java.net.InetSocketAddress;
|
||||
import java.nio.ByteBuffer;
|
||||
import java.util.EnumSet;
|
||||
import java.util.Set;
|
||||
import java.util.Collection;
|
||||
import java.util.Map;
|
||||
import java.util.HashMap;
|
||||
import java.util.concurrent.CompletionService;
|
||||
@ -51,7 +65,6 @@
|
||||
import java.util.concurrent.Callable;
|
||||
import java.util.concurrent.Future;
|
||||
|
||||
|
||||
/******************************************************************************
|
||||
* DFSStripedInputStream reads from striped block groups, illustrated below:
|
||||
*
|
||||
@ -125,6 +138,7 @@ boolean include(long pos) {
|
||||
private final short parityBlkNum;
|
||||
/** the buffer for a complete stripe */
|
||||
private ByteBuffer curStripeBuf;
|
||||
private final ECSchema schema;
|
||||
/**
|
||||
* indicate the start/end offset of the current buffered stripe in the
|
||||
* block group
|
||||
@ -137,6 +151,7 @@ boolean include(long pos) {
|
||||
super(dfsClient, src, verifyChecksum);
|
||||
|
||||
assert schema != null;
|
||||
this.schema = schema;
|
||||
cellSize = schema.getChunkSize();
|
||||
dataBlkNum = (short) schema.getNumDataUnits();
|
||||
parityBlkNum = (short) schema.getNumParityUnits();
|
||||
@ -472,12 +487,10 @@ private int copy(ReaderStrategy strategy, int offset, int length) {
|
||||
*/
|
||||
@Override
|
||||
protected LocatedBlock getBlockAt(long blkStartOffset) throws IOException {
|
||||
LocatedBlock lb = super.getBlockAt(blkStartOffset);
|
||||
assert lb instanceof LocatedStripedBlock : "NameNode should return a " +
|
||||
"LocatedStripedBlock for a striped file";
|
||||
LocatedBlock lb = getBlockGroupAt(blkStartOffset);
|
||||
|
||||
int idx = (int) (((blkStartOffset - lb.getStartOffset()) / cellSize)
|
||||
% dataBlkNum);
|
||||
int idx = (int) ((blkStartOffset - lb.getStartOffset())
|
||||
% (dataBlkNum + parityBlkNum));
|
||||
// If indexing information is returned, iterate through the index array
|
||||
// to find the entry for position idx in the group
|
||||
LocatedStripedBlock lsb = (LocatedStripedBlock) lb;
|
||||
@ -509,48 +522,121 @@ protected void fetchBlockByteRange(long blockStartOffset, long start,
|
||||
long end, byte[] buf, int offset,
|
||||
Map<ExtendedBlock, Set<DatanodeInfo>> corruptedBlockMap)
|
||||
throws IOException {
|
||||
Map<Future<Void>, Integer> futures = new HashMap<>();
|
||||
CompletionService<Void> stripedReadsService =
|
||||
new ExecutorCompletionService<>(dfsClient.getStripedReadsThreadPool());
|
||||
int len = (int) (end - start + 1);
|
||||
|
||||
// Refresh the striped block group
|
||||
LocatedStripedBlock blockGroup = getBlockGroupAt(blockStartOffset);
|
||||
|
||||
AlignedStripe[] stripes = divideByteRangeIntoStripes(schema, blockGroup,
|
||||
start, end, buf, offset);
|
||||
for (AlignedStripe stripe : stripes) {
|
||||
fetchOneStripe(blockGroup, buf, stripe, corruptedBlockMap);
|
||||
}
|
||||
}
|
||||
|
||||
// Planning the portion of I/O for each shard
|
||||
ReadPortion[] readPortions = planReadPortions(dataBlkNum, cellSize, start,
|
||||
len, offset);
|
||||
|
||||
private void fetchOneStripe(LocatedStripedBlock blockGroup,
|
||||
byte[] buf, AlignedStripe alignedStripe, Map<ExtendedBlock,
|
||||
Set<DatanodeInfo>> corruptedBlockMap) throws IOException {
|
||||
Map<Future<Void>, Integer> futures = new HashMap<>();
|
||||
CompletionService<Void> service =
|
||||
new ExecutorCompletionService<>(dfsClient.getStripedReadsThreadPool());
|
||||
if (alignedStripe.getSpanInBlock() == 0) {
|
||||
DFSClient.LOG.warn("Trying to read an empty stripe from" + blockGroup);
|
||||
return;
|
||||
}
|
||||
// Parse group to get chosen DN location
|
||||
LocatedBlock[] blks = StripedBlockUtil.
|
||||
parseStripedBlockGroup(blockGroup, cellSize, dataBlkNum, parityBlkNum);
|
||||
|
||||
for (short i = 0; i < dataBlkNum; i++) {
|
||||
ReadPortion rp = readPortions[i];
|
||||
if (rp.getReadLength() <= 0) {
|
||||
continue;
|
||||
if (alignedStripe.chunks[i] != null
|
||||
&& alignedStripe.chunks[i].state != StripingChunk.ALLZERO) {
|
||||
fetchOneStripingChunk(futures, service, blks[i], alignedStripe, i,
|
||||
corruptedBlockMap);
|
||||
}
|
||||
DatanodeInfo loc = blks[i].getLocations()[0];
|
||||
StorageType type = blks[i].getStorageTypes()[0];
|
||||
DNAddrPair dnAddr = new DNAddrPair(loc, NetUtils.createSocketAddr(
|
||||
loc.getXferAddr(dfsClient.getConf().isConnectToDnViaHostname())),
|
||||
type);
|
||||
Callable<Void> readCallable = getFromOneDataNode(dnAddr,
|
||||
blks[i].getStartOffset(), rp.getStartOffsetInBlock(),
|
||||
rp.getStartOffsetInBlock() + rp.getReadLength() - 1, buf,
|
||||
rp.getOffsets(), rp.getLengths(), corruptedBlockMap, i);
|
||||
Future<Void> getFromDNRequest = stripedReadsService.submit(readCallable);
|
||||
DFSClient.LOG.debug("Submitting striped read request for " + blks[i]);
|
||||
futures.put(getFromDNRequest, (int) i);
|
||||
}
|
||||
// Input buffers for potential decode operation, which remains null until
|
||||
// first read failure
|
||||
byte[][] decodeInputs = null;
|
||||
while (!futures.isEmpty()) {
|
||||
try {
|
||||
waitNextCompletion(stripedReadsService, futures);
|
||||
StripingChunkReadResult r = getNextCompletedStripedRead(
|
||||
service, futures, 0);
|
||||
if (DFSClient.LOG.isDebugEnabled()) {
|
||||
DFSClient.LOG.debug("Read task returned: " + r + ", for stripe " + alignedStripe);
|
||||
}
|
||||
StripingChunk returnedChunk = alignedStripe.chunks[r.index];
|
||||
Preconditions.checkNotNull(returnedChunk);
|
||||
Preconditions.checkState(returnedChunk.state == StripingChunk.PENDING);
|
||||
if (r.state == StripingChunkReadResult.SUCCESSFUL) {
|
||||
returnedChunk.state = StripingChunk.FETCHED;
|
||||
alignedStripe.fetchedChunksNum++;
|
||||
if (alignedStripe.fetchedChunksNum == dataBlkNum) {
|
||||
clearFutures(futures.keySet());
|
||||
break;
|
||||
}
|
||||
} else {
|
||||
returnedChunk.state = StripingChunk.MISSING;
|
||||
alignedStripe.missingChunksNum++;
|
||||
if (alignedStripe.missingChunksNum > parityBlkNum) {
|
||||
clearFutures(futures.keySet());
|
||||
throw new IOException("Too many blocks are missing: " + alignedStripe);
|
||||
}
|
||||
// When seeing first missing block, initialize decode input buffers
|
||||
if (decodeInputs == null) {
|
||||
decodeInputs = initDecodeInputs(alignedStripe, dataBlkNum, parityBlkNum);
|
||||
}
|
||||
for (int i = 0; i < alignedStripe.chunks.length; i++) {
|
||||
StripingChunk chunk = alignedStripe.chunks[i];
|
||||
Preconditions.checkNotNull(chunk);
|
||||
if (chunk.state == StripingChunk.REQUESTED && i <= dataBlkNum) {
|
||||
fetchOneStripingChunk(futures, service, blks[i], alignedStripe, i,
|
||||
corruptedBlockMap);
|
||||
}
|
||||
}
|
||||
}
|
||||
} catch (InterruptedException ie) {
|
||||
// Ignore and retry
|
||||
String err = "Read request interrupted";
|
||||
DFSClient.LOG.error(err);
|
||||
clearFutures(futures.keySet());
|
||||
// Don't decode if read interrupted
|
||||
throw new InterruptedIOException(err);
|
||||
}
|
||||
}
|
||||
|
||||
if (alignedStripe.missingChunksNum > 0) {
|
||||
decodeAndFillBuffer(decodeInputs, buf, alignedStripe,
|
||||
dataBlkNum, parityBlkNum);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Schedule a single read request to an internal block
|
||||
* @param block The internal block
|
||||
* @param index Index of the internal block in the group
|
||||
* @param corruptedBlockMap Map of corrupted blocks
|
||||
*/
|
||||
private void fetchOneStripingChunk(Map<Future<Void>, Integer> futures,
|
||||
final CompletionService<Void> service, final LocatedBlock block,
|
||||
final AlignedStripe alignedStripe, final int index,
|
||||
Map<ExtendedBlock, Set<DatanodeInfo>> corruptedBlockMap) {
|
||||
DatanodeInfo loc = block.getLocations()[0];
|
||||
StorageType type = block.getStorageTypes()[0];
|
||||
DNAddrPair dnAddr = new DNAddrPair(loc, NetUtils.createSocketAddr(
|
||||
loc.getXferAddr(dfsClient.getConf().isConnectToDnViaHostname())),
|
||||
type);
|
||||
StripingChunk chunk = alignedStripe.chunks[index];
|
||||
chunk.state = StripingChunk.PENDING;
|
||||
Callable<Void> readCallable = getFromOneDataNode(dnAddr,
|
||||
block.getStartOffset(), alignedStripe.getOffsetInBlock(),
|
||||
alignedStripe.getOffsetInBlock() + alignedStripe.getSpanInBlock() - 1, chunk.buf,
|
||||
chunk.getOffsets(), chunk.getLengths(),
|
||||
corruptedBlockMap, index);
|
||||
Future<Void> getFromDNRequest = service.submit(readCallable);
|
||||
if (DFSClient.LOG.isDebugEnabled()) {
|
||||
DFSClient.LOG.debug("Submitting striped read request for " + index +
|
||||
". Info of the block: " + block + ", offset in block is " +
|
||||
alignedStripe.getOffsetInBlock() + ", end is " +
|
||||
(alignedStripe.getOffsetInBlock() + alignedStripe.getSpanInBlock() - 1));
|
||||
}
|
||||
futures.put(getFromDNRequest, index);
|
||||
}
|
||||
|
||||
private Callable<Void> getFromOneDataNode(final DNAddrPair datanode,
|
||||
@ -609,4 +695,12 @@ public synchronized void releaseBuffer(ByteBuffer buffer) {
|
||||
throw new UnsupportedOperationException(
|
||||
"Not support enhanced byte buffer access.");
|
||||
}
|
||||
|
||||
/** A variation to {@link DFSInputStream#cancelAll} */
|
||||
private void clearFutures(Collection<Future<Void>> futures) {
|
||||
for (Future<Void> future : futures) {
|
||||
future.cancel(false);
|
||||
}
|
||||
futures.clear();
|
||||
}
|
||||
}
|
||||
|
@ -67,7 +67,7 @@
|
||||
import org.apache.hadoop.hdfs.server.datanode.DataNode;
|
||||
import org.apache.hadoop.hdfs.server.protocol.BlockECRecoveryCommand.BlockECRecoveryInfo;
|
||||
import org.apache.hadoop.hdfs.util.StripedBlockUtil;
|
||||
import org.apache.hadoop.hdfs.util.StripedBlockUtil.StripedReadResult;
|
||||
import org.apache.hadoop.hdfs.util.StripedBlockUtil.StripingChunkReadResult;
|
||||
import org.apache.hadoop.io.IOUtils;
|
||||
import org.apache.hadoop.io.erasurecode.ECSchema;
|
||||
import org.apache.hadoop.io.erasurecode.rawcoder.RSRawDecoder;
|
||||
@ -462,10 +462,10 @@ private int readMinimumStripedData4Recovery(int[] success) {
|
||||
int nsuccess = 0;
|
||||
while (!futures.isEmpty()) {
|
||||
try {
|
||||
StripedReadResult result =
|
||||
StripingChunkReadResult result =
|
||||
StripedBlockUtil.getNextCompletedStripedRead(
|
||||
readService, futures, STRIPED_READ_THRESHOLD_MILLIS);
|
||||
if (result.state == StripedReadResult.SUCCESSFUL) {
|
||||
if (result.state == StripingChunkReadResult.SUCCESSFUL) {
|
||||
success[nsuccess++] = result.index;
|
||||
if (nsuccess >= dataBlkNum) {
|
||||
// cancel remaining reads if we read successfully from minimum
|
||||
@ -474,14 +474,14 @@ private int readMinimumStripedData4Recovery(int[] success) {
|
||||
futures.clear();
|
||||
break;
|
||||
}
|
||||
} else if (result.state == StripedReadResult.FAILED) {
|
||||
} else if (result.state == StripingChunkReadResult.FAILED) {
|
||||
// If read failed for some source, we should not use it anymore
|
||||
// and schedule read from a new source.
|
||||
StripedReader failedReader = stripedReaders.get(result.index);
|
||||
closeBlockReader(failedReader.blockReader);
|
||||
failedReader.blockReader = null;
|
||||
scheduleNewRead(used);
|
||||
} else if (result.state == StripedReadResult.TIMEOUT) {
|
||||
} else if (result.state == StripingChunkReadResult.TIMEOUT) {
|
||||
// If timeout, we also schedule a new read.
|
||||
scheduleNewRead(used);
|
||||
}
|
||||
|
@ -22,16 +22,18 @@
|
||||
|
||||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
import org.apache.hadoop.fs.StorageType;
|
||||
import org.apache.hadoop.hdfs.DFSClient;
|
||||
import org.apache.hadoop.hdfs.DFSStripedOutputStream;
|
||||
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
||||
import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
|
||||
import org.apache.hadoop.hdfs.protocol.LocatedBlock;
|
||||
import org.apache.hadoop.hdfs.protocol.LocatedStripedBlock;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
import org.apache.hadoop.io.erasurecode.ECSchema;
|
||||
import org.apache.hadoop.io.erasurecode.rawcoder.RSRawDecoder;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.*;
|
||||
import java.util.concurrent.CancellationException;
|
||||
import java.util.concurrent.CompletionService;
|
||||
import java.util.concurrent.ExecutionException;
|
||||
@ -85,7 +87,7 @@ public static LocatedBlock constructInternalBlock(LocatedStripedBlock bg,
|
||||
new DatanodeInfo[]{bg.getLocations()[idxInReturnedLocs]},
|
||||
new String[]{bg.getStorageIDs()[idxInReturnedLocs]},
|
||||
new StorageType[]{bg.getStorageTypes()[idxInReturnedLocs]},
|
||||
bg.getStartOffset() + idxInBlockGroup * cellSize, bg.isCorrupt(),
|
||||
bg.getStartOffset() + idxInBlockGroup, bg.isCorrupt(),
|
||||
null);
|
||||
}
|
||||
|
||||
@ -238,33 +240,37 @@ public static ReadPortion[] planReadPortions(final int dataBlkNum,
|
||||
/**
|
||||
* Get the next completed striped read task
|
||||
*
|
||||
* @return {@link StripedReadResult} indicating the status of the read task
|
||||
* @return {@link StripingChunkReadResult} indicating the status of the read task
|
||||
* succeeded, and the block index of the task. If the method times
|
||||
* out without getting any completed read tasks, -1 is returned as
|
||||
* block index.
|
||||
* @throws InterruptedException
|
||||
*/
|
||||
public static StripedReadResult getNextCompletedStripedRead(
|
||||
public static StripingChunkReadResult getNextCompletedStripedRead(
|
||||
CompletionService<Void> readService, Map<Future<Void>, Integer> futures,
|
||||
final long threshold) throws InterruptedException {
|
||||
Preconditions.checkArgument(!futures.isEmpty());
|
||||
Preconditions.checkArgument(threshold > 0);
|
||||
Future<Void> future = null;
|
||||
try {
|
||||
future = readService.poll(threshold, TimeUnit.MILLISECONDS);
|
||||
if (threshold > 0) {
|
||||
future = readService.poll(threshold, TimeUnit.MILLISECONDS);
|
||||
} else {
|
||||
future = readService.take();
|
||||
}
|
||||
if (future != null) {
|
||||
future.get();
|
||||
return new StripedReadResult(futures.remove(future),
|
||||
StripedReadResult.SUCCESSFUL);
|
||||
return new StripingChunkReadResult(futures.remove(future),
|
||||
StripingChunkReadResult.SUCCESSFUL);
|
||||
} else {
|
||||
return new StripedReadResult(StripedReadResult.TIMEOUT);
|
||||
return new StripingChunkReadResult(StripingChunkReadResult.TIMEOUT);
|
||||
}
|
||||
} catch (ExecutionException e) {
|
||||
return new StripedReadResult(futures.remove(future),
|
||||
StripedReadResult.FAILED);
|
||||
DFSClient.LOG.error("ExecutionException " + e);
|
||||
return new StripingChunkReadResult(futures.remove(future),
|
||||
StripingChunkReadResult.FAILED);
|
||||
} catch (CancellationException e) {
|
||||
return new StripedReadResult(futures.remove(future),
|
||||
StripedReadResult.CANCELLED);
|
||||
return new StripingChunkReadResult(futures.remove(future),
|
||||
StripingChunkReadResult.CANCELLED);
|
||||
}
|
||||
}
|
||||
|
||||
@ -291,26 +297,247 @@ public static long spaceConsumedByStripedBlock(long numDataBlkBytes,
|
||||
}
|
||||
|
||||
/**
|
||||
* This class represents the portion of I/O associated with each block in the
|
||||
* striped block group.
|
||||
* Initialize the decoding input buffers based on the chunk states in an
|
||||
* AlignedStripe
|
||||
*/
|
||||
public static class ReadPortion {
|
||||
public static byte[][] initDecodeInputs(AlignedStripe alignedStripe,
|
||||
int dataBlkNum, int parityBlkNum) {
|
||||
byte[][] decodeInputs =
|
||||
new byte[dataBlkNum + parityBlkNum][(int) alignedStripe.getSpanInBlock()];
|
||||
for (int i = 0; i < alignedStripe.chunks.length; i++) {
|
||||
StripingChunk chunk = alignedStripe.chunks[i];
|
||||
if (chunk == null) {
|
||||
alignedStripe.chunks[i] = new StripingChunk(decodeInputs[i]);
|
||||
alignedStripe.chunks[i].offsetsInBuf.add(0);
|
||||
alignedStripe.chunks[i].lengthsInBuf.add((int) alignedStripe.getSpanInBlock());
|
||||
} else if (chunk.state == StripingChunk.FETCHED) {
|
||||
int posInBuf = 0;
|
||||
for (int j = 0; j < chunk.offsetsInBuf.size(); j++) {
|
||||
System.arraycopy(chunk.buf, chunk.offsetsInBuf.get(j),
|
||||
decodeInputs[i], posInBuf, chunk.lengthsInBuf.get(j));
|
||||
posInBuf += chunk.lengthsInBuf.get(j);
|
||||
}
|
||||
} else if (chunk.state == StripingChunk.ALLZERO) {
|
||||
Arrays.fill(decodeInputs[i], (byte)0);
|
||||
}
|
||||
}
|
||||
return decodeInputs;
|
||||
}
|
||||
|
||||
/**
|
||||
* Decode based on the given input buffers and schema
|
||||
*/
|
||||
public static void decodeAndFillBuffer(final byte[][] decodeInputs, byte[] buf,
|
||||
AlignedStripe alignedStripe, int dataBlkNum, int parityBlkNum) {
|
||||
int[] decodeIndices = new int[parityBlkNum];
|
||||
int pos = 0;
|
||||
for (int i = 0; i < alignedStripe.chunks.length; i++) {
|
||||
if (alignedStripe.chunks[i].state != StripingChunk.FETCHED &&
|
||||
alignedStripe.chunks[i].state != StripingChunk.ALLZERO) {
|
||||
decodeIndices[pos++] = i;
|
||||
}
|
||||
}
|
||||
|
||||
byte[][] outputs = new byte[parityBlkNum][(int) alignedStripe.getSpanInBlock()];
|
||||
RSRawDecoder rsRawDecoder = new RSRawDecoder();
|
||||
rsRawDecoder.initialize(dataBlkNum, parityBlkNum, (int) alignedStripe.getSpanInBlock());
|
||||
rsRawDecoder.decode(decodeInputs, decodeIndices, outputs);
|
||||
|
||||
for (int i = 0; i < dataBlkNum + parityBlkNum; i++) {
|
||||
StripingChunk chunk = alignedStripe.chunks[i];
|
||||
if (chunk.state == StripingChunk.MISSING) {
|
||||
int srcPos = 0;
|
||||
for (int j = 0; j < chunk.offsetsInBuf.size(); j++) {
|
||||
//TODO: workaround (filling fixed bytes), to remove after HADOOP-11938
|
||||
// System.arraycopy(outputs[i], srcPos, buf, chunk.offsetsInBuf.get(j),
|
||||
// chunk.lengthsInBuf.get(j));
|
||||
Arrays.fill(buf, chunk.offsetsInBuf.get(j),
|
||||
chunk.offsetsInBuf.get(j) + chunk.lengthsInBuf.get(j), (byte)7);
|
||||
srcPos += chunk.lengthsInBuf.get(j);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* This method divides a requested byte range into an array of
|
||||
* {@link AlignedStripe}
|
||||
*
|
||||
*
|
||||
* At most 5 stripes will be generated from each logical range
|
||||
* TODO: cleanup and get rid of planReadPortions
|
||||
*/
|
||||
public static AlignedStripe[] divideByteRangeIntoStripes (
|
||||
ECSchema ecSchema, LocatedStripedBlock blockGroup, long start, long end,
|
||||
byte[] buf, int offsetInBuf) {
|
||||
// TODO: change ECSchema naming to use cell size instead of chunk size
|
||||
|
||||
// Step 0: analyze range and calculate basic parameters
|
||||
int cellSize = ecSchema.getChunkSize();
|
||||
int dataBlkNum = ecSchema.getNumDataUnits();
|
||||
int len = (int) (end - start + 1);
|
||||
int firstCellIdxInBG = (int) (start / cellSize);
|
||||
int lastCellIdxInBG = (int) (end / cellSize);
|
||||
int firstCellSize = Math.min(cellSize - (int) (start % cellSize), len);
|
||||
long firstCellOffsetInBlk = start % cellSize;
|
||||
int lastCellSize = lastCellIdxInBG == firstCellIdxInBG ?
|
||||
firstCellSize : (int) (end % cellSize) + 1;
|
||||
|
||||
// Step 1: get the unmerged ranges on each internal block
|
||||
// TODO: StripingCell should carry info on size and start offset (HDFS-8320)
|
||||
VerticalRange[] ranges = getRangesForInternalBlocks(ecSchema,
|
||||
firstCellIdxInBG, lastCellIdxInBG, firstCellSize, firstCellOffsetInBlk,
|
||||
lastCellSize);
|
||||
|
||||
// Step 2: merge into at most 5 stripes
|
||||
AlignedStripe[] stripes = mergeRangesForInternalBlocks(ecSchema, ranges);
|
||||
|
||||
// Step 3: calculate each chunk's position in destination buffer
|
||||
calcualteChunkPositionsInBuf(ecSchema, blockGroup, buf, offsetInBuf,
|
||||
firstCellIdxInBG, lastCellIdxInBG, firstCellSize, firstCellOffsetInBlk,
|
||||
lastCellSize, stripes);
|
||||
|
||||
// Step 4: prepare ALLZERO blocks
|
||||
prepareAllZeroChunks(blockGroup, buf, stripes, cellSize, dataBlkNum);
|
||||
|
||||
return stripes;
|
||||
}
|
||||
|
||||
private static VerticalRange[] getRangesForInternalBlocks (ECSchema ecSchema,
|
||||
int firstCellIdxInBG, int lastCellIdxInBG, int firstCellSize,
|
||||
long firstCellOffsetInBlk, int lastCellSize) {
|
||||
int cellSize = ecSchema.getChunkSize();
|
||||
int dataBlkNum = ecSchema.getNumDataUnits();
|
||||
|
||||
StripingCell firstCell = new StripingCell(ecSchema, firstCellIdxInBG);
|
||||
StripingCell lastCell = new StripingCell(ecSchema, lastCellIdxInBG);
|
||||
|
||||
VerticalRange ranges[] = new VerticalRange[dataBlkNum];
|
||||
ranges[firstCell.idxInStripe] =
|
||||
new VerticalRange(firstCellOffsetInBlk, firstCellSize);
|
||||
for (int i = firstCellIdxInBG + 1; i < lastCellIdxInBG; i++) {
|
||||
// iterate through all cells and update the list of StripeRanges
|
||||
StripingCell cell = new StripingCell(ecSchema, i);
|
||||
if (ranges[cell.idxInStripe] == null) {
|
||||
ranges[cell.idxInStripe] = new VerticalRange(
|
||||
cell.idxInInternalBlk * cellSize, cellSize);
|
||||
} else {
|
||||
ranges[cell.idxInStripe].spanInBlock += cellSize;
|
||||
}
|
||||
}
|
||||
if (ranges[lastCell.idxInStripe] == null) {
|
||||
ranges[lastCell.idxInStripe] = new VerticalRange(
|
||||
lastCell.idxInInternalBlk * cellSize, lastCellSize);
|
||||
} else if (lastCell.idxInBlkGroup != firstCell.idxInBlkGroup) {
|
||||
ranges[lastCell.idxInStripe].spanInBlock += lastCellSize;
|
||||
}
|
||||
|
||||
return ranges;
|
||||
}
|
||||
|
||||
private static AlignedStripe[] mergeRangesForInternalBlocks(ECSchema ecSchema,
|
||||
VerticalRange[] ranges) {
|
||||
int dataBlkNum = ecSchema.getNumDataUnits();
|
||||
int parityBlkNum = ecSchema.getNumParityUnits();
|
||||
List<AlignedStripe> stripes = new ArrayList<>();
|
||||
SortedSet<Long> stripePoints = new TreeSet<>();
|
||||
for (VerticalRange r : ranges) {
|
||||
if (r != null) {
|
||||
stripePoints.add(r.offsetInBlock);
|
||||
stripePoints.add(r.offsetInBlock + r.spanInBlock);
|
||||
}
|
||||
}
|
||||
|
||||
long prev = -1;
|
||||
for (long point : stripePoints) {
|
||||
if (prev >= 0) {
|
||||
stripes.add(new AlignedStripe(prev, point - prev,
|
||||
dataBlkNum + parityBlkNum));
|
||||
}
|
||||
prev = point;
|
||||
}
|
||||
return stripes.toArray(new AlignedStripe[stripes.size()]);
|
||||
}
|
||||
|
||||
private static void calcualteChunkPositionsInBuf(ECSchema ecSchema,
|
||||
LocatedStripedBlock blockGroup, byte[] buf, int offsetInBuf,
|
||||
int firstCellIdxInBG, int lastCellIdxInBG, int firstCellSize,
|
||||
long firstCellOffsetInBlk, int lastCellSize, AlignedStripe[] stripes) {
|
||||
int cellSize = ecSchema.getChunkSize();
|
||||
int dataBlkNum = ecSchema.getNumDataUnits();
|
||||
// Step 3: calculate each chunk's position in destination buffer
|
||||
/**
|
||||
* startOffsetInBlock
|
||||
* |
|
||||
* v
|
||||
* |<-lengths[0]->|<- lengths[1] ->|<-lengths[2]->|
|
||||
* | <--------------- AlignedStripe --------------->|
|
||||
*
|
||||
* |<- length_0 ->|<-- length_1 -->|<- length_2 ->|
|
||||
* +------------------+------------------+----------------+
|
||||
* | cell_0 | cell_3 | cell_6 | <- blk_0
|
||||
* | cell_0_0_0 | cell_3_1_0 | cell_6_2_0 | <- blk_0
|
||||
* +------------------+------------------+----------------+
|
||||
* _/ \_______________________
|
||||
* | |
|
||||
* v offsetsInBuf[0] v offsetsInBuf[1]
|
||||
* +------------------------------------------------------+
|
||||
* | cell_0 | cell_1 and cell_2 |cell_3 ...| <- buf
|
||||
* | (partial) | (from blk_1 and blk_2) | |
|
||||
* +------------------------------------------------------+
|
||||
* v offset_0 v offset_1
|
||||
* +----------------------------------------------------------+
|
||||
* | cell_0_0_0 | cell_1_0_1 and cell_2_0_2 |cell_3_1_0 ...| <- buf
|
||||
* | (partial) | (from blk_1 and blk_2) | |
|
||||
* +----------------------------------------------------------+
|
||||
*
|
||||
* Cell indexing convention defined in {@link StripingCell}
|
||||
*/
|
||||
int done = 0;
|
||||
for (int i = firstCellIdxInBG; i <= lastCellIdxInBG; i++) {
|
||||
StripingCell cell = new StripingCell(ecSchema, i);
|
||||
long cellStart = i == firstCellIdxInBG ?
|
||||
firstCellOffsetInBlk : cell.idxInInternalBlk * cellSize;
|
||||
int cellLen;
|
||||
if (i == firstCellIdxInBG) {
|
||||
cellLen = firstCellSize;
|
||||
} else if (i == lastCellIdxInBG) {
|
||||
cellLen = lastCellSize;
|
||||
} else {
|
||||
cellLen = cellSize;
|
||||
}
|
||||
long cellEnd = cellStart + cellLen - 1;
|
||||
for (AlignedStripe s : stripes) {
|
||||
long stripeEnd = s.getOffsetInBlock() + s.getSpanInBlock() - 1;
|
||||
long overlapStart = Math.max(cellStart, s.getOffsetInBlock());
|
||||
long overlapEnd = Math.min(cellEnd, stripeEnd);
|
||||
int overLapLen = (int) (overlapEnd - overlapStart + 1);
|
||||
if (overLapLen <= 0) {
|
||||
continue;
|
||||
}
|
||||
if (s.chunks[cell.idxInStripe] == null) {
|
||||
s.chunks[cell.idxInStripe] = new StripingChunk(buf);
|
||||
}
|
||||
|
||||
s.chunks[cell.idxInStripe].offsetsInBuf.
|
||||
add((int)(offsetInBuf + done + overlapStart - cellStart));
|
||||
s.chunks[cell.idxInStripe].lengthsInBuf.add(overLapLen);
|
||||
}
|
||||
done += cellLen;
|
||||
}
|
||||
}
|
||||
|
||||
private static void prepareAllZeroChunks(LocatedStripedBlock blockGroup,
|
||||
byte[] buf, AlignedStripe[] stripes, int cellSize, int dataBlkNum) {
|
||||
for (AlignedStripe s : stripes) {
|
||||
for (int i = 0; i < dataBlkNum; i++) {
|
||||
long internalBlkLen = getInternalBlockLength(blockGroup.getBlockSize(),
|
||||
cellSize, dataBlkNum, i);
|
||||
if (internalBlkLen <= s.getOffsetInBlock()) {
|
||||
Preconditions.checkState(s.chunks[i] == null);
|
||||
s.chunks[i] = new StripingChunk(buf);
|
||||
s.chunks[i].state = StripingChunk.ALLZERO;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* This class represents the portion of I/O associated with each block in the
|
||||
* striped block group.
|
||||
* TODO: consolidate ReadPortion with AlignedStripe
|
||||
*/
|
||||
public static class ReadPortion {
|
||||
private long startOffsetInBlock = 0;
|
||||
private int readLength = 0;
|
||||
public final List<Integer> offsetsInBuf = new ArrayList<>();
|
||||
@ -349,12 +576,235 @@ void addReadLength(int extraLength) {
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* The unit of encoding used in {@link DFSStripedOutputStream}
|
||||
* | <------- Striped Block Group -------> |
|
||||
* blk_0 blk_1 blk_2
|
||||
* | | |
|
||||
* v v v
|
||||
* +----------+ +----------+ +----------+
|
||||
* |cell_0_0_0| |cell_1_0_1| |cell_2_0_2|
|
||||
* +----------+ +----------+ +----------+
|
||||
* |cell_3_1_0| |cell_4_1_1| |cell_5_1_2| <- {@link idxInBlkGroup} = 5
|
||||
* +----------+ +----------+ +----------+ {@link idxInInternalBlk} = 1
|
||||
* {@link idxInStripe} = 2
|
||||
* A StripingCell is a special instance of {@link StripingChunk} whose offset
|
||||
* and size align with the cell used when writing data.
|
||||
* TODO: consider parity cells
|
||||
*/
|
||||
public static class StripingCell {
|
||||
public final ECSchema schema;
|
||||
/** Logical order in a block group, used when doing I/O to a block group */
|
||||
public final int idxInBlkGroup;
|
||||
public final int idxInInternalBlk;
|
||||
public final int idxInStripe;
|
||||
|
||||
public StripingCell(ECSchema ecSchema, int idxInBlkGroup) {
|
||||
this.schema = ecSchema;
|
||||
this.idxInBlkGroup = idxInBlkGroup;
|
||||
this.idxInInternalBlk = idxInBlkGroup / ecSchema.getNumDataUnits();
|
||||
this.idxInStripe = idxInBlkGroup -
|
||||
this.idxInInternalBlk * ecSchema.getNumDataUnits();
|
||||
}
|
||||
|
||||
public StripingCell(ECSchema ecSchema, int idxInInternalBlk,
|
||||
int idxInStripe) {
|
||||
this.schema = ecSchema;
|
||||
this.idxInInternalBlk = idxInInternalBlk;
|
||||
this.idxInStripe = idxInStripe;
|
||||
this.idxInBlkGroup =
|
||||
idxInInternalBlk * ecSchema.getNumDataUnits() + idxInStripe;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Given a requested byte range on a striped block group, an AlignedStripe
|
||||
* represents a {@link VerticalRange} that is aligned with both the byte range
|
||||
* and boundaries of all internal blocks. As illustrated in the diagram, any
|
||||
* given byte range on a block group leads to 1~5 AlignedStripe's.
|
||||
*
|
||||
* |<-------- Striped Block Group -------->|
|
||||
* blk_0 blk_1 blk_2 blk_3 blk_4
|
||||
* +----+ | +----+ +----+
|
||||
* |full| | | | | | <- AlignedStripe0:
|
||||
* +----+ |~~~~| | |~~~~| |~~~~| 1st cell is partial
|
||||
* |part| | | | | | | | <- AlignedStripe1: byte range
|
||||
* +----+ +----+ +----+ | |~~~~| |~~~~| doesn't start at 1st block
|
||||
* |full| |full| |full| | | | | |
|
||||
* |cell| |cell| |cell| | | | | | <- AlignedStripe2 (full stripe)
|
||||
* | | | | | | | | | | |
|
||||
* +----+ +----+ +----+ | |~~~~| |~~~~|
|
||||
* |full| |part| | | | | | <- AlignedStripe3: byte range
|
||||
* |~~~~| +----+ | |~~~~| |~~~~| doesn't end at last block
|
||||
* | | | | | | | <- AlignedStripe4:
|
||||
* +----+ | +----+ +----+ last cell is partial
|
||||
* |
|
||||
* <---- data blocks ----> | <--- parity --->
|
||||
*
|
||||
* An AlignedStripe is the basic unit of reading from a striped block group,
|
||||
* because within the AlignedStripe, all internal blocks can be processed in
|
||||
* a uniform manner.
|
||||
*
|
||||
* The coverage of an AlignedStripe on an internal block is represented as a
|
||||
* {@link StripingChunk}.
|
||||
* To simplify the logic of reading a logical byte range from a block group,
|
||||
* a StripingChunk is either completely in the requested byte range or
|
||||
* completely outside the requested byte range.
|
||||
*/
|
||||
public static class AlignedStripe {
|
||||
public VerticalRange range;
|
||||
/** status of each chunk in the stripe */
|
||||
public final StripingChunk[] chunks;
|
||||
public int fetchedChunksNum = 0;
|
||||
public int missingChunksNum = 0;
|
||||
|
||||
public AlignedStripe(long offsetInBlock, long length, int width) {
|
||||
Preconditions.checkArgument(offsetInBlock >= 0 && length >= 0);
|
||||
this.range = new VerticalRange(offsetInBlock, length);
|
||||
this.chunks = new StripingChunk[width];
|
||||
}
|
||||
|
||||
public AlignedStripe(VerticalRange range, int width) {
|
||||
this.range = range;
|
||||
this.chunks = new StripingChunk[width];
|
||||
}
|
||||
|
||||
public boolean include(long pos) {
|
||||
return range.include(pos);
|
||||
}
|
||||
|
||||
public long getOffsetInBlock() {
|
||||
return range.offsetInBlock;
|
||||
}
|
||||
|
||||
public long getSpanInBlock() {
|
||||
return range.spanInBlock;
|
||||
}
|
||||
|
||||
@Override
|
||||
public String toString() {
|
||||
return "Offset=" + range.offsetInBlock + ", length=" + range.spanInBlock +
|
||||
", fetchedChunksNum=" + fetchedChunksNum +
|
||||
", missingChunksNum=" + missingChunksNum;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* A simple utility class representing an arbitrary vertical inclusive range
|
||||
* starting at {@link offsetInBlock} and lasting for {@link length} bytes in
|
||||
* an internal block. Note that VerticalRange doesn't necessarily align with
|
||||
* {@link StripingCell}.
|
||||
*
|
||||
* |<- Striped Block Group ->|
|
||||
* blk_0
|
||||
* |
|
||||
* v
|
||||
* +-----+
|
||||
* |~~~~~| <-- {@link offsetInBlock}
|
||||
* | | ^
|
||||
* | | |
|
||||
* | | | {@link spanInBlock}
|
||||
* | | v
|
||||
* |~~~~~| ---
|
||||
* | |
|
||||
* +-----+
|
||||
*/
|
||||
public static class VerticalRange {
|
||||
/** start offset in the block group (inclusive) */
|
||||
public long offsetInBlock;
|
||||
/** length of the stripe range */
|
||||
public long spanInBlock;
|
||||
|
||||
public VerticalRange(long offsetInBlock, long length) {
|
||||
Preconditions.checkArgument(offsetInBlock >= 0 && length >= 0);
|
||||
this.offsetInBlock = offsetInBlock;
|
||||
this.spanInBlock = length;
|
||||
}
|
||||
|
||||
/** whether a position is in the range */
|
||||
public boolean include(long pos) {
|
||||
return pos >= offsetInBlock && pos < offsetInBlock + spanInBlock;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Indicates the coverage of an {@link AlignedStripe} on an internal block,
|
||||
* and the state of the chunk in the context of the read request.
|
||||
*
|
||||
* |<---------------- Striped Block Group --------------->|
|
||||
* blk_0 blk_1 blk_2 blk_3 blk_4
|
||||
* +---------+ | +----+ +----+
|
||||
* null null |REQUESTED| | |null| |null| <- AlignedStripe0
|
||||
* +---------+ |---------| | |----| |----|
|
||||
* null |REQUESTED| |REQUESTED| | |null| |null| <- AlignedStripe1
|
||||
* +---------+ +---------+ +---------+ | +----+ +----+
|
||||
* |REQUESTED| |REQUESTED| ALLZERO | |null| |null| <- AlignedStripe2
|
||||
* +---------+ +---------+ | +----+ +----+
|
||||
* <----------- data blocks ------------> | <--- parity --->
|
||||
*
|
||||
* The class also carries {@link buf}, {@link offsetsInBuf}, and
|
||||
* {@link lengthsInBuf} to define how read task for this chunk should deliver
|
||||
* the returned data.
|
||||
*/
|
||||
public static class StripingChunk {
|
||||
/** Chunk has been successfully fetched */
|
||||
public static final int FETCHED = 0x01;
|
||||
/** Chunk has encountered failed when being fetched */
|
||||
public static final int MISSING = 0x02;
|
||||
/** Chunk being fetched (fetching task is in-flight) */
|
||||
public static final int PENDING = 0x04;
|
||||
/**
|
||||
* Chunk is requested either by application or for decoding, need to
|
||||
* schedule read task
|
||||
*/
|
||||
public static final int REQUESTED = 0X08;
|
||||
/**
|
||||
* Internal block is short and has no overlap with chunk. Chunk considered
|
||||
* all-zero bytes in codec calculations.
|
||||
*/
|
||||
public static final int ALLZERO = 0X0f;
|
||||
|
||||
/**
|
||||
* If a chunk is completely in requested range, the state transition is:
|
||||
* REQUESTED (when AlignedStripe created) -> PENDING -> {FETCHED | MISSING}
|
||||
* If a chunk is completely outside requested range (including parity
|
||||
* chunks), state transition is:
|
||||
* null (AlignedStripe created) -> REQUESTED (upon failure) -> PENDING ...
|
||||
*/
|
||||
public int state = REQUESTED;
|
||||
public byte[] buf;
|
||||
public List<Integer> offsetsInBuf;
|
||||
public List<Integer> lengthsInBuf;
|
||||
|
||||
public StripingChunk(byte[] buf) {
|
||||
this.buf = buf;
|
||||
this.offsetsInBuf = new ArrayList<>();
|
||||
this.lengthsInBuf = new ArrayList<>();
|
||||
}
|
||||
|
||||
public int[] getOffsets() {
|
||||
int[] offsets = new int[offsetsInBuf.size()];
|
||||
for (int i = 0; i < offsets.length; i++) {
|
||||
offsets[i] = offsetsInBuf.get(i);
|
||||
}
|
||||
return offsets;
|
||||
}
|
||||
|
||||
public int[] getLengths() {
|
||||
int[] lens = new int[this.lengthsInBuf.size()];
|
||||
for (int i = 0; i < lens.length; i++) {
|
||||
lens[i] = this.lengthsInBuf.get(i);
|
||||
}
|
||||
return lens;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* This class represents result from a striped read request.
|
||||
* If the task was successful or the internal computation failed,
|
||||
* an index is also returned.
|
||||
*/
|
||||
public static class StripedReadResult {
|
||||
public static class StripingChunkReadResult {
|
||||
public static final int SUCCESSFUL = 0x01;
|
||||
public static final int FAILED = 0x02;
|
||||
public static final int TIMEOUT = 0x04;
|
||||
@ -363,18 +813,23 @@ public static class StripedReadResult {
|
||||
public final int index;
|
||||
public final int state;
|
||||
|
||||
public StripedReadResult(int state) {
|
||||
public StripingChunkReadResult(int state) {
|
||||
Preconditions.checkArgument(state == TIMEOUT,
|
||||
"Only timeout result should return negative index.");
|
||||
this.index = -1;
|
||||
this.state = state;
|
||||
}
|
||||
|
||||
public StripedReadResult(int index, int state) {
|
||||
public StripingChunkReadResult(int index, int state) {
|
||||
Preconditions.checkArgument(state != TIMEOUT,
|
||||
"Timeout result should return negative index.");
|
||||
this.index = index;
|
||||
this.state = state;
|
||||
}
|
||||
|
||||
@Override
|
||||
public String toString() {
|
||||
return "(index=" + index + ", state =" + state + ")";
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -36,6 +36,7 @@
|
||||
import org.apache.hadoop.hdfs.server.namenode.ErasureCodingSchemaManager;
|
||||
import org.apache.hadoop.hdfs.util.StripedBlockUtil;
|
||||
import org.apache.hadoop.io.erasurecode.ECSchema;
|
||||
import org.apache.hadoop.io.erasurecode.rawcoder.RSRawDecoder;
|
||||
import org.junit.After;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
@ -133,8 +134,102 @@ public void testPread() throws Exception {
|
||||
byte[] readBuffer = new byte[readSize];
|
||||
int ret = in.read(0, readBuffer, 0, readSize);
|
||||
|
||||
byte[] expected = new byte[readSize];
|
||||
/** A variation of {@link DFSTestUtil#fillExpectedBuf} for striped blocks */
|
||||
for (int i = 0; i < NUM_STRIPE_PER_BLOCK; i++) {
|
||||
for (int j = 0; j < DATA_BLK_NUM; j++) {
|
||||
for (int k = 0; k < CELLSIZE; k++) {
|
||||
int posInBlk = i * CELLSIZE + k;
|
||||
int posInFile = i * CELLSIZE * DATA_BLK_NUM + j * CELLSIZE + k;
|
||||
expected[posInFile] = SimulatedFSDataset.simulatedByte(
|
||||
new Block(bg.getBlock().getBlockId() + j), posInBlk);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
assertEquals(readSize, ret);
|
||||
// TODO: verify read results with patterned data from HDFS-8117
|
||||
assertArrayEquals(expected, readBuffer);
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testPreadWithDNFailure() throws Exception {
|
||||
final int numBlocks = 4;
|
||||
final int failedDNIdx = 2;
|
||||
DFSTestUtil.createStripedFile(cluster, filePath, null, numBlocks,
|
||||
NUM_STRIPE_PER_BLOCK, false);
|
||||
LocatedBlocks lbs = fs.getClient().namenode.getBlockLocations(
|
||||
filePath.toString(), 0, BLOCK_GROUP_SIZE);
|
||||
|
||||
assert lbs.get(0) instanceof LocatedStripedBlock;
|
||||
LocatedStripedBlock bg = (LocatedStripedBlock)(lbs.get(0));
|
||||
for (int i = 0; i < DATA_BLK_NUM + PARITY_BLK_NUM; i++) {
|
||||
Block blk = new Block(bg.getBlock().getBlockId() + i,
|
||||
NUM_STRIPE_PER_BLOCK * CELLSIZE,
|
||||
bg.getBlock().getGenerationStamp());
|
||||
blk.setGenerationStamp(bg.getBlock().getGenerationStamp());
|
||||
cluster.injectBlocks(i, Arrays.asList(blk),
|
||||
bg.getBlock().getBlockPoolId());
|
||||
}
|
||||
DFSStripedInputStream in =
|
||||
new DFSStripedInputStream(fs.getClient(), filePath.toString(), false,
|
||||
ErasureCodingSchemaManager.getSystemDefaultSchema());
|
||||
int readSize = BLOCK_GROUP_SIZE;
|
||||
byte[] readBuffer = new byte[readSize];
|
||||
byte[] expected = new byte[readSize];
|
||||
cluster.stopDataNode(failedDNIdx);
|
||||
/** A variation of {@link DFSTestUtil#fillExpectedBuf} for striped blocks */
|
||||
for (int i = 0; i < NUM_STRIPE_PER_BLOCK; i++) {
|
||||
for (int j = 0; j < DATA_BLK_NUM; j++) {
|
||||
for (int k = 0; k < CELLSIZE; k++) {
|
||||
int posInBlk = i * CELLSIZE + k;
|
||||
int posInFile = i * CELLSIZE * DATA_BLK_NUM + j * CELLSIZE + k;
|
||||
expected[posInFile] = SimulatedFSDataset.simulatedByte(
|
||||
new Block(bg.getBlock().getBlockId() + j), posInBlk);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Update the expected content for decoded data
|
||||
for (int i = 0; i < NUM_STRIPE_PER_BLOCK; i++) {
|
||||
byte[][] decodeInputs = new byte[DATA_BLK_NUM + PARITY_BLK_NUM][CELLSIZE];
|
||||
int[] missingBlkIdx = new int[]{failedDNIdx, DATA_BLK_NUM+1, DATA_BLK_NUM+2};
|
||||
byte[][] decodeOutputs = new byte[PARITY_BLK_NUM][CELLSIZE];
|
||||
for (int j = 0; j < DATA_BLK_NUM; j++) {
|
||||
int posInBuf = i * CELLSIZE * DATA_BLK_NUM + j * CELLSIZE;
|
||||
if (j != failedDNIdx) {
|
||||
System.arraycopy(expected, posInBuf, decodeInputs[j], 0, CELLSIZE);
|
||||
}
|
||||
}
|
||||
for (int k = 0; k < CELLSIZE; k++) {
|
||||
int posInBlk = i * CELLSIZE + k;
|
||||
decodeInputs[DATA_BLK_NUM][k] = SimulatedFSDataset.simulatedByte(
|
||||
new Block(bg.getBlock().getBlockId() + DATA_BLK_NUM), posInBlk);
|
||||
}
|
||||
// RSRawDecoder rsRawDecoder = new RSRawDecoder();
|
||||
// rsRawDecoder.initialize(DATA_BLK_NUM, PARITY_BLK_NUM, CELLSIZE);
|
||||
// rsRawDecoder.decode(decodeInputs, missingBlkIdx, decodeOutputs);
|
||||
int posInBuf = i * CELLSIZE * DATA_BLK_NUM + failedDNIdx * CELLSIZE;
|
||||
// System.arraycopy(decodeOutputs[0], 0, expected, posInBuf, CELLSIZE);
|
||||
//TODO: workaround (filling fixed bytes), to remove after HADOOP-11938
|
||||
Arrays.fill(expected, posInBuf, posInBuf + CELLSIZE, (byte)7);
|
||||
}
|
||||
int delta = 10;
|
||||
int done = 0;
|
||||
// read a small delta, shouldn't trigger decode
|
||||
// |cell_0 |
|
||||
// |10 |
|
||||
done += in.read(0, readBuffer, 0, delta);
|
||||
assertEquals(delta, done);
|
||||
// both head and trail cells are partial
|
||||
// |c_0 |c_1 |c_2 |c_3 |c_4 |c_5 |
|
||||
// |256K - 10|missing|256K|256K|256K - 10|not in range|
|
||||
done += in.read(delta, readBuffer, delta,
|
||||
CELLSIZE * (DATA_BLK_NUM - 1) - 2 * delta);
|
||||
assertEquals(CELLSIZE * (DATA_BLK_NUM - 1) - delta, done);
|
||||
// read the rest
|
||||
done += in.read(done, readBuffer, done, readSize - done);
|
||||
assertEquals(readSize, done);
|
||||
assertArrayEquals(expected, readBuffer);
|
||||
}
|
||||
|
||||
@Test
|
||||
|
@ -18,10 +18,13 @@
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.BlockLocation;
|
||||
import org.apache.hadoop.fs.FSDataInputStream;
|
||||
import org.apache.hadoop.fs.FileStatus;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
||||
import org.apache.hadoop.hdfs.server.datanode.DataNode;
|
||||
import org.apache.hadoop.io.erasurecode.rawcoder.RSRawDecoder;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Assert;
|
||||
import org.junit.BeforeClass;
|
||||
@ -321,4 +324,50 @@ private void testOneFileUsingDFSStripedInputStream(String src, int writeBytes)
|
||||
Assert.assertArrayEquals(bytes, result.array());
|
||||
}
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testWritePreadWithDNFailure() throws IOException {
|
||||
final int failedDNIdx = 2;
|
||||
final int length = cellSize * (dataBlocks + 2);
|
||||
Path testPath = new Path("/foo");
|
||||
final byte[] bytes = generateBytes(length);
|
||||
DFSTestUtil.writeFile(fs, testPath, new String(bytes));
|
||||
|
||||
// shut down the DN that holds the last internal data block
|
||||
BlockLocation[] locs = fs.getFileBlockLocations(testPath, cellSize * 5,
|
||||
cellSize);
|
||||
String name = (locs[0].getNames())[failedDNIdx];
|
||||
for (DataNode dn : cluster.getDataNodes()) {
|
||||
int port = dn.getXferPort();
|
||||
if (name.contains(Integer.toString(port))) {
|
||||
dn.shutdown();
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
// pread
|
||||
int startOffsetInFile = cellSize * 5;
|
||||
try (FSDataInputStream fsdis = fs.open(testPath)) {
|
||||
byte[] buf = new byte[length];
|
||||
int readLen = fsdis.read(startOffsetInFile, buf, 0, buf.length);
|
||||
Assert.assertEquals("The length of file should be the same to write size",
|
||||
length - startOffsetInFile, readLen);
|
||||
|
||||
RSRawDecoder rsRawDecoder = new RSRawDecoder();
|
||||
rsRawDecoder.initialize(dataBlocks, parityBlocks, 1);
|
||||
byte[] expected = new byte[readLen];
|
||||
for (int i = startOffsetInFile; i < length; i++) {
|
||||
//TODO: workaround (filling fixed bytes), to remove after HADOOP-11938
|
||||
if ((i / cellSize) % dataBlocks == failedDNIdx) {
|
||||
expected[i - startOffsetInFile] = (byte)7;
|
||||
} else {
|
||||
expected[i - startOffsetInFile] = getByte(i);
|
||||
}
|
||||
}
|
||||
for (int i = startOffsetInFile; i < length; i++) {
|
||||
Assert.assertEquals("Byte at " + i + " should be the same",
|
||||
expected[i - startOffsetInFile], buf[i - startOffsetInFile]);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user