HDFS-12884. BlockUnderConstructionFeature.truncateBlock should be of type BlockInfo. Contributed by chencan.
This commit is contained in:
parent
5aa7052e31
commit
8d898ab25f
@ -60,7 +60,7 @@ public class BlockUnderConstructionFeature {
|
|||||||
/**
|
/**
|
||||||
* The block source to use in the event of copy-on-write truncate.
|
* The block source to use in the event of copy-on-write truncate.
|
||||||
*/
|
*/
|
||||||
private Block truncateBlock;
|
private BlockInfo truncateBlock;
|
||||||
|
|
||||||
public BlockUnderConstructionFeature(Block blk,
|
public BlockUnderConstructionFeature(Block blk,
|
||||||
BlockUCState state, DatanodeStorageInfo[] targets, BlockType blockType) {
|
BlockUCState state, DatanodeStorageInfo[] targets, BlockType blockType) {
|
||||||
@ -193,11 +193,11 @@ public long getBlockRecoveryId() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/** Get recover block */
|
/** Get recover block */
|
||||||
public Block getTruncateBlock() {
|
public BlockInfo getTruncateBlock() {
|
||||||
return truncateBlock;
|
return truncateBlock;
|
||||||
}
|
}
|
||||||
|
|
||||||
public void setTruncateBlock(Block recoveryBlock) {
|
public void setTruncateBlock(BlockInfo recoveryBlock) {
|
||||||
this.truncateBlock = recoveryBlock;
|
this.truncateBlock = recoveryBlock;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -104,7 +104,7 @@ static TruncateResult truncate(final FSNamesystem fsn, final String srcArg,
|
|||||||
final BlockInfo last = file.getLastBlock();
|
final BlockInfo last = file.getLastBlock();
|
||||||
if (last != null && last.getBlockUCState()
|
if (last != null && last.getBlockUCState()
|
||||||
== BlockUCState.UNDER_RECOVERY) {
|
== BlockUCState.UNDER_RECOVERY) {
|
||||||
final Block truncatedBlock = last.getUnderConstructionFeature()
|
final BlockInfo truncatedBlock = last.getUnderConstructionFeature()
|
||||||
.getTruncateBlock();
|
.getTruncateBlock();
|
||||||
if (truncatedBlock != null) {
|
if (truncatedBlock != null) {
|
||||||
final long truncateLength = file.computeFileSize(false, false)
|
final long truncateLength = file.computeFileSize(false, false)
|
||||||
@ -259,7 +259,8 @@ static Block prepareFileForTruncate(FSNamesystem fsn, INodesInPath iip,
|
|||||||
oldBlock = file.getLastBlock();
|
oldBlock = file.getLastBlock();
|
||||||
assert !oldBlock.isComplete() : "oldBlock should be under construction";
|
assert !oldBlock.isComplete() : "oldBlock should be under construction";
|
||||||
BlockUnderConstructionFeature uc = oldBlock.getUnderConstructionFeature();
|
BlockUnderConstructionFeature uc = oldBlock.getUnderConstructionFeature();
|
||||||
uc.setTruncateBlock(new Block(oldBlock));
|
uc.setTruncateBlock(new BlockInfoContiguous(oldBlock,
|
||||||
|
oldBlock.getReplication()));
|
||||||
uc.getTruncateBlock().setNumBytes(oldBlock.getNumBytes() - lastBlockDelta);
|
uc.getTruncateBlock().setNumBytes(oldBlock.getNumBytes() - lastBlockDelta);
|
||||||
uc.getTruncateBlock().setGenerationStamp(newBlock.getGenerationStamp());
|
uc.getTruncateBlock().setGenerationStamp(newBlock.getGenerationStamp());
|
||||||
truncatedBlockUC = oldBlock;
|
truncatedBlockUC = oldBlock;
|
||||||
|
@ -3376,7 +3376,7 @@ boolean internalReleaseLease(Lease lease, String src, INodesInPath iip,
|
|||||||
BlockUnderConstructionFeature uc =
|
BlockUnderConstructionFeature uc =
|
||||||
lastBlock.getUnderConstructionFeature();
|
lastBlock.getUnderConstructionFeature();
|
||||||
// determine if last block was intended to be truncated
|
// determine if last block was intended to be truncated
|
||||||
Block recoveryBlock = uc.getTruncateBlock();
|
BlockInfo recoveryBlock = uc.getTruncateBlock();
|
||||||
boolean truncateRecovery = recoveryBlock != null;
|
boolean truncateRecovery = recoveryBlock != null;
|
||||||
boolean copyOnTruncate = truncateRecovery &&
|
boolean copyOnTruncate = truncateRecovery &&
|
||||||
recoveryBlock.getBlockId() != lastBlock.getBlockId();
|
recoveryBlock.getBlockId() != lastBlock.getBlockId();
|
||||||
|
@ -33,7 +33,6 @@
|
|||||||
import org.apache.hadoop.fs.permission.FsPermission;
|
import org.apache.hadoop.fs.permission.FsPermission;
|
||||||
import org.apache.hadoop.fs.permission.PermissionStatus;
|
import org.apache.hadoop.fs.permission.PermissionStatus;
|
||||||
import org.apache.hadoop.hdfs.DFSUtilClient;
|
import org.apache.hadoop.hdfs.DFSUtilClient;
|
||||||
import org.apache.hadoop.hdfs.protocol.Block;
|
|
||||||
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
||||||
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfo;
|
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfo;
|
||||||
import org.apache.hadoop.hdfs.server.blockmanagement.BlockStoragePolicySuite;
|
import org.apache.hadoop.hdfs.server.blockmanagement.BlockStoragePolicySuite;
|
||||||
@ -1054,12 +1053,11 @@ public void addDeleteBlock(BlockInfo toDelete) {
|
|||||||
if(uc == null) {
|
if(uc == null) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
Block truncateBlock = uc.getTruncateBlock();
|
BlockInfo truncateBlock = uc.getTruncateBlock();
|
||||||
if(truncateBlock == null || truncateBlock.equals(toDelete)) {
|
if(truncateBlock == null || truncateBlock.equals(toDelete)) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
assert truncateBlock instanceof BlockInfo : "should be BlockInfo";
|
addDeleteBlock(truncateBlock);
|
||||||
addDeleteBlock((BlockInfo) truncateBlock);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
public void addUpdateReplicationFactor(BlockInfo block, short targetRepl) {
|
public void addUpdateReplicationFactor(BlockInfo block, short targetRepl) {
|
||||||
|
@ -17,7 +17,6 @@
|
|||||||
*/
|
*/
|
||||||
package org.apache.hadoop.hdfs.server.namenode.snapshot;
|
package org.apache.hadoop.hdfs.server.namenode.snapshot;
|
||||||
|
|
||||||
import org.apache.hadoop.hdfs.protocol.Block;
|
|
||||||
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfo;
|
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfo;
|
||||||
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoContiguous;
|
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoContiguous;
|
||||||
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants;
|
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants;
|
||||||
@ -129,7 +128,7 @@ void combineAndCollectSnapshotBlocks(
|
|||||||
}
|
}
|
||||||
// Check if last block is part of truncate recovery
|
// Check if last block is part of truncate recovery
|
||||||
BlockInfo lastBlock = file.getLastBlock();
|
BlockInfo lastBlock = file.getLastBlock();
|
||||||
Block dontRemoveBlock = null;
|
BlockInfo dontRemoveBlock = null;
|
||||||
if (lastBlock != null && lastBlock.getBlockUCState().equals(
|
if (lastBlock != null && lastBlock.getBlockUCState().equals(
|
||||||
HdfsServerConstants.BlockUCState.UNDER_RECOVERY)) {
|
HdfsServerConstants.BlockUCState.UNDER_RECOVERY)) {
|
||||||
dontRemoveBlock = lastBlock.getUnderConstructionFeature()
|
dontRemoveBlock = lastBlock.getUnderConstructionFeature()
|
||||||
|
Loading…
Reference in New Issue
Block a user