HADOOP-16158. DistCp to support checksum validation when copy blocks in parallel (#919)
* DistCp to support checksum validation when copy blocks in parallel * address review comments * add checksums comparison test for combine mode
This commit is contained in:
parent
a707bb7c1b
commit
c765584eb2
@ -73,6 +73,7 @@ public class CopyCommitter extends FileOutputCommitter {
|
|||||||
private boolean overwrite = false;
|
private boolean overwrite = false;
|
||||||
private boolean targetPathExists = true;
|
private boolean targetPathExists = true;
|
||||||
private boolean ignoreFailures = false;
|
private boolean ignoreFailures = false;
|
||||||
|
private boolean skipCrc = false;
|
||||||
private int blocksPerChunk = 0;
|
private int blocksPerChunk = 0;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -87,6 +88,9 @@ public CopyCommitter(Path outputPath, TaskAttemptContext context) throws IOExcep
|
|||||||
blocksPerChunk = context.getConfiguration().getInt(
|
blocksPerChunk = context.getConfiguration().getInt(
|
||||||
DistCpOptionSwitch.BLOCKS_PER_CHUNK.getConfigLabel(), 0);
|
DistCpOptionSwitch.BLOCKS_PER_CHUNK.getConfigLabel(), 0);
|
||||||
LOG.debug("blocks per chunk {}", blocksPerChunk);
|
LOG.debug("blocks per chunk {}", blocksPerChunk);
|
||||||
|
skipCrc = context.getConfiguration().getBoolean(
|
||||||
|
DistCpOptionSwitch.SKIP_CRC.getConfigLabel(), false);
|
||||||
|
LOG.debug("skip CRC is {}", skipCrc);
|
||||||
this.taskAttemptContext = context;
|
this.taskAttemptContext = context;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -247,7 +251,8 @@ private void concatFileChunks(Configuration conf) throws IOException {
|
|||||||
== srcFileStatus.getLen()) {
|
== srcFileStatus.getLen()) {
|
||||||
// This is the last chunk of the splits, consolidate allChunkPaths
|
// This is the last chunk of the splits, consolidate allChunkPaths
|
||||||
try {
|
try {
|
||||||
concatFileChunks(conf, targetFile, allChunkPaths);
|
concatFileChunks(conf, srcFileStatus.getPath(), targetFile,
|
||||||
|
allChunkPaths);
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
// If the concat failed because a chunk file doesn't exist,
|
// If the concat failed because a chunk file doesn't exist,
|
||||||
// then we assume that the CopyMapper has skipped copying this
|
// then we assume that the CopyMapper has skipped copying this
|
||||||
@ -603,8 +608,9 @@ private void commitData(Configuration conf) throws IOException {
|
|||||||
/**
|
/**
|
||||||
* Concat the passed chunk files into one and rename it the targetFile.
|
* Concat the passed chunk files into one and rename it the targetFile.
|
||||||
*/
|
*/
|
||||||
private void concatFileChunks(Configuration conf, Path targetFile,
|
private void concatFileChunks(Configuration conf, Path sourceFile,
|
||||||
LinkedList<Path> allChunkPaths) throws IOException {
|
Path targetFile, LinkedList<Path> allChunkPaths)
|
||||||
|
throws IOException {
|
||||||
if (allChunkPaths.size() == 1) {
|
if (allChunkPaths.size() == 1) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
@ -613,6 +619,7 @@ private void concatFileChunks(Configuration conf, Path targetFile,
|
|||||||
+ allChunkPaths.size());
|
+ allChunkPaths.size());
|
||||||
}
|
}
|
||||||
FileSystem dstfs = targetFile.getFileSystem(conf);
|
FileSystem dstfs = targetFile.getFileSystem(conf);
|
||||||
|
FileSystem srcfs = sourceFile.getFileSystem(conf);
|
||||||
|
|
||||||
Path firstChunkFile = allChunkPaths.removeFirst();
|
Path firstChunkFile = allChunkPaths.removeFirst();
|
||||||
Path[] restChunkFiles = new Path[allChunkPaths.size()];
|
Path[] restChunkFiles = new Path[allChunkPaths.size()];
|
||||||
@ -630,6 +637,8 @@ private void concatFileChunks(Configuration conf, Path targetFile,
|
|||||||
LOG.debug("concat: result: " + dstfs.getFileStatus(firstChunkFile));
|
LOG.debug("concat: result: " + dstfs.getFileStatus(firstChunkFile));
|
||||||
}
|
}
|
||||||
rename(dstfs, firstChunkFile, targetFile);
|
rename(dstfs, firstChunkFile, targetFile);
|
||||||
|
DistCpUtils.compareFileLengthsAndChecksums(
|
||||||
|
srcfs, sourceFile, null, dstfs, targetFile, skipCrc);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -143,15 +143,8 @@ private long doCopy(CopyListingFileStatus source, Path target,
|
|||||||
offset, context, fileAttributes, sourceChecksum);
|
offset, context, fileAttributes, sourceChecksum);
|
||||||
|
|
||||||
if (!source.isSplit()) {
|
if (!source.isSplit()) {
|
||||||
compareFileLengths(source, targetPath, configuration, bytesRead
|
DistCpUtils.compareFileLengthsAndChecksums(sourceFS, sourcePath,
|
||||||
+ offset);
|
sourceChecksum, targetFS, targetPath, skipCrc);
|
||||||
}
|
|
||||||
//At this point, src&dest lengths are same. if length==0, we skip checksum
|
|
||||||
if ((bytesRead != 0) && (!skipCrc)) {
|
|
||||||
if (!source.isSplit()) {
|
|
||||||
compareCheckSums(sourceFS, source.getPath(), sourceChecksum,
|
|
||||||
targetFS, targetPath);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
// it's not append or direct write (preferred for s3a) case, thus we first
|
// it's not append or direct write (preferred for s3a) case, thus we first
|
||||||
// write to a temporary file, then rename it to the target path.
|
// write to a temporary file, then rename it to the target path.
|
||||||
@ -216,51 +209,6 @@ private long copyToFile(Path targetPath, FileSystem targetFS,
|
|||||||
context);
|
context);
|
||||||
}
|
}
|
||||||
|
|
||||||
private void compareFileLengths(CopyListingFileStatus source, Path target,
|
|
||||||
Configuration configuration, long targetLen)
|
|
||||||
throws IOException {
|
|
||||||
final Path sourcePath = source.getPath();
|
|
||||||
FileSystem fs = sourcePath.getFileSystem(configuration);
|
|
||||||
long srcLen = fs.getFileStatus(sourcePath).getLen();
|
|
||||||
if (srcLen != targetLen)
|
|
||||||
throw new IOException("Mismatch in length of source:" + sourcePath + " (" + srcLen +
|
|
||||||
") and target:" + target + " (" + targetLen + ")");
|
|
||||||
}
|
|
||||||
|
|
||||||
private void compareCheckSums(FileSystem sourceFS, Path source,
|
|
||||||
FileChecksum sourceChecksum, FileSystem targetFS, Path target)
|
|
||||||
throws IOException {
|
|
||||||
if (!DistCpUtils.checksumsAreEqual(sourceFS, source, sourceChecksum,
|
|
||||||
targetFS, target)) {
|
|
||||||
StringBuilder errorMessage =
|
|
||||||
new StringBuilder("Checksum mismatch between ")
|
|
||||||
.append(source).append(" and ").append(target).append(".");
|
|
||||||
boolean addSkipHint = false;
|
|
||||||
String srcScheme = sourceFS.getScheme();
|
|
||||||
String targetScheme = targetFS.getScheme();
|
|
||||||
if (!srcScheme.equals(targetScheme)
|
|
||||||
&& !(srcScheme.contains("hdfs") && targetScheme.contains("hdfs"))) {
|
|
||||||
// the filesystems are different and they aren't both hdfs connectors
|
|
||||||
errorMessage.append("Source and destination filesystems are of"
|
|
||||||
+ " different types\n")
|
|
||||||
.append("Their checksum algorithms may be incompatible");
|
|
||||||
addSkipHint = true;
|
|
||||||
} else if (sourceFS.getFileStatus(source).getBlockSize() !=
|
|
||||||
targetFS.getFileStatus(target).getBlockSize()) {
|
|
||||||
errorMessage.append(" Source and target differ in block-size.\n")
|
|
||||||
.append(" Use -pb to preserve block-sizes during copy.");
|
|
||||||
addSkipHint = true;
|
|
||||||
}
|
|
||||||
if (addSkipHint) {
|
|
||||||
errorMessage.append(" You can skip checksum-checks altogether "
|
|
||||||
+ " with -skipcrccheck.\n")
|
|
||||||
.append(" (NOTE: By skipping checksums, one runs the risk of " +
|
|
||||||
"masking data-corruption during file-transfer.)\n");
|
|
||||||
}
|
|
||||||
throw new IOException(errorMessage.toString());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
//If target file exists and unable to delete target - fail
|
//If target file exists and unable to delete target - fail
|
||||||
//If target doesn't exist and unable to create parent folder - fail
|
//If target doesn't exist and unable to create parent folder - fail
|
||||||
//If target is successfully deleted and parent exists, if rename fails - fail
|
//If target is successfully deleted and parent exists, if rename fails - fail
|
||||||
|
@ -583,6 +583,66 @@ public static boolean checksumsAreEqual(FileSystem sourceFS, Path source,
|
|||||||
sourceChecksum.equals(targetChecksum));
|
sourceChecksum.equals(targetChecksum));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Utility to compare file lengths and checksums for source and target.
|
||||||
|
*
|
||||||
|
* @param sourceFS FileSystem for the source path.
|
||||||
|
* @param source The source path.
|
||||||
|
* @param sourceChecksum The checksum of the source file. If it is null we
|
||||||
|
* still need to retrieve it through sourceFS.
|
||||||
|
* @param targetFS FileSystem for the target path.
|
||||||
|
* @param target The target path.
|
||||||
|
* @param skipCrc The flag to indicate whether to skip checksums.
|
||||||
|
* @throws IOException if there's a mismatch in file lengths or checksums.
|
||||||
|
*/
|
||||||
|
public static void compareFileLengthsAndChecksums(
|
||||||
|
FileSystem sourceFS, Path source, FileChecksum sourceChecksum,
|
||||||
|
FileSystem targetFS, Path target, boolean skipCrc) throws IOException {
|
||||||
|
long srcLen = sourceFS.getFileStatus(source).getLen();
|
||||||
|
long tgtLen = targetFS.getFileStatus(target).getLen();
|
||||||
|
if (srcLen != tgtLen) {
|
||||||
|
throw new IOException(
|
||||||
|
"Mismatch in length of source:" + source + " (" + srcLen
|
||||||
|
+ ") and target:" + target + " (" + tgtLen + ")");
|
||||||
|
}
|
||||||
|
|
||||||
|
//At this point, src & dest lengths are same. if length==0, we skip checksum
|
||||||
|
if ((srcLen != 0) && (!skipCrc)) {
|
||||||
|
if (!checksumsAreEqual(sourceFS, source, sourceChecksum,
|
||||||
|
targetFS, target)) {
|
||||||
|
StringBuilder errorMessage =
|
||||||
|
new StringBuilder("Checksum mismatch between ")
|
||||||
|
.append(source).append(" and ").append(target).append(".");
|
||||||
|
boolean addSkipHint = false;
|
||||||
|
String srcScheme = sourceFS.getScheme();
|
||||||
|
String targetScheme = targetFS.getScheme();
|
||||||
|
if (!srcScheme.equals(targetScheme)) {
|
||||||
|
// the filesystems are different and they aren't both hdfs connectors
|
||||||
|
errorMessage.append("Source and destination filesystems are of"
|
||||||
|
+ " different types\n")
|
||||||
|
.append("Their checksum algorithms may be incompatible");
|
||||||
|
addSkipHint = true;
|
||||||
|
} else if (sourceFS.getFileStatus(source).getBlockSize() !=
|
||||||
|
targetFS.getFileStatus(target).getBlockSize()) {
|
||||||
|
errorMessage.append(" Source and target differ in block-size.\n")
|
||||||
|
.append(" Use -pb to preserve block-sizes during copy.");
|
||||||
|
addSkipHint = true;
|
||||||
|
}
|
||||||
|
if (addSkipHint) {
|
||||||
|
errorMessage
|
||||||
|
.append(" You can choose file-level checksum validation via "
|
||||||
|
+ "-Ddfs.checksum.combine.mode=COMPOSITE_CRC when block-sizes"
|
||||||
|
+ " or filesystems are different.")
|
||||||
|
.append(" Or you can skip checksum-checks altogether "
|
||||||
|
+ " with -skipcrccheck.\n")
|
||||||
|
.append(" (NOTE: By skipping checksums, one runs the risk of " +
|
||||||
|
"masking data-corruption during file-transfer.)\n");
|
||||||
|
}
|
||||||
|
throw new IOException(errorMessage.toString());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Return the Path for a given chunk.
|
* Return the Path for a given chunk.
|
||||||
* Used when splitting large file into chunks to copy in parallel.
|
* Used when splitting large file into chunks to copy in parallel.
|
||||||
|
@ -25,6 +25,9 @@
|
|||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.fs.permission.FsPermission;
|
import org.apache.hadoop.fs.permission.FsPermission;
|
||||||
|
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||||
|
import org.apache.hadoop.hdfs.DFSTestUtil;
|
||||||
|
import org.apache.hadoop.hdfs.DistributedFileSystem;
|
||||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||||
import org.apache.hadoop.io.IOUtils;
|
import org.apache.hadoop.io.IOUtils;
|
||||||
import org.apache.hadoop.io.SequenceFile;
|
import org.apache.hadoop.io.SequenceFile;
|
||||||
@ -33,6 +36,7 @@
|
|||||||
import org.apache.hadoop.mapreduce.task.TaskAttemptContextImpl;
|
import org.apache.hadoop.mapreduce.task.TaskAttemptContextImpl;
|
||||||
import org.apache.hadoop.mapreduce.task.JobContextImpl;
|
import org.apache.hadoop.mapreduce.task.JobContextImpl;
|
||||||
import org.apache.hadoop.mapreduce.lib.output.NullOutputFormat;
|
import org.apache.hadoop.mapreduce.lib.output.NullOutputFormat;
|
||||||
|
import org.apache.hadoop.test.GenericTestUtils;
|
||||||
import org.apache.hadoop.tools.CopyListing;
|
import org.apache.hadoop.tools.CopyListing;
|
||||||
import org.apache.hadoop.tools.CopyListingFileStatus;
|
import org.apache.hadoop.tools.CopyListingFileStatus;
|
||||||
import org.apache.hadoop.tools.DistCpConstants;
|
import org.apache.hadoop.tools.DistCpConstants;
|
||||||
@ -40,6 +44,7 @@
|
|||||||
import org.apache.hadoop.tools.DistCpOptions;
|
import org.apache.hadoop.tools.DistCpOptions;
|
||||||
import org.apache.hadoop.tools.DistCpOptions.FileAttribute;
|
import org.apache.hadoop.tools.DistCpOptions.FileAttribute;
|
||||||
import org.apache.hadoop.tools.GlobbedCopyListing;
|
import org.apache.hadoop.tools.GlobbedCopyListing;
|
||||||
|
import org.apache.hadoop.tools.util.DistCpUtils;
|
||||||
import org.apache.hadoop.tools.util.TestDistCpUtils;
|
import org.apache.hadoop.tools.util.TestDistCpUtils;
|
||||||
import org.apache.hadoop.security.Credentials;
|
import org.apache.hadoop.security.Credentials;
|
||||||
import org.junit.*;
|
import org.junit.*;
|
||||||
@ -55,13 +60,16 @@ public class TestCopyCommitter {
|
|||||||
|
|
||||||
private static final Random rand = new Random();
|
private static final Random rand = new Random();
|
||||||
|
|
||||||
|
private static final long BLOCK_SIZE = 1024;
|
||||||
private static final Credentials CREDENTIALS = new Credentials();
|
private static final Credentials CREDENTIALS = new Credentials();
|
||||||
public static final int PORT = 39737;
|
public static final int PORT = 39737;
|
||||||
|
|
||||||
|
|
||||||
private static Configuration config;
|
private static Configuration clusterConfig;
|
||||||
private static MiniDFSCluster cluster;
|
private static MiniDFSCluster cluster;
|
||||||
|
|
||||||
|
private Configuration config;
|
||||||
|
|
||||||
private static Job getJobForClient() throws IOException {
|
private static Job getJobForClient() throws IOException {
|
||||||
Job job = Job.getInstance(new Configuration());
|
Job job = Job.getInstance(new Configuration());
|
||||||
job.getConfiguration().set("mapred.job.tracker", "localhost:" + PORT);
|
job.getConfiguration().set("mapred.job.tracker", "localhost:" + PORT);
|
||||||
@ -73,10 +81,17 @@ private static Job getJobForClient() throws IOException {
|
|||||||
|
|
||||||
@BeforeClass
|
@BeforeClass
|
||||||
public static void create() throws IOException {
|
public static void create() throws IOException {
|
||||||
config = getJobForClient().getConfiguration();
|
clusterConfig = getJobForClient().getConfiguration();
|
||||||
config.setLong(DistCpConstants.CONF_LABEL_TOTAL_BYTES_TO_BE_COPIED, 0);
|
clusterConfig.setLong(
|
||||||
cluster = new MiniDFSCluster.Builder(config).numDataNodes(1).format(true)
|
DistCpConstants.CONF_LABEL_TOTAL_BYTES_TO_BE_COPIED, 0);
|
||||||
.build();
|
clusterConfig.setLong(
|
||||||
|
DFSConfigKeys.DFS_NAMENODE_MIN_BLOCK_SIZE_KEY, BLOCK_SIZE);
|
||||||
|
clusterConfig.setLong(
|
||||||
|
DFSConfigKeys.DFS_BLOCK_SIZE_KEY, BLOCK_SIZE);
|
||||||
|
cluster = new MiniDFSCluster.Builder(clusterConfig)
|
||||||
|
.numDataNodes(1)
|
||||||
|
.format(true)
|
||||||
|
.build();
|
||||||
}
|
}
|
||||||
|
|
||||||
@AfterClass
|
@AfterClass
|
||||||
@ -88,6 +103,7 @@ public static void destroy() {
|
|||||||
|
|
||||||
@Before
|
@Before
|
||||||
public void createMetaFolder() throws IOException {
|
public void createMetaFolder() throws IOException {
|
||||||
|
config = new Configuration(clusterConfig);
|
||||||
config.set(DistCpConstants.CONF_LABEL_META_FOLDER, "/meta");
|
config.set(DistCpConstants.CONF_LABEL_META_FOLDER, "/meta");
|
||||||
Path meta = new Path("/meta");
|
Path meta = new Path("/meta");
|
||||||
cluster.getFileSystem().mkdirs(meta);
|
cluster.getFileSystem().mkdirs(meta);
|
||||||
@ -397,6 +413,141 @@ public void testAtomicCommitExistingFinal() throws IOException {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testCommitWithChecksumMismatchAndSkipCrc() throws IOException {
|
||||||
|
testCommitWithChecksumMismatch(true);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testCommitWithChecksumMismatchWithoutSkipCrc()
|
||||||
|
throws IOException {
|
||||||
|
testCommitWithChecksumMismatch(false);
|
||||||
|
}
|
||||||
|
|
||||||
|
private void testCommitWithChecksumMismatch(boolean skipCrc)
|
||||||
|
throws IOException {
|
||||||
|
|
||||||
|
TaskAttemptContext taskAttemptContext = getTaskAttemptContext(config);
|
||||||
|
JobContext jobContext = new JobContextImpl(
|
||||||
|
taskAttemptContext.getConfiguration(),
|
||||||
|
taskAttemptContext.getTaskAttemptID().getJobID());
|
||||||
|
Configuration conf = jobContext.getConfiguration();
|
||||||
|
|
||||||
|
String sourceBase;
|
||||||
|
String targetBase;
|
||||||
|
FileSystem fs = null;
|
||||||
|
try {
|
||||||
|
fs = FileSystem.get(conf);
|
||||||
|
sourceBase = "/tmp1/" + String.valueOf(rand.nextLong());
|
||||||
|
targetBase = "/tmp1/" + String.valueOf(rand.nextLong());
|
||||||
|
|
||||||
|
int blocksPerChunk = 5;
|
||||||
|
String srcFilename = "/srcdata";
|
||||||
|
createSrcAndWorkFilesWithDifferentChecksum(fs, targetBase, sourceBase,
|
||||||
|
srcFilename, blocksPerChunk);
|
||||||
|
|
||||||
|
DistCpOptions options = new DistCpOptions.Builder(
|
||||||
|
Collections.singletonList(new Path(sourceBase)),
|
||||||
|
new Path("/out"))
|
||||||
|
.withBlocksPerChunk(blocksPerChunk)
|
||||||
|
.withCRC(skipCrc)
|
||||||
|
.build();
|
||||||
|
options.appendToConf(conf);
|
||||||
|
conf.setBoolean(
|
||||||
|
DistCpConstants.CONF_LABEL_SIMPLE_LISTING_RANDOMIZE_FILES, false);
|
||||||
|
DistCpContext context = new DistCpContext(options);
|
||||||
|
context.setTargetPathExists(false);
|
||||||
|
|
||||||
|
CopyListing listing = new GlobbedCopyListing(conf, CREDENTIALS);
|
||||||
|
Path listingFile = new Path("/tmp1/"
|
||||||
|
+ String.valueOf(rand.nextLong()));
|
||||||
|
listing.buildListing(listingFile, context);
|
||||||
|
|
||||||
|
conf.set(DistCpConstants.CONF_LABEL_TARGET_WORK_PATH, targetBase);
|
||||||
|
conf.set(DistCpConstants.CONF_LABEL_TARGET_FINAL_PATH, targetBase);
|
||||||
|
|
||||||
|
OutputCommitter committer = new CopyCommitter(
|
||||||
|
null, taskAttemptContext);
|
||||||
|
try {
|
||||||
|
committer.commitJob(jobContext);
|
||||||
|
if (!skipCrc) {
|
||||||
|
Assert.fail("Expected commit to fail");
|
||||||
|
}
|
||||||
|
Assert.assertFalse(DistCpUtils.checksumsAreEqual(
|
||||||
|
fs, new Path(sourceBase + srcFilename), null,
|
||||||
|
fs, new Path(targetBase + srcFilename)));
|
||||||
|
} catch(IOException exception) {
|
||||||
|
if (skipCrc) {
|
||||||
|
LOG.error("Unexpected exception is found", exception);
|
||||||
|
throw exception;
|
||||||
|
}
|
||||||
|
Throwable cause = exception.getCause();
|
||||||
|
GenericTestUtils.assertExceptionContains(
|
||||||
|
"Checksum mismatch", cause);
|
||||||
|
}
|
||||||
|
} finally {
|
||||||
|
TestDistCpUtils.delete(fs, "/tmp1");
|
||||||
|
TestDistCpUtils.delete(fs, "/meta");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Create a source file and its DistCp working files with different checksum
|
||||||
|
* to test the checksum validation for copying blocks in parallel.
|
||||||
|
*
|
||||||
|
* For the ease of construction, it assumes a source file can be broken down
|
||||||
|
* into 2 working files (or 2 chunks).
|
||||||
|
*
|
||||||
|
* So for a source file with length =
|
||||||
|
* BLOCK_SIZE * blocksPerChunk + BLOCK_SIZE / 2,
|
||||||
|
* its 1st working file will have length =
|
||||||
|
* BLOCK_SIZE * blocksPerChunk,
|
||||||
|
* then the 2nd working file will have length =
|
||||||
|
* BLOCK_SIZE / 2.
|
||||||
|
* And the working files are generated with a different seed to mimic
|
||||||
|
* same length but different checksum scenario.
|
||||||
|
*
|
||||||
|
* @param fs the FileSystem
|
||||||
|
* @param targetBase the path to the working files
|
||||||
|
* @param sourceBase the path to a source file
|
||||||
|
* @param filename the filename to copy and work on
|
||||||
|
* @param blocksPerChunk the blocks per chunk config that enables copying
|
||||||
|
* blocks in parallel
|
||||||
|
* @throws IOException when it fails to create files
|
||||||
|
*/
|
||||||
|
private void createSrcAndWorkFilesWithDifferentChecksum(FileSystem fs,
|
||||||
|
String targetBase,
|
||||||
|
String sourceBase,
|
||||||
|
String filename,
|
||||||
|
int blocksPerChunk)
|
||||||
|
throws IOException {
|
||||||
|
|
||||||
|
long srcSeed = System.currentTimeMillis();
|
||||||
|
long dstSeed = srcSeed + rand.nextLong();
|
||||||
|
int bufferLen = 128;
|
||||||
|
short replFactor = 2;
|
||||||
|
Path srcData = new Path(sourceBase + filename);
|
||||||
|
|
||||||
|
// create data with 2 chunks: the 2nd chunk has half of the block size
|
||||||
|
long firstChunkLength = BLOCK_SIZE * blocksPerChunk;
|
||||||
|
long secondChunkLength = BLOCK_SIZE / 2;
|
||||||
|
|
||||||
|
DFSTestUtil.createFile(fs, srcData,
|
||||||
|
bufferLen, firstChunkLength, BLOCK_SIZE, replFactor,
|
||||||
|
srcSeed);
|
||||||
|
DFSTestUtil.appendFileNewBlock((DistributedFileSystem) fs, srcData,
|
||||||
|
(int) secondChunkLength);
|
||||||
|
|
||||||
|
DFSTestUtil.createFile(fs, new Path(targetBase
|
||||||
|
+ filename + ".____distcpSplit____0."
|
||||||
|
+ firstChunkLength), bufferLen,
|
||||||
|
firstChunkLength, BLOCK_SIZE, replFactor, dstSeed);
|
||||||
|
DFSTestUtil.createFile(fs, new Path(targetBase
|
||||||
|
+ filename + ".____distcpSplit____"
|
||||||
|
+ firstChunkLength + "." + secondChunkLength), bufferLen,
|
||||||
|
secondChunkLength, BLOCK_SIZE, replFactor, dstSeed);
|
||||||
|
}
|
||||||
|
|
||||||
private TaskAttemptContext getTaskAttemptContext(Configuration conf) {
|
private TaskAttemptContext getTaskAttemptContext(Configuration conf) {
|
||||||
return new TaskAttemptContextImpl(conf,
|
return new TaskAttemptContextImpl(conf,
|
||||||
new TaskAttemptID("200707121733", 1, TaskType.MAP, 1, 1));
|
new TaskAttemptID("200707121733", 1, TaskType.MAP, 1, 1));
|
||||||
|
@ -28,10 +28,12 @@
|
|||||||
import org.apache.hadoop.fs.permission.AclEntry;
|
import org.apache.hadoop.fs.permission.AclEntry;
|
||||||
import org.apache.hadoop.fs.permission.FsPermission;
|
import org.apache.hadoop.fs.permission.FsPermission;
|
||||||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||||
|
import org.apache.hadoop.hdfs.DFSTestUtil;
|
||||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||||
import org.apache.hadoop.hdfs.server.namenode.INodeFile;
|
import org.apache.hadoop.hdfs.server.namenode.INodeFile;
|
||||||
import org.apache.hadoop.hdfs.tools.ECAdmin;
|
import org.apache.hadoop.hdfs.tools.ECAdmin;
|
||||||
import org.apache.hadoop.io.IOUtils;
|
import org.apache.hadoop.io.IOUtils;
|
||||||
|
import org.apache.hadoop.test.GenericTestUtils;
|
||||||
import org.apache.hadoop.tools.CopyListingFileStatus;
|
import org.apache.hadoop.tools.CopyListingFileStatus;
|
||||||
import org.apache.hadoop.tools.DistCpOptionSwitch;
|
import org.apache.hadoop.tools.DistCpOptionSwitch;
|
||||||
import org.apache.hadoop.tools.DistCpOptions.FileAttribute;
|
import org.apache.hadoop.tools.DistCpOptions.FileAttribute;
|
||||||
@ -1205,6 +1207,71 @@ public void testPreserveOnDirectoryDownwardRecursion() throws IOException {
|
|||||||
Assert.assertFalse(srcStatus.getReplication() == f2Status.getReplication());
|
Assert.assertFalse(srcStatus.getReplication() == f2Status.getReplication());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testCompareFileLengthsAndChecksums() throws IOException {
|
||||||
|
|
||||||
|
String base = "/tmp/verify-checksum/";
|
||||||
|
long srcSeed = System.currentTimeMillis();
|
||||||
|
long dstSeed = srcSeed + rand.nextLong();
|
||||||
|
short replFactor = 2;
|
||||||
|
|
||||||
|
FileSystem fs = FileSystem.get(config);
|
||||||
|
Path basePath = new Path(base);
|
||||||
|
fs.mkdirs(basePath);
|
||||||
|
|
||||||
|
// empty lengths comparison
|
||||||
|
Path srcWithLen0 = new Path(base + "srcLen0");
|
||||||
|
Path dstWithLen0 = new Path(base + "dstLen0");
|
||||||
|
fs.create(srcWithLen0).close();
|
||||||
|
fs.create(dstWithLen0).close();
|
||||||
|
DistCpUtils.compareFileLengthsAndChecksums(fs, srcWithLen0,
|
||||||
|
null, fs, dstWithLen0, false);
|
||||||
|
|
||||||
|
// different lengths comparison
|
||||||
|
Path srcWithLen1 = new Path(base + "srcLen1");
|
||||||
|
Path dstWithLen2 = new Path(base + "dstLen2");
|
||||||
|
DFSTestUtil.createFile(fs, srcWithLen1, 1, replFactor, srcSeed);
|
||||||
|
DFSTestUtil.createFile(fs, dstWithLen2, 2, replFactor, srcSeed);
|
||||||
|
try {
|
||||||
|
DistCpUtils.compareFileLengthsAndChecksums(fs, srcWithLen1,
|
||||||
|
null, fs, dstWithLen2, false);
|
||||||
|
Assert.fail("Expected different lengths comparison to fail!");
|
||||||
|
} catch (IOException e) {
|
||||||
|
GenericTestUtils.assertExceptionContains(
|
||||||
|
"Mismatch in length", e);
|
||||||
|
}
|
||||||
|
|
||||||
|
// checksums matched
|
||||||
|
Path srcWithChecksum1 = new Path(base + "srcChecksum1");
|
||||||
|
Path dstWithChecksum1 = new Path(base + "dstChecksum1");
|
||||||
|
DFSTestUtil.createFile(fs, srcWithChecksum1, 1024,
|
||||||
|
replFactor, srcSeed);
|
||||||
|
DFSTestUtil.createFile(fs, dstWithChecksum1, 1024,
|
||||||
|
replFactor, srcSeed);
|
||||||
|
DistCpUtils.compareFileLengthsAndChecksums(fs, srcWithChecksum1,
|
||||||
|
null, fs, dstWithChecksum1, false);
|
||||||
|
DistCpUtils.compareFileLengthsAndChecksums(fs, srcWithChecksum1,
|
||||||
|
fs.getFileChecksum(srcWithChecksum1), fs, dstWithChecksum1,
|
||||||
|
false);
|
||||||
|
|
||||||
|
// checksums mismatched
|
||||||
|
Path dstWithChecksum2 = new Path(base + "dstChecksum2");
|
||||||
|
DFSTestUtil.createFile(fs, dstWithChecksum2, 1024,
|
||||||
|
replFactor, dstSeed);
|
||||||
|
try {
|
||||||
|
DistCpUtils.compareFileLengthsAndChecksums(fs, srcWithChecksum1,
|
||||||
|
null, fs, dstWithChecksum2, false);
|
||||||
|
Assert.fail("Expected different checksums comparison to fail!");
|
||||||
|
} catch (IOException e) {
|
||||||
|
GenericTestUtils.assertExceptionContains(
|
||||||
|
"Checksum mismatch", e);
|
||||||
|
}
|
||||||
|
|
||||||
|
// checksums mismatched but skipped
|
||||||
|
DistCpUtils.compareFileLengthsAndChecksums(fs, srcWithChecksum1,
|
||||||
|
null, fs, dstWithChecksum2, true);
|
||||||
|
}
|
||||||
|
|
||||||
private static Random rand = new Random();
|
private static Random rand = new Random();
|
||||||
|
|
||||||
public static String createTestSetup(FileSystem fs) throws IOException {
|
public static String createTestSetup(FileSystem fs) throws IOException {
|
||||||
|
@ -0,0 +1,115 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.tools.util;
|
||||||
|
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
|
import org.apache.hadoop.fs.Path;
|
||||||
|
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||||
|
import org.apache.hadoop.hdfs.DFSTestUtil;
|
||||||
|
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||||
|
import org.apache.hadoop.test.GenericTestUtils;
|
||||||
|
import org.junit.After;
|
||||||
|
import org.junit.Assert;
|
||||||
|
import org.junit.Before;
|
||||||
|
import org.junit.Rule;
|
||||||
|
import org.junit.Test;
|
||||||
|
import org.junit.rules.TestName;
|
||||||
|
import org.slf4j.Logger;
|
||||||
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Test length and checksums comparison with checksum combine mode.
|
||||||
|
* When the combine mode is COMPOSITE_CRC, it should tolerate different file
|
||||||
|
* systems and different block sizes.
|
||||||
|
*/
|
||||||
|
public class TestDistCpUtilsWithCombineMode {
|
||||||
|
private static final Logger LOG = LoggerFactory.getLogger(TestDistCpUtils.class);
|
||||||
|
|
||||||
|
private Configuration config;
|
||||||
|
private MiniDFSCluster cluster;
|
||||||
|
|
||||||
|
@Rule
|
||||||
|
public TestName testName = new TestName();
|
||||||
|
|
||||||
|
@Before
|
||||||
|
public void create() throws IOException {
|
||||||
|
config = new Configuration();
|
||||||
|
if (testName.getMethodName().contains("WithCombineMode")) {
|
||||||
|
config.set("dfs.checksum.combine.mode", "COMPOSITE_CRC");
|
||||||
|
}
|
||||||
|
config.setLong(
|
||||||
|
DFSConfigKeys.DFS_NAMENODE_MIN_BLOCK_SIZE_KEY, 512);
|
||||||
|
cluster = new MiniDFSCluster.Builder(config)
|
||||||
|
.numDataNodes(2)
|
||||||
|
.format(true)
|
||||||
|
.build();
|
||||||
|
}
|
||||||
|
|
||||||
|
@After
|
||||||
|
public void destroy() {
|
||||||
|
if (cluster != null) {
|
||||||
|
cluster.shutdown();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testChecksumsComparisonWithCombineMode() throws IOException {
|
||||||
|
try {
|
||||||
|
compareSameContentButDiffBlockSizes();
|
||||||
|
} catch (IOException e) {
|
||||||
|
LOG.error("Unexpected exception is found", e);
|
||||||
|
throw e;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testChecksumsComparisonWithoutCombineMode() {
|
||||||
|
try {
|
||||||
|
compareSameContentButDiffBlockSizes();
|
||||||
|
Assert.fail("Expected comparison to fail");
|
||||||
|
} catch (IOException e) {
|
||||||
|
GenericTestUtils.assertExceptionContains(
|
||||||
|
"Checksum mismatch", e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private void compareSameContentButDiffBlockSizes() throws IOException {
|
||||||
|
String base = "/tmp/verify-checksum-" + testName.getMethodName() + "/";
|
||||||
|
long seed = System.currentTimeMillis();
|
||||||
|
short rf = 2;
|
||||||
|
|
||||||
|
FileSystem fs = FileSystem.get(config);
|
||||||
|
Path basePath = new Path(base);
|
||||||
|
fs.mkdirs(basePath);
|
||||||
|
|
||||||
|
// create 2 files of same content but different block-sizes
|
||||||
|
Path src = new Path(base + "src");
|
||||||
|
Path dst = new Path(base + "dst");
|
||||||
|
DFSTestUtil.createFile(fs, src, 256, 1024, 512,
|
||||||
|
rf, seed);
|
||||||
|
DFSTestUtil.createFile(fs, dst, 256, 1024, 1024,
|
||||||
|
rf, seed);
|
||||||
|
// then compare
|
||||||
|
DistCpUtils.compareFileLengthsAndChecksums(fs, src,
|
||||||
|
null, fs, dst, false);
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue
Block a user