MAPREDUCE-3170. Fixed job output commit for deep hierarchies. Contributed by Hitesh Shah.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1183185 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
c46dbedaf9
commit
c7fb49b3c5
@ -1608,6 +1608,9 @@ Release 0.23.0 - Unreleased
|
||||
|
||||
MAPREDUCE-2789. Complete schedulingInfo on CLI. (Eric Payne via acmurthy)
|
||||
|
||||
MAPREDUCE-3170. Fixed job output commit for deep hierarchies. (Hitesh Shah
|
||||
via acmurthy)
|
||||
|
||||
Release 0.22.0 - Unreleased
|
||||
|
||||
INCOMPATIBLE CHANGES
|
||||
|
@ -71,6 +71,7 @@ public void commitJob(JobContext context) throws IOException {
|
||||
//delete the task temp directory from the current jobtempdir
|
||||
JobConf conf = context.getJobConf();
|
||||
Path outputPath = FileOutputFormat.getOutputPath(conf);
|
||||
if (outputPath != null) {
|
||||
FileSystem outputFileSystem = outputPath.getFileSystem(conf);
|
||||
Path tmpDir = new Path(outputPath, getJobAttemptBaseDirName(context) +
|
||||
Path.SEPARATOR + FileOutputCommitter.TEMP_DIR_NAME);
|
||||
@ -84,7 +85,8 @@ public void commitJob(JobContext context) throws IOException {
|
||||
//move the job output to final place
|
||||
Path jobOutputPath =
|
||||
new Path(outputPath, getJobAttemptBaseDirName(context));
|
||||
moveJobOutputs(outputFileSystem, outputPath, jobOutputPath);
|
||||
moveJobOutputs(outputFileSystem,
|
||||
jobOutputPath, outputPath, jobOutputPath);
|
||||
|
||||
// delete the _temporary folder in the output folder
|
||||
cleanupJob(context);
|
||||
@ -94,6 +96,7 @@ public void commitJob(JobContext context) throws IOException {
|
||||
markOutputDirSuccessful(context);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Create a _success file in the job's output folder
|
||||
private void markOutputDirSuccessful(JobContext context) throws IOException {
|
||||
@ -109,10 +112,14 @@ private void markOutputDirSuccessful(JobContext context) throws IOException {
|
||||
}
|
||||
}
|
||||
|
||||
private void moveJobOutputs(FileSystem fs,
|
||||
private void moveJobOutputs(FileSystem fs, final Path origJobOutputPath,
|
||||
Path finalOutputDir, Path jobOutput) throws IOException {
|
||||
LOG.debug("Told to move job output from " + jobOutput
|
||||
+ " to " + finalOutputDir +
|
||||
" and orig job output path is " + origJobOutputPath);
|
||||
if (fs.isFile(jobOutput)) {
|
||||
Path finalOutputPath = getFinalPath(finalOutputDir, jobOutput, jobOutput);
|
||||
Path finalOutputPath =
|
||||
getFinalPath(fs, finalOutputDir, jobOutput, origJobOutputPath);
|
||||
if (!fs.rename(jobOutput, finalOutputPath)) {
|
||||
if (!fs.delete(finalOutputPath, true)) {
|
||||
throw new IOException("Failed to delete earlier output of job");
|
||||
@ -121,18 +128,23 @@ private void moveJobOutputs(FileSystem fs,
|
||||
throw new IOException("Failed to save output of job");
|
||||
}
|
||||
}
|
||||
LOG.debug("Moved " + jobOutput + " to " + finalOutputPath);
|
||||
LOG.debug("Moved job output file from " + jobOutput + " to " +
|
||||
finalOutputPath);
|
||||
} else if (fs.getFileStatus(jobOutput).isDirectory()) {
|
||||
LOG.debug("Job output file " + jobOutput + " is a dir");
|
||||
FileStatus[] paths = fs.listStatus(jobOutput);
|
||||
Path finalOutputPath = getFinalPath(finalOutputDir, jobOutput, jobOutput);
|
||||
Path finalOutputPath =
|
||||
getFinalPath(fs, finalOutputDir, jobOutput, origJobOutputPath);
|
||||
fs.mkdirs(finalOutputPath);
|
||||
LOG.debug("Creating dirs along job output path " + finalOutputPath);
|
||||
if (paths != null) {
|
||||
for (FileStatus path : paths) {
|
||||
moveJobOutputs(fs, finalOutputDir, path.getPath());
|
||||
moveJobOutputs(fs, origJobOutputPath, finalOutputDir, path.getPath());
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
@Deprecated
|
||||
public void cleanupJob(JobContext context) throws IOException {
|
||||
@ -199,8 +211,10 @@ private void moveTaskOutputs(TaskAttemptContext context,
|
||||
throws IOException {
|
||||
TaskAttemptID attemptId = context.getTaskAttemptID();
|
||||
context.getProgressible().progress();
|
||||
LOG.debug("Told to move taskoutput from " + taskOutput
|
||||
+ " to " + jobOutputDir);
|
||||
if (fs.isFile(taskOutput)) {
|
||||
Path finalOutputPath = getFinalPath(jobOutputDir, taskOutput,
|
||||
Path finalOutputPath = getFinalPath(fs, jobOutputDir, taskOutput,
|
||||
getTempTaskOutputPath(context));
|
||||
if (!fs.rename(taskOutput, finalOutputPath)) {
|
||||
if (!fs.delete(finalOutputPath, true)) {
|
||||
@ -214,10 +228,12 @@ private void moveTaskOutputs(TaskAttemptContext context,
|
||||
}
|
||||
LOG.debug("Moved " + taskOutput + " to " + finalOutputPath);
|
||||
} else if(fs.getFileStatus(taskOutput).isDirectory()) {
|
||||
LOG.debug("Taskoutput " + taskOutput + " is a dir");
|
||||
FileStatus[] paths = fs.listStatus(taskOutput);
|
||||
Path finalOutputPath = getFinalPath(jobOutputDir, taskOutput,
|
||||
Path finalOutputPath = getFinalPath(fs, jobOutputDir, taskOutput,
|
||||
getTempTaskOutputPath(context));
|
||||
fs.mkdirs(finalOutputPath);
|
||||
LOG.debug("Creating dirs along path " + finalOutputPath);
|
||||
if (paths != null) {
|
||||
for (FileStatus path : paths) {
|
||||
moveTaskOutputs(context, fs, jobOutputDir, path.getPath());
|
||||
@ -235,14 +251,16 @@ public void abortTask(TaskAttemptContext context) throws IOException {
|
||||
}
|
||||
}
|
||||
|
||||
private Path getFinalPath(Path jobOutputDir, Path taskOutput,
|
||||
@SuppressWarnings("deprecation")
|
||||
private Path getFinalPath(FileSystem fs, Path jobOutputDir, Path taskOutput,
|
||||
Path taskOutputPath) throws IOException {
|
||||
URI taskOutputUri = taskOutput.toUri();
|
||||
URI relativePath = taskOutputPath.toUri().relativize(taskOutputUri);
|
||||
URI taskOutputUri = taskOutput.makeQualified(fs).toUri();
|
||||
URI taskOutputPathUri = taskOutputPath.makeQualified(fs).toUri();
|
||||
URI relativePath = taskOutputPathUri.relativize(taskOutputUri);
|
||||
if (taskOutputUri == relativePath) {
|
||||
//taskOutputPath is not a parent of taskOutput
|
||||
throw new IOException("Can not get the relative path: base = " +
|
||||
taskOutputPath + " child = " + taskOutput);
|
||||
taskOutputPathUri + " child = " + taskOutputUri);
|
||||
}
|
||||
if (relativePath.getPath().length() > 0) {
|
||||
return new Path(jobOutputDir, relativePath.getPath());
|
||||
@ -325,7 +343,10 @@ public void recoverTask(TaskAttemptContext context)
|
||||
new Path(outputPath, getJobAttemptBaseDirName(previousAttempt));
|
||||
if (outputFileSystem.exists(pathToRecover)) {
|
||||
// Move the task outputs to their final place
|
||||
moveJobOutputs(outputFileSystem, jobOutputPath, pathToRecover);
|
||||
LOG.debug("Trying to recover task from " + pathToRecover
|
||||
+ " into " + jobOutputPath);
|
||||
moveJobOutputs(outputFileSystem,
|
||||
pathToRecover, jobOutputPath, pathToRecover);
|
||||
LOG.info("Saved output of job to " + jobOutputPath);
|
||||
}
|
||||
}
|
||||
|
@ -111,6 +111,7 @@ private void markOutputDirSuccessful(MRJobConfig context) throws IOException {
|
||||
* @param context the job's context
|
||||
*/
|
||||
public void commitJob(JobContext context) throws IOException {
|
||||
if (outputPath != null) {
|
||||
//delete the task temp directory from the current jobtempdir
|
||||
Path tmpDir = new Path(outputPath, getJobAttemptBaseDirName(context) +
|
||||
Path.SEPARATOR + FileOutputCommitter.TEMP_DIR_NAME);
|
||||
@ -124,7 +125,7 @@ public void commitJob(JobContext context) throws IOException {
|
||||
//move the job output to final place
|
||||
Path jobOutputPath =
|
||||
new Path(outputPath, getJobAttemptBaseDirName(context));
|
||||
moveJobOutputs(outputFileSystem, outputPath, jobOutputPath);
|
||||
moveJobOutputs(outputFileSystem, jobOutputPath, outputPath, jobOutputPath);
|
||||
|
||||
// delete the _temporary folder and create a _done file in the o/p folder
|
||||
cleanupJob(context);
|
||||
@ -132,11 +133,26 @@ public void commitJob(JobContext context) throws IOException {
|
||||
markOutputDirSuccessful(context);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
private void moveJobOutputs(FileSystem fs,
|
||||
/**
|
||||
* Move job output to final location
|
||||
* @param fs Filesystem handle
|
||||
* @param origJobOutputPath The original location of the job output
|
||||
* Required to generate the relative path for correct moving of data.
|
||||
* @param finalOutputDir The final output directory to which the job output
|
||||
* needs to be moved
|
||||
* @param jobOutput The current job output directory being moved
|
||||
* @throws IOException
|
||||
*/
|
||||
private void moveJobOutputs(FileSystem fs, final Path origJobOutputPath,
|
||||
Path finalOutputDir, Path jobOutput) throws IOException {
|
||||
LOG.debug("Told to move job output from " + jobOutput
|
||||
+ " to " + finalOutputDir +
|
||||
" and orig job output path is " + origJobOutputPath);
|
||||
if (fs.isFile(jobOutput)) {
|
||||
Path finalOutputPath = getFinalPath(finalOutputDir, jobOutput, jobOutput);
|
||||
Path finalOutputPath =
|
||||
getFinalPath(finalOutputDir, jobOutput, origJobOutputPath);
|
||||
if (!fs.rename(jobOutput, finalOutputPath)) {
|
||||
if (!fs.delete(finalOutputPath, true)) {
|
||||
throw new IOException("Failed to delete earlier output of job");
|
||||
@ -145,14 +161,18 @@ private void moveJobOutputs(FileSystem fs,
|
||||
throw new IOException("Failed to save output of job");
|
||||
}
|
||||
}
|
||||
LOG.debug("Moved " + jobOutput + " to " + finalOutputPath);
|
||||
LOG.debug("Moved job output file from " + jobOutput + " to " +
|
||||
finalOutputPath);
|
||||
} else if (fs.getFileStatus(jobOutput).isDirectory()) {
|
||||
LOG.debug("Job output file " + jobOutput + " is a dir");
|
||||
FileStatus[] paths = fs.listStatus(jobOutput);
|
||||
Path finalOutputPath = getFinalPath(finalOutputDir, jobOutput, jobOutput);
|
||||
Path finalOutputPath =
|
||||
getFinalPath(finalOutputDir, jobOutput, origJobOutputPath);
|
||||
fs.mkdirs(finalOutputPath);
|
||||
LOG.debug("Creating dirs along job output path " + finalOutputPath);
|
||||
if (paths != null) {
|
||||
for (FileStatus path : paths) {
|
||||
moveJobOutputs(fs, finalOutputDir, path.getPath());
|
||||
moveJobOutputs(fs, origJobOutputPath, finalOutputDir, path.getPath());
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -233,6 +253,8 @@ private void moveTaskOutputs(TaskAttemptContext context,
|
||||
throws IOException {
|
||||
TaskAttemptID attemptId = context.getTaskAttemptID();
|
||||
context.progress();
|
||||
LOG.debug("Told to move taskoutput from " + taskOutput
|
||||
+ " to " + jobOutputDir);
|
||||
if (fs.isFile(taskOutput)) {
|
||||
Path finalOutputPath = getFinalPath(jobOutputDir, taskOutput,
|
||||
workPath);
|
||||
@ -248,9 +270,11 @@ private void moveTaskOutputs(TaskAttemptContext context,
|
||||
}
|
||||
LOG.debug("Moved " + taskOutput + " to " + finalOutputPath);
|
||||
} else if(fs.getFileStatus(taskOutput).isDirectory()) {
|
||||
LOG.debug("Taskoutput " + taskOutput + " is a dir");
|
||||
FileStatus[] paths = fs.listStatus(taskOutput);
|
||||
Path finalOutputPath = getFinalPath(jobOutputDir, taskOutput, workPath);
|
||||
fs.mkdirs(finalOutputPath);
|
||||
LOG.debug("Creating dirs along path " + finalOutputPath);
|
||||
if (paths != null) {
|
||||
for (FileStatus path : paths) {
|
||||
moveTaskOutputs(context, fs, jobOutputDir, path.getPath());
|
||||
@ -282,11 +306,16 @@ public void abortTask(TaskAttemptContext context) throws IOException {
|
||||
*/
|
||||
private Path getFinalPath(Path jobOutputDir, Path taskOutput,
|
||||
Path taskOutputPath) throws IOException {
|
||||
URI taskOutputUri = taskOutput.toUri();
|
||||
URI relativePath = taskOutputPath.toUri().relativize(taskOutputUri);
|
||||
URI taskOutputUri = taskOutput.makeQualified(outputFileSystem.getUri(),
|
||||
outputFileSystem.getWorkingDirectory()).toUri();
|
||||
URI taskOutputPathUri =
|
||||
taskOutputPath.makeQualified(
|
||||
outputFileSystem.getUri(),
|
||||
outputFileSystem.getWorkingDirectory()).toUri();
|
||||
URI relativePath = taskOutputPathUri.relativize(taskOutputUri);
|
||||
if (taskOutputUri == relativePath) {
|
||||
throw new IOException("Can not get the relative path: base = " +
|
||||
taskOutputPath + " child = " + taskOutput);
|
||||
taskOutputPathUri + " child = " + taskOutputUri);
|
||||
}
|
||||
if (relativePath.getPath().length() > 0) {
|
||||
return new Path(jobOutputDir, relativePath.getPath());
|
||||
@ -334,9 +363,12 @@ public void recoverTask(TaskAttemptContext context)
|
||||
|
||||
Path pathToRecover =
|
||||
new Path(outputPath, getJobAttemptBaseDirName(previousAttempt));
|
||||
LOG.debug("Trying to recover task from " + pathToRecover
|
||||
+ " into " + jobOutputPath);
|
||||
if (outputFileSystem.exists(pathToRecover)) {
|
||||
// Move the task outputs to their final place
|
||||
moveJobOutputs(outputFileSystem, jobOutputPath, pathToRecover);
|
||||
moveJobOutputs(outputFileSystem,
|
||||
pathToRecover, jobOutputPath, pathToRecover);
|
||||
LOG.info("Saved output of job to " + jobOutputPath);
|
||||
}
|
||||
}
|
||||
|
@ -25,13 +25,17 @@
|
||||
|
||||
import junit.framework.TestCase;
|
||||
|
||||
import org.apache.hadoop.fs.FileStatus;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.FileUtil;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.fs.RawLocalFileSystem;
|
||||
import org.apache.hadoop.io.LongWritable;
|
||||
import org.apache.hadoop.io.MapFile;
|
||||
import org.apache.hadoop.io.NullWritable;
|
||||
import org.apache.hadoop.io.Text;
|
||||
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
public class TestFileOutputCommitter extends TestCase {
|
||||
private static Path outDir = new Path(System.getProperty("test.build.data",
|
||||
@ -65,6 +69,20 @@ private void writeOutput(RecordWriter theRecordWriter,
|
||||
}
|
||||
}
|
||||
|
||||
private void writeMapFileOutput(RecordWriter theRecordWriter,
|
||||
TaskAttemptContext context) throws IOException, InterruptedException {
|
||||
try {
|
||||
int key = 0;
|
||||
for (int i = 0 ; i < 10; ++i) {
|
||||
key = i;
|
||||
Text val = (i%2 == 1) ? val1 : val2;
|
||||
theRecordWriter.write(new LongWritable(key),
|
||||
val);
|
||||
}
|
||||
} finally {
|
||||
theRecordWriter.close(null);
|
||||
}
|
||||
}
|
||||
|
||||
public void testRecovery() throws Exception {
|
||||
JobConf conf = new JobConf();
|
||||
@ -93,8 +111,6 @@ public void testRecovery() throws Exception {
|
||||
assertTrue((new File(jobTempDir1.toString()).exists()));
|
||||
validateContent(jobTempDir1);
|
||||
|
||||
|
||||
|
||||
//now while running the second app attempt,
|
||||
//recover the task output from first attempt
|
||||
JobConf conf2 = new JobConf(conf);
|
||||
@ -131,6 +147,29 @@ private void validateContent(Path dir) throws IOException {
|
||||
assertEquals(output, expectedOutput.toString());
|
||||
}
|
||||
|
||||
private void validateMapFileOutputContent(
|
||||
FileSystem fs, Path dir) throws IOException {
|
||||
// map output is a directory with index and data files
|
||||
Path expectedMapDir = new Path(dir, partFile);
|
||||
assert(fs.getFileStatus(expectedMapDir).isDirectory());
|
||||
FileStatus[] files = fs.listStatus(expectedMapDir);
|
||||
int fileCount = 0;
|
||||
boolean dataFileFound = false;
|
||||
boolean indexFileFound = false;
|
||||
for (FileStatus f : files) {
|
||||
if (f.isFile()) {
|
||||
++fileCount;
|
||||
if (f.getPath().getName().equals(MapFile.INDEX_FILE_NAME)) {
|
||||
indexFileFound = true;
|
||||
}
|
||||
else if (f.getPath().getName().equals(MapFile.DATA_FILE_NAME)) {
|
||||
dataFileFound = true;
|
||||
}
|
||||
}
|
||||
}
|
||||
assert(fileCount > 0);
|
||||
assert(dataFileFound && indexFileFound);
|
||||
}
|
||||
|
||||
public void testCommitter() throws Exception {
|
||||
JobConf conf = new JobConf();
|
||||
@ -159,6 +198,31 @@ public void testCommitter() throws Exception {
|
||||
FileUtil.fullyDelete(new File(outDir.toString()));
|
||||
}
|
||||
|
||||
public void testMapFileOutputCommitter() throws Exception {
|
||||
JobConf conf = new JobConf();
|
||||
FileOutputFormat.setOutputPath(conf, outDir);
|
||||
conf.set(JobContext.TASK_ATTEMPT_ID, attempt);
|
||||
JobContext jContext = new JobContextImpl(conf, taskID.getJobID());
|
||||
TaskAttemptContext tContext = new TaskAttemptContextImpl(conf, taskID);
|
||||
FileOutputCommitter committer = new FileOutputCommitter();
|
||||
|
||||
// setup
|
||||
committer.setupJob(jContext);
|
||||
committer.setupTask(tContext);
|
||||
|
||||
// write output
|
||||
MapFileOutputFormat theOutputFormat = new MapFileOutputFormat();
|
||||
RecordWriter theRecordWriter = theOutputFormat.getRecordWriter(null, conf, partFile, null);
|
||||
writeMapFileOutput(theRecordWriter, tContext);
|
||||
|
||||
// do commit
|
||||
committer.commitTask(tContext);
|
||||
committer.commitJob(jContext);
|
||||
|
||||
// validate output
|
||||
validateMapFileOutputContent(FileSystem.get(conf), outDir);
|
||||
FileUtil.fullyDelete(new File(outDir.toString()));
|
||||
}
|
||||
|
||||
public void testAbort() throws IOException, InterruptedException {
|
||||
JobConf conf = new JobConf();
|
||||
|
@ -26,10 +26,13 @@
|
||||
import junit.framework.TestCase;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.FileStatus;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.FileUtil;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.fs.RawLocalFileSystem;
|
||||
import org.apache.hadoop.io.LongWritable;
|
||||
import org.apache.hadoop.io.MapFile;
|
||||
import org.apache.hadoop.io.NullWritable;
|
||||
import org.apache.hadoop.io.Text;
|
||||
import org.apache.hadoop.mapreduce.Job;
|
||||
@ -75,6 +78,20 @@ private void writeOutput(RecordWriter theRecordWriter,
|
||||
}
|
||||
}
|
||||
|
||||
private void writeMapFileOutput(RecordWriter theRecordWriter,
|
||||
TaskAttemptContext context) throws IOException, InterruptedException {
|
||||
try {
|
||||
int key = 0;
|
||||
for (int i = 0 ; i < 10; ++i) {
|
||||
key = i;
|
||||
Text val = (i%2 == 1) ? val1 : val2;
|
||||
theRecordWriter.write(new LongWritable(key),
|
||||
val);
|
||||
}
|
||||
} finally {
|
||||
theRecordWriter.close(context);
|
||||
}
|
||||
}
|
||||
|
||||
public void testRecovery() throws Exception {
|
||||
Job job = Job.getInstance();
|
||||
@ -103,8 +120,6 @@ public void testRecovery() throws Exception {
|
||||
assertTrue((new File(jobTempDir1.toString()).exists()));
|
||||
validateContent(jobTempDir1);
|
||||
|
||||
|
||||
|
||||
//now while running the second app attempt,
|
||||
//recover the task output from first attempt
|
||||
Configuration conf2 = job.getConfiguration();
|
||||
@ -141,6 +156,29 @@ private void validateContent(Path dir) throws IOException {
|
||||
assertEquals(output, expectedOutput.toString());
|
||||
}
|
||||
|
||||
private void validateMapFileOutputContent(
|
||||
FileSystem fs, Path dir) throws IOException {
|
||||
// map output is a directory with index and data files
|
||||
Path expectedMapDir = new Path(dir, partFile);
|
||||
assert(fs.getFileStatus(expectedMapDir).isDirectory());
|
||||
FileStatus[] files = fs.listStatus(expectedMapDir);
|
||||
int fileCount = 0;
|
||||
boolean dataFileFound = false;
|
||||
boolean indexFileFound = false;
|
||||
for (FileStatus f : files) {
|
||||
if (f.isFile()) {
|
||||
++fileCount;
|
||||
if (f.getPath().getName().equals(MapFile.INDEX_FILE_NAME)) {
|
||||
indexFileFound = true;
|
||||
}
|
||||
else if (f.getPath().getName().equals(MapFile.DATA_FILE_NAME)) {
|
||||
dataFileFound = true;
|
||||
}
|
||||
}
|
||||
}
|
||||
assert(fileCount > 0);
|
||||
assert(dataFileFound && indexFileFound);
|
||||
}
|
||||
|
||||
public void testCommitter() throws Exception {
|
||||
Job job = Job.getInstance();
|
||||
@ -169,6 +207,32 @@ public void testCommitter() throws Exception {
|
||||
FileUtil.fullyDelete(new File(outDir.toString()));
|
||||
}
|
||||
|
||||
public void testMapFileOutputCommitter() throws Exception {
|
||||
Job job = Job.getInstance();
|
||||
FileOutputFormat.setOutputPath(job, outDir);
|
||||
Configuration conf = job.getConfiguration();
|
||||
conf.set(MRJobConfig.TASK_ATTEMPT_ID, attempt);
|
||||
JobContext jContext = new JobContextImpl(conf, taskID.getJobID());
|
||||
TaskAttemptContext tContext = new TaskAttemptContextImpl(conf, taskID);
|
||||
FileOutputCommitter committer = new FileOutputCommitter(outDir, tContext);
|
||||
|
||||
// setup
|
||||
committer.setupJob(jContext);
|
||||
committer.setupTask(tContext);
|
||||
|
||||
// write output
|
||||
MapFileOutputFormat theOutputFormat = new MapFileOutputFormat();
|
||||
RecordWriter theRecordWriter = theOutputFormat.getRecordWriter(tContext);
|
||||
writeMapFileOutput(theRecordWriter, tContext);
|
||||
|
||||
// do commit
|
||||
committer.commitTask(tContext);
|
||||
committer.commitJob(jContext);
|
||||
|
||||
// validate output
|
||||
validateMapFileOutputContent(FileSystem.get(job.getConfiguration()), outDir);
|
||||
FileUtil.fullyDelete(new File(outDir.toString()));
|
||||
}
|
||||
|
||||
public void testAbort() throws IOException, InterruptedException {
|
||||
Job job = Job.getInstance();
|
||||
|
Loading…
Reference in New Issue
Block a user