MAPREDUCE-5756. CombineFileInputFormat.getSplits() including directories in its results. Contributed by Jason Dere
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1612400 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
0c1469ece3
commit
c2174a5536
@ -172,6 +172,9 @@ Release 2.6.0 - UNRELEASED
|
|||||||
MAPREDUCE-5957. AM throws ClassNotFoundException with job classloader
|
MAPREDUCE-5957. AM throws ClassNotFoundException with job classloader
|
||||||
enabled if custom output format/committer is used (Sangjin Lee via jlowe)
|
enabled if custom output format/committer is used (Sangjin Lee via jlowe)
|
||||||
|
|
||||||
|
MAPREDUCE-5756. CombineFileInputFormat.getSplits() including directories
|
||||||
|
in its results (Jason Dere via jlowe)
|
||||||
|
|
||||||
Release 2.5.0 - UNRELEASED
|
Release 2.5.0 - UNRELEASED
|
||||||
|
|
||||||
INCOMPATIBLE CHANGES
|
INCOMPATIBLE CHANGES
|
||||||
|
@ -579,7 +579,7 @@ static class OneFileInfo {
|
|||||||
blocks = new OneBlockInfo[0];
|
blocks = new OneBlockInfo[0];
|
||||||
} else {
|
} else {
|
||||||
|
|
||||||
if(locations.length == 0) {
|
if(locations.length == 0 && !stat.isDirectory()) {
|
||||||
locations = new BlockLocation[] { new BlockLocation() };
|
locations = new BlockLocation[] { new BlockLocation() };
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1274,6 +1274,61 @@ public void testForEmptyFile() throws Exception {
|
|||||||
fileSys.delete(file.getParent(), true);
|
fileSys.delete(file.getParent(), true);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Test that directories do not get included as part of getSplits()
|
||||||
|
*/
|
||||||
|
@Test
|
||||||
|
public void testGetSplitsWithDirectory() throws Exception {
|
||||||
|
MiniDFSCluster dfs = null;
|
||||||
|
try {
|
||||||
|
Configuration conf = new Configuration();
|
||||||
|
dfs = new MiniDFSCluster.Builder(conf).racks(rack1).hosts(hosts1)
|
||||||
|
.build();
|
||||||
|
dfs.waitActive();
|
||||||
|
|
||||||
|
dfs = new MiniDFSCluster.Builder(conf).racks(rack1).hosts(hosts1)
|
||||||
|
.build();
|
||||||
|
dfs.waitActive();
|
||||||
|
|
||||||
|
FileSystem fileSys = dfs.getFileSystem();
|
||||||
|
|
||||||
|
// Set up the following directory structure:
|
||||||
|
// /dir1/: directory
|
||||||
|
// /dir1/file: regular file
|
||||||
|
// /dir1/dir2/: directory
|
||||||
|
Path dir1 = new Path("/dir1");
|
||||||
|
Path file = new Path("/dir1/file1");
|
||||||
|
Path dir2 = new Path("/dir1/dir2");
|
||||||
|
if (!fileSys.mkdirs(dir1)) {
|
||||||
|
throw new IOException("Mkdirs failed to create " + dir1.toString());
|
||||||
|
}
|
||||||
|
FSDataOutputStream out = fileSys.create(file);
|
||||||
|
out.write(new byte[0]);
|
||||||
|
out.close();
|
||||||
|
if (!fileSys.mkdirs(dir2)) {
|
||||||
|
throw new IOException("Mkdirs failed to create " + dir2.toString());
|
||||||
|
}
|
||||||
|
|
||||||
|
// split it using a CombinedFile input format
|
||||||
|
DummyInputFormat inFormat = new DummyInputFormat();
|
||||||
|
Job job = Job.getInstance(conf);
|
||||||
|
FileInputFormat.setInputPaths(job, "/dir1");
|
||||||
|
List<InputSplit> splits = inFormat.getSplits(job);
|
||||||
|
|
||||||
|
// directories should be omitted from getSplits() - we should only see file1 and not dir2
|
||||||
|
assertEquals(1, splits.size());
|
||||||
|
CombineFileSplit fileSplit = (CombineFileSplit) splits.get(0);
|
||||||
|
assertEquals(1, fileSplit.getNumPaths());
|
||||||
|
assertEquals(file.getName(), fileSplit.getPath(0).getName());
|
||||||
|
assertEquals(0, fileSplit.getOffset(0));
|
||||||
|
assertEquals(0, fileSplit.getLength(0));
|
||||||
|
} finally {
|
||||||
|
if (dfs != null) {
|
||||||
|
dfs.shutdown();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Test when input files are from non-default file systems
|
* Test when input files are from non-default file systems
|
||||||
*/
|
*/
|
||||||
|
Loading…
Reference in New Issue
Block a user