HDFS-7167. NPE while running Mover if the given path is for a file. Contributed by Jing Zhao.

This commit is contained in:
Jing Zhao 2014-09-30 11:13:35 -07:00
parent ea32a66f7d
commit cdf1af0e5a
3 changed files with 50 additions and 22 deletions

View File

@ -772,6 +772,11 @@ Release 2.6.0 - UNRELEASED
HDFS-7157. Using Time.now() for recording start/end time of reconfiguration HDFS-7157. Using Time.now() for recording start/end time of reconfiguration
tasks (Lei Xu via cmccabe) tasks (Lei Xu via cmccabe)
HDFS-6664. HDFS permissions guide documentation states incorrect default
group mapping class. (Ray Chiang via aw)
HDFS-4227. Document dfs.namenode.resource.* (Daisuke Kobayashi via aw)
BREAKDOWN OF HDFS-6134 AND HADOOP-10150 SUBTASKS AND RELATED JIRAS BREAKDOWN OF HDFS-6134 AND HADOOP-10150 SUBTASKS AND RELATED JIRAS
HDFS-6387. HDFS CLI admin tool for creating & deleting an HDFS-6387. HDFS CLI admin tool for creating & deleting an
@ -1003,10 +1008,7 @@ Release 2.6.0 - UNRELEASED
HDFS-7140. Add a tool to list all the existing block storage policies. HDFS-7140. Add a tool to list all the existing block storage policies.
(jing9) (jing9)
HDFS-6664. HDFS permissions guide documentation states incorrect default HDFS-7167. NPE while running Mover if the given path is for a file. (jing9)
group mapping class. (Ray Chiang via aw)
HDFS-4227. Document dfs.namenode.resource.* (Daisuke Kobayashi via aw)
Release 2.5.1 - 2014-09-05 Release 2.5.1 - 2014-09-05

View File

@ -252,14 +252,9 @@ private boolean isSnapshotPathInCurrent(String path) throws IOException {
*/ */
private boolean processNamespace() { private boolean processNamespace() {
getSnapshottableDirs(); getSnapshottableDirs();
boolean hasRemaining = true; boolean hasRemaining = false;
try { for (Path target : targetPaths) {
for (Path target : targetPaths) { hasRemaining |= processPath(target.toUri().getPath());
hasRemaining = processDirRecursively("", dfs.getFileInfo(target
.toUri().getPath()));
}
} catch (IOException e) {
LOG.warn("Failed to get root directory status. Ignore and continue.", e);
} }
// wait for pending move to finish and retry the failed migration // wait for pending move to finish and retry the failed migration
hasRemaining |= Dispatcher.waitForMoveCompletion(storages.targets.values()); hasRemaining |= Dispatcher.waitForMoveCompletion(storages.targets.values());
@ -270,7 +265,7 @@ private boolean processNamespace() {
* @return whether there is still remaing migration work for the next * @return whether there is still remaing migration work for the next
* round * round
*/ */
private boolean processChildrenList(String fullPath) { private boolean processPath(String fullPath) {
boolean hasRemaining = false; boolean hasRemaining = false;
for (byte[] lastReturnedName = HdfsFileStatus.EMPTY_NAME;;) { for (byte[] lastReturnedName = HdfsFileStatus.EMPTY_NAME;;) {
final DirectoryListing children; final DirectoryListing children;
@ -285,7 +280,7 @@ private boolean processChildrenList(String fullPath) {
return hasRemaining; return hasRemaining;
} }
for (HdfsFileStatus child : children.getPartialListing()) { for (HdfsFileStatus child : children.getPartialListing()) {
hasRemaining |= processDirRecursively(fullPath, child); hasRemaining |= processRecursively(fullPath, child);
} }
if (children.hasMore()) { if (children.hasMore()) {
lastReturnedName = children.getLastName(); lastReturnedName = children.getLastName();
@ -296,8 +291,7 @@ private boolean processChildrenList(String fullPath) {
} }
/** @return whether the migration requires next round */ /** @return whether the migration requires next round */
private boolean processDirRecursively(String parent, private boolean processRecursively(String parent, HdfsFileStatus status) {
HdfsFileStatus status) {
String fullPath = status.getFullName(parent); String fullPath = status.getFullName(parent);
boolean hasRemaining = false; boolean hasRemaining = false;
if (status.isDir()) { if (status.isDir()) {
@ -305,11 +299,11 @@ private boolean processDirRecursively(String parent,
fullPath = fullPath + Path.SEPARATOR; fullPath = fullPath + Path.SEPARATOR;
} }
hasRemaining = processChildrenList(fullPath); hasRemaining = processPath(fullPath);
// process snapshots if this is a snapshottable directory // process snapshots if this is a snapshottable directory
if (snapshottableDirs.contains(fullPath)) { if (snapshottableDirs.contains(fullPath)) {
final String dirSnapshot = fullPath + HdfsConstants.DOT_SNAPSHOT_DIR; final String dirSnapshot = fullPath + HdfsConstants.DOT_SNAPSHOT_DIR;
hasRemaining |= processChildrenList(dirSnapshot); hasRemaining |= processPath(dirSnapshot);
} }
} else if (!status.isSymlink()) { // file } else if (!status.isSymlink()) { // file
try { try {

View File

@ -17,7 +17,6 @@
*/ */
package org.apache.hadoop.hdfs.server.mover; package org.apache.hadoop.hdfs.server.mover;
import java.io.File;
import java.io.IOException; import java.io.IOException;
import java.net.URI; import java.net.URI;
import java.util.ArrayList; import java.util.ArrayList;
@ -27,12 +26,10 @@
import java.util.List; import java.util.List;
import java.util.Map; import java.util.Map;
import com.google.common.base.Joiner;
import org.apache.commons.logging.Log; import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory; import org.apache.commons.logging.LogFactory;
import org.apache.commons.logging.impl.Log4JLogger; import org.apache.commons.logging.impl.Log4JLogger;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.conf.ReconfigurationException;
import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FSDataInputStream;
import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
@ -46,7 +43,6 @@
import org.apache.hadoop.hdfs.MiniDFSCluster; import org.apache.hadoop.hdfs.MiniDFSCluster;
import org.apache.hadoop.hdfs.StorageType; import org.apache.hadoop.hdfs.StorageType;
import org.apache.hadoop.hdfs.protocol.DirectoryListing; import org.apache.hadoop.hdfs.protocol.DirectoryListing;
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
import org.apache.hadoop.hdfs.protocol.HdfsFileStatus; import org.apache.hadoop.hdfs.protocol.HdfsFileStatus;
import org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus; import org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus;
import org.apache.hadoop.hdfs.protocol.LocatedBlock; import org.apache.hadoop.hdfs.protocol.LocatedBlock;
@ -514,6 +510,42 @@ static void banner(String string) {
"==================================================\n\n"); "==================================================\n\n");
} }
/**
* Run Mover with arguments specifying files and directories
*/
@Test
public void testMoveSpecificPaths() throws Exception {
LOG.info("testMoveSpecificPaths");
final Path foo = new Path("/foo");
final Path barFile = new Path(foo, "bar");
final Path foo2 = new Path("/foo2");
final Path bar2File = new Path(foo2, "bar2");
Map<Path, BlockStoragePolicy> policyMap = Maps.newHashMap();
policyMap.put(foo, COLD);
policyMap.put(foo2, WARM);
NamespaceScheme nsScheme = new NamespaceScheme(Arrays.asList(foo, foo2),
Arrays.asList(barFile, bar2File), BLOCK_SIZE, null, policyMap);
ClusterScheme clusterScheme = new ClusterScheme(DEFAULT_CONF,
NUM_DATANODES, REPL, genStorageTypes(NUM_DATANODES), null);
MigrationTest test = new MigrationTest(clusterScheme, nsScheme);
test.setupCluster();
try {
test.prepareNamespace();
test.setStoragePolicy();
Map<URI, List<Path>> map = Mover.Cli.getNameNodePathsToMove(test.conf,
"-p", "/foo/bar", "/foo2");
int result = Mover.run(map, test.conf);
Assert.assertEquals(ExitStatus.SUCCESS.getExitCode(), result);
Thread.sleep(5000);
test.verify(true);
} finally {
test.shutdownCluster();
}
}
/** /**
* Move an open file into archival storage * Move an open file into archival storage
*/ */