HDFS-3626. Creating file with invalid path can corrupt edit log. Contributed by Todd Lipcon.

git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1365801 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Todd Lipcon 2012-07-25 21:52:38 +00:00
parent 56b0912b6c
commit 972953bd77
9 changed files with 185 additions and 25 deletions

View File

@ -611,7 +611,7 @@ public static class HardLink extends org.apache.hadoop.fs.HardLink {
*/ */
public static int symLink(String target, String linkname) throws IOException{ public static int symLink(String target, String linkname) throws IOException{
String cmd = "ln -s " + target + " " + linkname; String cmd = "ln -s " + target + " " + linkname;
Process p = Runtime.getRuntime().exec(cmd, null); Process p = Runtime.getRuntime().exec(new String[]{"ln","-s",target,linkname}, null);
int returnVal = -1; int returnVal = -1;
try{ try{
returnVal = p.waitFor(); returnVal = p.waitFor();

View File

@ -139,7 +139,7 @@ public Path(String pathString) {
* Construct a path from a URI * Construct a path from a URI
*/ */
public Path(URI aUri) { public Path(URI aUri) {
uri = aUri; uri = aUri.normalize();
} }
/** Construct a Path from components. */ /** Construct a Path from components. */

View File

@ -61,7 +61,7 @@ private void toStringTest(String pathString) {
assertEquals(pathString, new Path(pathString).toString()); assertEquals(pathString, new Path(pathString).toString());
} }
public void testNormalize() { public void testNormalize() throws URISyntaxException {
assertEquals("", new Path(".").toString()); assertEquals("", new Path(".").toString());
assertEquals("..", new Path("..").toString()); assertEquals("..", new Path("..").toString());
assertEquals("/", new Path("/").toString()); assertEquals("/", new Path("/").toString());
@ -75,6 +75,8 @@ public void testNormalize() {
assertEquals("foo", new Path("foo/").toString()); assertEquals("foo", new Path("foo/").toString());
assertEquals("foo", new Path("foo//").toString()); assertEquals("foo", new Path("foo//").toString());
assertEquals("foo/bar", new Path("foo//bar").toString()); assertEquals("foo/bar", new Path("foo//bar").toString());
assertEquals("hdfs://foo/foo2/bar/baz/",
new Path(new URI("hdfs://foo//foo2///bar/baz///")).toString());
if (Path.WINDOWS) { if (Path.WINDOWS) {
assertEquals("c:/a/b", new Path("c:\\a\\b").toString()); assertEquals("c:/a/b", new Path("c:\\a\\b").toString());
} }

View File

@ -531,6 +531,8 @@ Branch-2 ( Unreleased changes )
HDFS-3720. hdfs.h must get packaged. (Colin Patrick McCabe via atm) HDFS-3720. hdfs.h must get packaged. (Colin Patrick McCabe via atm)
HDFS-3626. Creating file with invalid path can corrupt edit log (todd)
BREAKDOWN OF HDFS-3042 SUBTASKS BREAKDOWN OF HDFS-3042 SUBTASKS
HDFS-2185. HDFS portion of ZK-based FailoverController (todd) HDFS-2185. HDFS portion of ZK-based FailoverController (todd)

View File

@ -56,6 +56,7 @@
import org.apache.hadoop.net.NetUtils; import org.apache.hadoop.net.NetUtils;
import org.apache.hadoop.net.NodeBase; import org.apache.hadoop.net.NodeBase;
import org.apache.hadoop.security.UserGroupInformation; import org.apache.hadoop.security.UserGroupInformation;
import org.apache.hadoop.util.StringUtils;
import org.apache.commons.logging.Log; import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory; import org.apache.commons.logging.LogFactory;
@ -118,7 +119,7 @@ public boolean match(InetSocketAddress s) {
/** /**
* Whether the pathname is valid. Currently prohibits relative paths, * Whether the pathname is valid. Currently prohibits relative paths,
* and names which contain a ":" or "/" * names which contain a ":" or "//", or other non-canonical paths.
*/ */
public static boolean isValidName(String src) { public static boolean isValidName(String src) {
// Path must be absolute. // Path must be absolute.
@ -127,15 +128,22 @@ public static boolean isValidName(String src) {
} }
// Check for ".." "." ":" "/" // Check for ".." "." ":" "/"
StringTokenizer tokens = new StringTokenizer(src, Path.SEPARATOR); String[] components = StringUtils.split(src, '/');
while(tokens.hasMoreTokens()) { for (int i = 0; i < components.length; i++) {
String element = tokens.nextToken(); String element = components[i];
if (element.equals("..") || if (element.equals("..") ||
element.equals(".") || element.equals(".") ||
(element.indexOf(":") >= 0) || (element.indexOf(":") >= 0) ||
(element.indexOf("/") >= 0)) { (element.indexOf("/") >= 0)) {
return false; return false;
} }
// The string may start or end with a /, but not have
// "//" in the middle.
if (element.isEmpty() && i != components.length - 1 &&
i != 0) {
return false;
}
} }
return true; return true;
} }

View File

@ -230,8 +230,15 @@ INodeFileUnderConstruction addFile(String path,
// Always do an implicit mkdirs for parent directory tree. // Always do an implicit mkdirs for parent directory tree.
long modTime = now(); long modTime = now();
if (!mkdirs(new Path(path).getParent().toString(), permissions, true,
modTime)) { Path parent = new Path(path).getParent();
if (parent == null) {
// Trying to add "/" as a file - this path has no
// parent -- avoids an NPE below.
return null;
}
if (!mkdirs(parent.toString(), permissions, true, modTime)) {
return null; return null;
} }
INodeFileUnderConstruction newNode = new INodeFileUnderConstruction( INodeFileUnderConstruction newNode = new INodeFileUnderConstruction(

View File

@ -17,8 +17,7 @@
*/ */
package org.apache.hadoop.hdfs; package org.apache.hadoop.hdfs;
import static org.junit.Assert.assertFalse; import static org.junit.Assert.*;
import static org.junit.Assert.assertTrue;
import java.io.DataOutputStream; import java.io.DataOutputStream;
import java.io.FileNotFoundException; import java.io.FileNotFoundException;
@ -26,33 +25,38 @@
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.InvalidPathException;
import org.apache.hadoop.fs.ParentNotDirectoryException; import org.apache.hadoop.fs.ParentNotDirectoryException;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.permission.FsPermission; import org.apache.hadoop.fs.permission.FsPermission;
import org.apache.hadoop.hdfs.server.protocol.NamenodeProtocol;
import org.apache.hadoop.hdfs.server.protocol.NamenodeProtocols;
import org.apache.hadoop.test.GenericTestUtils;
import org.apache.hadoop.util.Time; import org.apache.hadoop.util.Time;
import org.junit.Test; import org.junit.Test;
/** /**
* This class tests that the DFS command mkdirs cannot create subdirectories * This class tests that the DFS command mkdirs only creates valid
* from a file when passed an illegal path. HADOOP-281. * directories, and generally behaves as expected.
*/ */
public class TestDFSMkdirs { public class TestDFSMkdirs {
private Configuration conf = new HdfsConfiguration();
private void writeFile(FileSystem fileSys, Path name) throws IOException { private static final String[] NON_CANONICAL_PATHS = new String[] {
DataOutputStream stm = fileSys.create(name); "//test1",
stm.writeBytes("wchien"); "/test2/..",
stm.close(); "/test2//bar",
} "/test2/../test4",
"/test5/."
};
/** /**
* Tests mkdirs can create a directory that does not exist and will * Tests mkdirs can create a directory that does not exist and will
* not create a subdirectory off a file. * not create a subdirectory off a file. Regression test for HADOOP-281.
*/ */
@Test @Test
public void testDFSMkdirs() throws IOException { public void testDFSMkdirs() throws IOException {
Configuration conf = new HdfsConfiguration(); MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(1).build();
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
FileSystem fileSys = cluster.getFileSystem(); FileSystem fileSys = cluster.getFileSystem();
try { try {
// First create a new directory with mkdirs // First create a new directory with mkdirs
@ -63,7 +67,7 @@ public void testDFSMkdirs() throws IOException {
// Second, create a file in that directory. // Second, create a file in that directory.
Path myFile = new Path("/test/mkdirs/myFile"); Path myFile = new Path("/test/mkdirs/myFile");
writeFile(fileSys, myFile); DFSTestUtil.writeFile(fileSys, myFile, "hello world");
// Third, use mkdir to create a subdirectory off of that file, // Third, use mkdir to create a subdirectory off of that file,
// and check that it fails. // and check that it fails.
@ -99,7 +103,7 @@ public void testMkdir() throws IOException {
// Create a dir when parent dir exists as a file, should fail // Create a dir when parent dir exists as a file, should fail
IOException expectedException = null; IOException expectedException = null;
String filePath = "/mkdir-file-" + Time.now(); String filePath = "/mkdir-file-" + Time.now();
writeFile(dfs, new Path(filePath)); DFSTestUtil.writeFile(dfs, new Path(filePath), "hello world");
try { try {
dfs.mkdir(new Path(filePath + "/mkdir"), FsPermission.getDefault()); dfs.mkdir(new Path(filePath + "/mkdir"), FsPermission.getDefault());
} catch (IOException e) { } catch (IOException e) {
@ -126,4 +130,29 @@ public void testMkdir() throws IOException {
cluster.shutdown(); cluster.shutdown();
} }
} }
/**
* Regression test for HDFS-3626. Creates a file using a non-canonical path
* (i.e. with extra slashes between components) and makes sure that the NN
* rejects it.
*/
@Test
public void testMkdirRpcNonCanonicalPath() throws IOException {
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(0).build();
try {
NamenodeProtocols nnrpc = cluster.getNameNodeRpc();
for (String pathStr : NON_CANONICAL_PATHS) {
try {
nnrpc.mkdirs(pathStr, new FsPermission((short)0755), true);
fail("Did not fail when called with a non-canonicalized path: "
+ pathStr);
} catch (InvalidPathException ipe) {
// expected
}
}
} finally {
cluster.shutdown();
}
}
} }

View File

@ -620,4 +620,12 @@ public void testGetNNUris() throws Exception {
assertEquals(1, uris.size()); assertEquals(1, uris.size());
assertTrue(uris.contains(new URI("hdfs://" + NN1_SRVC_ADDR))); assertTrue(uris.contains(new URI("hdfs://" + NN1_SRVC_ADDR)));
} }
@Test
public void testIsValidName() {
assertFalse(DFSUtil.isValidName("/foo/../bar"));
assertFalse(DFSUtil.isValidName("/foo//bar"));
assertTrue(DFSUtil.isValidName("/"));
assertTrue(DFSUtil.isValidName("/bar/"));
}
} }

View File

@ -41,6 +41,8 @@
import java.io.FileReader; import java.io.FileReader;
import java.io.IOException; import java.io.IOException;
import java.net.InetSocketAddress; import java.net.InetSocketAddress;
import java.net.URI;
import java.net.URISyntaxException;
import java.net.UnknownHostException; import java.net.UnknownHostException;
import java.util.EnumSet; import java.util.EnumSet;
@ -53,6 +55,7 @@
import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.FsServerDefaults; import org.apache.hadoop.fs.FsServerDefaults;
import org.apache.hadoop.fs.InvalidPathException;
import org.apache.hadoop.fs.ParentNotDirectoryException; import org.apache.hadoop.fs.ParentNotDirectoryException;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.permission.FsPermission; import org.apache.hadoop.fs.permission.FsPermission;
@ -69,7 +72,10 @@
import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsDatasetSpi; import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsDatasetSpi;
import org.apache.hadoop.hdfs.server.namenode.FSNamesystem; import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
import org.apache.hadoop.hdfs.server.namenode.LeaseManager; import org.apache.hadoop.hdfs.server.namenode.LeaseManager;
import org.apache.hadoop.hdfs.server.protocol.NamenodeProtocols;
import org.apache.hadoop.io.EnumSetWritable;
import org.apache.hadoop.io.IOUtils; import org.apache.hadoop.io.IOUtils;
import org.apache.hadoop.test.GenericTestUtils;
import org.apache.hadoop.util.Time; import org.apache.hadoop.util.Time;
import org.apache.log4j.Level; import org.apache.log4j.Level;
import org.junit.Ignore; import org.junit.Ignore;
@ -94,6 +100,15 @@ public class TestFileCreation {
static final int fileSize = numBlocks * blockSize + 1; static final int fileSize = numBlocks * blockSize + 1;
boolean simulatedStorage = false; boolean simulatedStorage = false;
private static final String[] NON_CANONICAL_PATHS = new String[] {
"//foo",
"///foo2",
"//dir//file",
"////test2/file",
"/dir/./file2",
"/dir/../file3"
};
// creates a file but does not close it // creates a file but does not close it
public static FSDataOutputStream createFile(FileSystem fileSys, Path name, int repl) public static FSDataOutputStream createFile(FileSystem fileSys, Path name, int repl)
throws IOException { throws IOException {
@ -988,4 +1003,93 @@ public void testFsCloseAfterClusterShutdown() throws IOException {
} }
} }
} }
/**
* Regression test for HDFS-3626. Creates a file using a non-canonical path
* (i.e. with extra slashes between components) and makes sure that the NN
* can properly restart.
*
* This test RPCs directly to the NN, to ensure that even an old client
* which passes an invalid path won't cause corrupt edits.
*/
@Test
public void testCreateNonCanonicalPathAndRestartRpc() throws Exception {
doCreateTest(CreationMethod.DIRECT_NN_RPC);
}
/**
* Another regression test for HDFS-3626. This one creates files using
* a Path instantiated from a string object.
*/
@Test
public void testCreateNonCanonicalPathAndRestartFromString()
throws Exception {
doCreateTest(CreationMethod.PATH_FROM_STRING);
}
/**
* Another regression test for HDFS-3626. This one creates files using
* a Path instantiated from a URI object.
*/
@Test
public void testCreateNonCanonicalPathAndRestartFromUri()
throws Exception {
doCreateTest(CreationMethod.PATH_FROM_URI);
}
private static enum CreationMethod {
DIRECT_NN_RPC,
PATH_FROM_URI,
PATH_FROM_STRING
};
private void doCreateTest(CreationMethod method) throws Exception {
Configuration conf = new HdfsConfiguration();
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf)
.numDataNodes(1).build();
try {
FileSystem fs = cluster.getFileSystem();
NamenodeProtocols nnrpc = cluster.getNameNodeRpc();
for (String pathStr : NON_CANONICAL_PATHS) {
System.out.println("Creating " + pathStr + " by " + method);
switch (method) {
case DIRECT_NN_RPC:
try {
nnrpc.create(pathStr, new FsPermission((short)0755), "client",
new EnumSetWritable<CreateFlag>(EnumSet.of(CreateFlag.CREATE)),
true, (short)1, 128*1024*1024L);
fail("Should have thrown exception when creating '"
+ pathStr + "'" + " by " + method);
} catch (InvalidPathException ipe) {
// When we create by direct NN RPC, the NN just rejects the
// non-canonical paths, rather than trying to normalize them.
// So, we expect all of them to fail.
}
break;
case PATH_FROM_URI:
case PATH_FROM_STRING:
// Unlike the above direct-to-NN case, we expect these to succeed,
// since the Path constructor should normalize the path.
Path p;
if (method == CreationMethod.PATH_FROM_URI) {
p = new Path(new URI(fs.getUri() + pathStr));
} else {
p = new Path(fs.getUri() + pathStr);
}
FSDataOutputStream stm = fs.create(p);
IOUtils.closeStream(stm);
break;
default:
throw new AssertionError("bad method: " + method);
}
}
cluster.restartNameNode();
} finally {
cluster.shutdown();
}
}
} }