HDFS-3187. Upgrade guava to 11.0.2. Contributed by Todd Lipcon.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1309181 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
84ff2d6d06
commit
9972a9633d
@ -306,6 +306,8 @@ Release 2.0.0 - UNRELEASED
|
|||||||
|
|
||||||
HDFS-3120. Enable hsync and hflush by default. (eli)
|
HDFS-3120. Enable hsync and hflush by default. (eli)
|
||||||
|
|
||||||
|
HDFS-3187. Upgrade guava to 11.0.2 (todd)
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
HDFS-3024. Improve performance of stringification in addStoredBlock (todd)
|
HDFS-3024. Improve performance of stringification in addStoredBlock (todd)
|
||||||
|
@ -731,7 +731,9 @@ private void copyNameDirs(Collection<URI> srcDirs, Collection<URI> dstDirs,
|
|||||||
Preconditions.checkArgument(!dstDir.equals(srcDir));
|
Preconditions.checkArgument(!dstDir.equals(srcDir));
|
||||||
File dstDirF = new File(dstDir);
|
File dstDirF = new File(dstDir);
|
||||||
if (dstDirF.exists()) {
|
if (dstDirF.exists()) {
|
||||||
Files.deleteRecursively(dstDirF);
|
if (!FileUtil.fullyDelete(dstDirF)) {
|
||||||
|
throw new IOException("Unable to delete: " + dstDirF);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
LOG.info("Copying namedir from primary node dir "
|
LOG.info("Copying namedir from primary node dir "
|
||||||
+ srcDir + " to " + dstDir);
|
+ srcDir + " to " + dstDir);
|
||||||
|
@ -37,6 +37,7 @@
|
|||||||
import org.apache.commons.lang.StringUtils;
|
import org.apache.commons.lang.StringUtils;
|
||||||
import org.apache.commons.logging.Log;
|
import org.apache.commons.logging.Log;
|
||||||
import org.apache.commons.logging.LogFactory;
|
import org.apache.commons.logging.LogFactory;
|
||||||
|
import org.apache.hadoop.fs.FileUtil;
|
||||||
import org.apache.hadoop.fs.permission.FsPermission;
|
import org.apache.hadoop.fs.permission.FsPermission;
|
||||||
import org.apache.hadoop.fs.permission.PermissionStatus;
|
import org.apache.hadoop.fs.permission.PermissionStatus;
|
||||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||||
@ -181,7 +182,9 @@ public static FSImageTransactionalStorageInspector inspectStorageDirectory(
|
|||||||
public static FSEditLog createStandaloneEditLog(File logDir)
|
public static FSEditLog createStandaloneEditLog(File logDir)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
assertTrue(logDir.mkdirs() || logDir.exists());
|
assertTrue(logDir.mkdirs() || logDir.exists());
|
||||||
Files.deleteDirectoryContents(logDir);
|
if (!FileUtil.fullyDeleteContents(logDir)) {
|
||||||
|
throw new IOException("Unable to delete contents of " + logDir);
|
||||||
|
}
|
||||||
NNStorage storage = Mockito.mock(NNStorage.class);
|
NNStorage storage = Mockito.mock(NNStorage.class);
|
||||||
StorageDirectory sd
|
StorageDirectory sd
|
||||||
= FSImageTestUtil.mockStorageDirectory(logDir, NameNodeDirType.EDITS);
|
= FSImageTestUtil.mockStorageDirectory(logDir, NameNodeDirType.EDITS);
|
||||||
|
@ -253,7 +253,7 @@
|
|||||||
<dependency>
|
<dependency>
|
||||||
<groupId>com.google.guava</groupId>
|
<groupId>com.google.guava</groupId>
|
||||||
<artifactId>guava</artifactId>
|
<artifactId>guava</artifactId>
|
||||||
<version>r09</version>
|
<version>11.0.2</version>
|
||||||
</dependency>
|
</dependency>
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>commons-cli</groupId>
|
<groupId>commons-cli</groupId>
|
||||||
|
Loading…
Reference in New Issue
Block a user