diff --git a/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DFSClient.java b/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DFSClient.java index 1f6022ca35..8acda61199 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DFSClient.java +++ b/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DFSClient.java @@ -2883,12 +2883,9 @@ private void initThreadsNumForStripedReads(int numThreads) { } synchronized (DFSClient.class) { if (STRIPED_READ_THREAD_POOL == null) { - // Only after thread pool is fully constructed then save it to - // volatile field. - ThreadPoolExecutor threadPool = DFSUtilClient.getThreadPoolExecutor(1, + STRIPED_READ_THREAD_POOL = DFSUtilClient.getThreadPoolExecutor(1, numThreads, 60, "StripedRead-", true); - threadPool.allowCoreThreadTimeOut(true); - STRIPED_READ_THREAD_POOL = threadPool; + STRIPED_READ_THREAD_POOL.allowCoreThreadTimeOut(true); } } } diff --git a/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/server/protocol/SlowDiskReports.java b/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/server/protocol/SlowDiskReports.java index 496389a1dd..8095c2a690 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/server/protocol/SlowDiskReports.java +++ b/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/server/protocol/SlowDiskReports.java @@ -101,9 +101,8 @@ public boolean equals(Object o) { } boolean areEqual; - for (Map.Entry> entry : this.slowDisks - .entrySet()) { - if (!entry.getValue().equals(that.slowDisks.get(entry.getKey()))) { + for (String disk : this.slowDisks.keySet()) { + if (!this.slowDisks.get(disk).equals(that.slowDisks.get(disk))) { return false; } } diff --git a/hadoop-hdfs-project/hadoop-hdfs/dev-support/findbugsExcludeFile.xml b/hadoop-hdfs-project/hadoop-hdfs/dev-support/findbugsExcludeFile.xml index 9270990e0b..be54efb242 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/dev-support/findbugsExcludeFile.xml +++ b/hadoop-hdfs-project/hadoop-hdfs/dev-support/findbugsExcludeFile.xml @@ -252,9 +252,4 @@ - - - - - diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/JournalNode.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/JournalNode.java index 6056e34d8f..af7a84f672 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/JournalNode.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/JournalNode.java @@ -299,18 +299,14 @@ public boolean accept(File file) { return file.isDirectory(); } }); - - if (journalDirs != null) { - for (File journalDir : journalDirs) { - String jid = journalDir.getName(); - if (!status.containsKey(jid)) { - Map jMap = new HashMap(); - jMap.put("Formatted", "true"); - status.put(jid, jMap); - } + for (File journalDir : journalDirs) { + String jid = journalDir.getName(); + if (!status.containsKey(jid)) { + Map jMap = new HashMap(); + jMap.put("Formatted", "true"); + status.put(jid, jMap); } } - return JSON.toString(status); } diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/common/HdfsServerConstants.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/common/HdfsServerConstants.java index ab50eb4012..c3098f3cff 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/common/HdfsServerConstants.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/common/HdfsServerConstants.java @@ -188,10 +188,8 @@ public NamenodeRole toNodeRole() { return NamenodeRole.NAMENODE; } } - + public void setClusterId(String cid) { - Preconditions.checkState(this == UPGRADE || this == UPGRADEONLY - || this == FORMAT); clusterId = cid; } @@ -216,7 +214,6 @@ public MetaRecoveryContext createRecoveryContext() { } public void setForce(int force) { - Preconditions.checkState(this == RECOVER); this.force = force; } @@ -229,7 +226,6 @@ public boolean getForceFormat() { } public void setForceFormat(boolean force) { - Preconditions.checkState(this == FORMAT); isForceFormat = force; } @@ -238,7 +234,6 @@ public boolean getInteractiveFormat() { } public void setInteractiveFormat(boolean interactive) { - Preconditions.checkState(this == FORMAT); isInteractiveFormat = interactive; } diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataStorage.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataStorage.java index 6d6e96a699..9a7108124f 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataStorage.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataStorage.java @@ -1336,14 +1336,10 @@ public boolean accept(File dir, String name) { return name.startsWith(BLOCK_SUBDIR_PREFIX); } }); - - if (otherNames != null) { - for (int i = 0; i < otherNames.length; i++) { - linkBlocksHelper(new File(from, otherNames[i]), - new File(to, otherNames[i]), oldLV, hl, upgradeToIdBasedLayout, - blockRoot, idBasedLayoutSingleLinks); - } - } + for(int i = 0; i < otherNames.length; i++) + linkBlocksHelper(new File(from, otherNames[i]), + new File(to, otherNames[i]), oldLV, hl, upgradeToIdBasedLayout, + blockRoot, idBasedLayoutSingleLinks); } /** diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NNStorageRetentionManager.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NNStorageRetentionManager.java index 2a83541d38..98b7e9ac52 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NNStorageRetentionManager.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NNStorageRetentionManager.java @@ -255,27 +255,24 @@ public boolean accept(File dir, String name) { }); // Check whether there is any work to do. - if (filesInStorage != null - && filesInStorage.length <= numCheckpointsToRetain) { + if (filesInStorage.length <= numCheckpointsToRetain) { return; } // Create a sorted list of txids from the file names. TreeSet sortedTxIds = new TreeSet(); - if (filesInStorage != null) { - for (String fName : filesInStorage) { - // Extract the transaction id from the file name. - long fTxId; - try { - fTxId = Long.parseLong(fName.substring(oivImagePrefix.length() + 1)); - } catch (NumberFormatException nfe) { - // This should not happen since we have already filtered it. - // Log and continue. - LOG.warn("Invalid file name. Skipping " + fName); - continue; - } - sortedTxIds.add(Long.valueOf(fTxId)); + for (String fName : filesInStorage) { + // Extract the transaction id from the file name. + long fTxId; + try { + fTxId = Long.parseLong(fName.substring(oivImagePrefix.length() + 1)); + } catch (NumberFormatException nfe) { + // This should not happen since we have already filtered it. + // Log and continue. + LOG.warn("Invalid file name. Skipping " + fName); + continue; } + sortedTxIds.add(Long.valueOf(fTxId)); } int numFilesToDelete = sortedTxIds.size() - numCheckpointsToRetain; diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DFSAdmin.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DFSAdmin.java index b771ff073d..06f408d0d3 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DFSAdmin.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DFSAdmin.java @@ -1977,7 +1977,7 @@ public int run(String[] argv) throws Exception { return exitCode; } } else if ("-report".equals(cmd)) { - if (argv.length > 6) { + if (argv.length < 1) { printUsage(cmd); return exitCode; } @@ -2007,7 +2007,7 @@ public int run(String[] argv) throws Exception { return exitCode; } } else if (RollingUpgradeCommand.matches(cmd)) { - if (argv.length > 2) { + if (argv.length < 1 || argv.length > 2) { printUsage(cmd); return exitCode; } @@ -2082,7 +2082,7 @@ public int run(String[] argv) throws Exception { return exitCode; } } else if ("-triggerBlockReport".equals(cmd)) { - if ((argv.length != 2) && (argv.length != 3)) { + if (argv.length < 1) { printUsage(cmd); return exitCode; } diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/ImageLoaderCurrent.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/ImageLoaderCurrent.java index 2e2eaf4e4d..f2c7427b51 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/ImageLoaderCurrent.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/ImageLoaderCurrent.java @@ -722,13 +722,9 @@ private void processINode(DataInputStream in, ImageVisitor v, if (supportSnapshot && supportInodeId) { dirNodeMap.put(inodeId, pathName); } - - v.visit(ImageElement.NS_QUOTA, in.readLong()); - if (NameNodeLayoutVersion.supports(Feature.DISKSPACE_QUOTA, - imageVersion)) { - v.visit(ImageElement.DS_QUOTA, in.readLong()); - } - + v.visit(ImageElement.NS_QUOTA, numBlocks == -1 ? in.readLong() : -1); + if (NameNodeLayoutVersion.supports(Feature.DISKSPACE_QUOTA, imageVersion)) + v.visit(ImageElement.DS_QUOTA, numBlocks == -1 ? in.readLong() : -1); if (supportSnapshot) { boolean snapshottable = in.readBoolean(); if (!snapshottable) {