diff --git a/hadoop-cloud-storage-project/hadoop-cos/dev-support/findbugs-exclude.xml b/hadoop-cloud-storage-project/hadoop-cos/dev-support/findbugs-exclude.xml
index e647e678a0..f8c3472640 100644
--- a/hadoop-cloud-storage-project/hadoop-cos/dev-support/findbugs-exclude.xml
+++ b/hadoop-cloud-storage-project/hadoop-cos/dev-support/findbugs-exclude.xml
@@ -16,8 +16,8 @@
-->
-
+
- h_LIB
+
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Server.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Server.java
index 907d55f9be..cc5d941903 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Server.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Server.java
@@ -3714,7 +3714,7 @@ void incrUserConnections(String user) {
if (count == null) {
count = 1;
} else {
- count++;
+ count = count + 1;
}
userToConnectionsMap.put(user, count);
}
@@ -3726,7 +3726,7 @@ void decrUserConnections(String user) {
if (count == null) {
return;
} else {
- count--;
+ count = count - 1;
}
if (count == 0) {
userToConnectionsMap.remove(user);
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/checker/DatasetVolumeChecker.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/checker/DatasetVolumeChecker.java
index 91582fe055..e31ac6f124 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/checker/DatasetVolumeChecker.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/checker/DatasetVolumeChecker.java
@@ -354,23 +354,29 @@ private class ResultHandler
}
@Override
- public void onSuccess(@Nonnull VolumeCheckResult result) {
- switch(result) {
- case HEALTHY:
- case DEGRADED:
- LOG.debug("Volume {} is {}.", reference.getVolume(), result);
- markHealthy();
- break;
- case FAILED:
- LOG.warn("Volume {} detected as being unhealthy",
+ public void onSuccess(VolumeCheckResult result) {
+ if (result == null) {
+ LOG.error("Unexpected health check result null for volume {}",
reference.getVolume());
- markFailed();
- break;
- default:
- LOG.error("Unexpected health check result {} for volume {}",
- result, reference.getVolume());
markHealthy();
- break;
+ } else {
+ switch(result) {
+ case HEALTHY:
+ case DEGRADED:
+ LOG.debug("Volume {} is {}.", reference.getVolume(), result);
+ markHealthy();
+ break;
+ case FAILED:
+ LOG.warn("Volume {} detected as being unhealthy",
+ reference.getVolume());
+ markFailed();
+ break;
+ default:
+ LOG.error("Unexpected health check result {} for volume {}",
+ result, reference.getVolume());
+ markHealthy();
+ break;
+ }
}
cleanup();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/checker/ThrottledAsyncChecker.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/checker/ThrottledAsyncChecker.java
index 6f04129b49..610c8fd97e 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/checker/ThrottledAsyncChecker.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/checker/ThrottledAsyncChecker.java
@@ -166,7 +166,7 @@ private void addResultCachingCallback(
Checkable target, ListenableFuture lf) {
Futures.addCallback(lf, new FutureCallback() {
@Override
- public void onSuccess(@Nullable V result) {
+ public void onSuccess(V result) {
synchronized (ThrottledAsyncChecker.this) {
checksInProgress.remove(target);
completedChecks.put(target, new LastCheckResult<>(
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogLoader.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogLoader.java
index e3694ba4f4..280e070950 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogLoader.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogLoader.java
@@ -1238,7 +1238,7 @@ private void incrOpCount(FSEditLogOpCodes opCode,
holder = new Holder(1);
opCounts.put(opCode, holder);
} else {
- holder.held++;
+ holder.held = holder.held + 1;
}
counter.increment();
}
diff --git a/hadoop-mapreduce-project/dev-support/findbugs-exclude.xml b/hadoop-mapreduce-project/dev-support/findbugs-exclude.xml
index 9b4d8c90f5..4e459b652b 100644
--- a/hadoop-mapreduce-project/dev-support/findbugs-exclude.xml
+++ b/hadoop-mapreduce-project/dev-support/findbugs-exclude.xml
@@ -533,5 +533,17 @@
-
+
+
+
+
+
+
+
+
+
+
diff --git a/hadoop-tools/hadoop-sls/src/main/java/org/apache/hadoop/yarn/sls/SLSRunner.java b/hadoop-tools/hadoop-sls/src/main/java/org/apache/hadoop/yarn/sls/SLSRunner.java
index 6f75bd17c6..5bfa8dc021 100644
--- a/hadoop-tools/hadoop-sls/src/main/java/org/apache/hadoop/yarn/sls/SLSRunner.java
+++ b/hadoop-tools/hadoop-sls/src/main/java/org/apache/hadoop/yarn/sls/SLSRunner.java
@@ -813,7 +813,7 @@ private void increaseQueueAppNum(String queue) throws YarnException {
if (appNum == null) {
appNum = 1;
} else {
- appNum++;
+ appNum = appNum + 1;
}
queueAppNumMap.put(queueName, appNum);
diff --git a/hadoop-yarn-project/hadoop-yarn/dev-support/findbugs-exclude.xml b/hadoop-yarn-project/hadoop-yarn/dev-support/findbugs-exclude.xml
index 3a37293357..95706f9e13 100644
--- a/hadoop-yarn-project/hadoop-yarn/dev-support/findbugs-exclude.xml
+++ b/hadoop-yarn-project/hadoop-yarn/dev-support/findbugs-exclude.xml
@@ -705,4 +705,10 @@
+
+
+
+
+
+
diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase-tests/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestTimelineReaderHBaseDown.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase-tests/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestTimelineReaderHBaseDown.java
index 1148b80d19..d83f130338 100644
--- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase-tests/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestTimelineReaderHBaseDown.java
+++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase-tests/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestTimelineReaderHBaseDown.java
@@ -181,14 +181,13 @@ private static void waitForHBaseDown(HBaseTimelineReaderImpl htr) throws
}
}
- private static void checkQuery(HBaseTimelineReaderImpl htr) throws
- IOException {
+ private static Set checkQuery(HBaseTimelineReaderImpl htr)
+ throws IOException {
TimelineReaderContext context =
new TimelineReaderContext(YarnConfiguration.DEFAULT_RM_CLUSTER_ID,
null, null, null, null, TimelineEntityType
.YARN_FLOW_ACTIVITY.toString(), null, null);
- Set entities = htr.getEntities(context, MONITOR_FILTERS,
- DATA_TO_RETRIEVE);
+ return htr.getEntities(context, MONITOR_FILTERS, DATA_TO_RETRIEVE);
}
private static void configure(HBaseTestingUtility util) {