From e216e8e2334519b7c833d99586218e99a39265f3 Mon Sep 17 00:00:00 2001 From: Daniel Templeton Date: Tue, 27 Dec 2016 11:12:05 -0800 Subject: [PATCH] HADOOP-13932. Fix indefinite article in comments (Contributed by LiXin Ge via Daniel Templeton) --- .../namenode/FSImagePreTransactionalStorageInspector.java | 2 +- .../hadoop-hdfs/src/site/markdown/HdfsRollingUpgrade.md | 8 ++++---- .../hadoop-hdfs/src/site/markdown/LibHdfs.md | 2 +- .../dev-support/jdiff/Apache_Hadoop_YARN_API_2.6.0.xml | 2 +- .../dev-support/jdiff/Apache_Hadoop_YARN_API_2.7.2.xml | 2 +- .../src/main/java/org/apache/hadoop/yarn/conf/HAUtil.java | 4 ++-- .../org/apache/hadoop/yarn/conf/YarnConfiguration.java | 2 +- .../src/main/resources/yarn-default.xml | 8 ++++---- .../resourcemanager/recovery/FileSystemRMStateStore.java | 2 +- .../resourcemanager/scheduler/fair/FSQueueMetrics.java | 2 +- .../org/apache/hadoop/yarn/server/MiniYARNCluster.java | 2 +- 11 files changed, 18 insertions(+), 18 deletions(-) diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImagePreTransactionalStorageInspector.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImagePreTransactionalStorageInspector.java index 3cd5ae1aee..64693852a5 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImagePreTransactionalStorageInspector.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImagePreTransactionalStorageInspector.java @@ -41,7 +41,7 @@ import org.apache.hadoop.io.IOUtils; /** - * Inspects a FSImage storage directory in the "old" (pre-HDFS-1073) format. + * Inspects an FSImage storage directory in the "old" (pre-HDFS-1073) format. * This format has the following data files: * - fsimage * - fsimage.ckpt (when checkpoint is being uploaded) diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/HdfsRollingUpgrade.md b/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/HdfsRollingUpgrade.md index f9b76b80f5..889d7a0a3b 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/HdfsRollingUpgrade.md +++ b/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/HdfsRollingUpgrade.md @@ -65,7 +65,7 @@ Note that rolling upgrade is supported only from Hadoop-2.4.0 onwards. ### Upgrade without Downtime -In a HA cluster, there are two or more *NameNodes (NNs)*, many *DataNodes (DNs)*, +In an HA cluster, there are two or more *NameNodes (NNs)*, many *DataNodes (DNs)*, a few *JournalNodes (JNs)* and a few *ZooKeeperNodes (ZKNs)*. *JNs* is relatively stable and does not require upgrade when upgrading HDFS in most of the cases. In the rolling upgrade procedure described here, @@ -76,7 +76,7 @@ Upgrading *JNs* and *ZKNs* may incur cluster downtime. Suppose there are two namenodes *NN1* and *NN2*, where *NN1* and *NN2* are respectively in active and standby states. -The following are the steps for upgrading a HA cluster: +The following are the steps for upgrading an HA cluster: 1. Prepare Rolling Upgrade 1. Run "[`hdfs dfsadmin -rollingUpgrade prepare`](#dfsadmin_-rollingUpgrade)" @@ -133,7 +133,7 @@ However, datanodes can still be upgraded in a rolling manner. In a non-HA cluster, there are a *NameNode (NN)*, a *SecondaryNameNode (SNN)* and many *DataNodes (DNs)*. -The procedure for upgrading a non-HA cluster is similar to upgrading a HA cluster +The procedure for upgrading a non-HA cluster is similar to upgrading an HA cluster except that Step 2 "Upgrade Active and Standby *NNs*" is changed to below: * Upgrade *NN* and *SNN* @@ -175,7 +175,7 @@ A newer release is downgradable to the pre-upgrade release only if both the namenode layout version and the datanode layout version are not changed between these two releases. -In a HA cluster, +In an HA cluster, when a rolling upgrade from an old software release to a new software release is in progress, it is possible to downgrade, in a rolling fashion, the upgraded machines back to the old software release. Same as before, suppose *NN1* and *NN2* are respectively in active and standby states. diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/LibHdfs.md b/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/LibHdfs.md index 7b869c80f5..2697cde82d 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/LibHdfs.md +++ b/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/LibHdfs.md @@ -76,7 +76,7 @@ libdhfs is thread safe. * Concurrency and Hadoop FS "handles" - The Hadoop FS implementation includes a FS handle cache which + The Hadoop FS implementation includes an FS handle cache which caches based on the URI of the namenode along with the user connecting. So, all calls to `hdfsConnect` will return the same handle but calls to `hdfsConnectAsUser` with different users will diff --git a/hadoop-yarn-project/hadoop-yarn/dev-support/jdiff/Apache_Hadoop_YARN_API_2.6.0.xml b/hadoop-yarn-project/hadoop-yarn/dev-support/jdiff/Apache_Hadoop_YARN_API_2.6.0.xml index 5d58600a0c..a03d7a7ded 100644 --- a/hadoop-yarn-project/hadoop-yarn/dev-support/jdiff/Apache_Hadoop_YARN_API_2.6.0.xml +++ b/hadoop-yarn-project/hadoop-yarn/dev-support/jdiff/Apache_Hadoop_YARN_API_2.6.0.xml @@ -9801,7 +9801,7 @@ name property as a - InetSocketAddress. On a HA cluster, + InetSocketAddress. On an HA cluster, this fetches the address corresponding to the RM identified by {@link #RM_HA_ID}. @param name property name. diff --git a/hadoop-yarn-project/hadoop-yarn/dev-support/jdiff/Apache_Hadoop_YARN_API_2.7.2.xml b/hadoop-yarn-project/hadoop-yarn/dev-support/jdiff/Apache_Hadoop_YARN_API_2.7.2.xml index 2328c8153d..a651da8b41 100644 --- a/hadoop-yarn-project/hadoop-yarn/dev-support/jdiff/Apache_Hadoop_YARN_API_2.7.2.xml +++ b/hadoop-yarn-project/hadoop-yarn/dev-support/jdiff/Apache_Hadoop_YARN_API_2.7.2.xml @@ -9416,7 +9416,7 @@ name property as a - InetSocketAddress. On a HA cluster, + InetSocketAddress. On an HA cluster, this fetches the address corresponding to the RM identified by {@link #RM_HA_ID}. @param name property name. diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/HAUtil.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/HAUtil.java index 594832e005..e4948e74b0 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/HAUtil.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/HAUtil.java @@ -208,7 +208,7 @@ public static String getRMHAId(Configuration conf) { @VisibleForTesting static String getNeedToSetValueMessage(String confKey) { - return confKey + " needs to be set in a HA configuration."; + return confKey + " needs to be set in an HA configuration."; } @VisibleForTesting @@ -223,7 +223,7 @@ static String getRMHAIdNeedToBeIncludedMessage(String ids, String rmId) { return YarnConfiguration.RM_HA_IDS + "(" + ids + ") need to contain " + YarnConfiguration.RM_HA_ID + "(" - + rmId + ") in a HA configuration."; + + rmId + ") in an HA configuration."; } @VisibleForTesting diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java index 5adbfa7059..bf0a0ab13c 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java @@ -2767,7 +2767,7 @@ public static List getServiceAddressConfKeys(Configuration conf) { /** * Get the socket address for name property as a - * InetSocketAddress. On a HA cluster, + * InetSocketAddress. On an HA cluster, * this fetches the address corresponding to the RM identified by * {@link #RM_HA_ID}. * @param name property name. diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/resources/yarn-default.xml b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/resources/yarn-default.xml index 3e739aba35..9d733088cb 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/resources/yarn-default.xml +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/resources/yarn-default.xml @@ -441,7 +441,7 @@ Host:Port of the ZooKeeper server to be used by the RM. This must be supplied when using the ZooKeeper based implementation of the - RM state store and/or embedded automatic failover in a HA setting. + RM state store and/or embedded automatic failover in an HA setting. yarn.resourcemanager.zk-address @@ -490,7 +490,7 @@ - ACLs to be used for the root znode when using ZKRMStateStore in a HA + ACLs to be used for the root znode when using ZKRMStateStore in an HA scenario for fencing. ZKRMStateStore supports implicit fencing to allow a single @@ -606,7 +606,7 @@ - Name of the cluster. In a HA setting, + Name of the cluster. In an HA setting, this is used to ensure the RM participates in leader election for this cluster and ensures it does not affect other clusters @@ -2188,7 +2188,7 @@ yarn.timeline-service.client.fd-retain-secs - How long the ATS v1.5 writer will keep a FSStream open. + How long the ATS v1.5 writer will keep an FSStream open. If this fsstream does not write anything for this configured time, it will be close. diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/FileSystemRMStateStore.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/FileSystemRMStateStore.java index b9a4374929..929e2dab11 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/FileSystemRMStateStore.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/FileSystemRMStateStore.java @@ -740,7 +740,7 @@ T runWithRetries() throws Exception { try { return run(); } catch (IOException e) { - LOG.info("Exception while executing a FS operation.", e); + LOG.info("Exception while executing an FS operation.", e); if (++retry > fsNumRetries) { LOG.info("Maxed out FS retries. Giving up!"); throw e; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FSQueueMetrics.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FSQueueMetrics.java index ca375f2fe2..22306a0f4a 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FSQueueMetrics.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FSQueueMetrics.java @@ -200,7 +200,7 @@ static FSQueueMetrics forQueue(String queueName, Queue parent, * @param parent parent queue * @param enableUserMetrics if user metrics is needed * @param conf configuration - * @return a FSQueueMetrics object + * @return an FSQueueMetrics object */ @VisibleForTesting public synchronized diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-tests/src/test/java/org/apache/hadoop/yarn/server/MiniYARNCluster.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-tests/src/test/java/org/apache/hadoop/yarn/server/MiniYARNCluster.java index 5ae74a6c82..77240c63f6 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-tests/src/test/java/org/apache/hadoop/yarn/server/MiniYARNCluster.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-tests/src/test/java/org/apache/hadoop/yarn/server/MiniYARNCluster.java @@ -385,7 +385,7 @@ public File getTestWorkDir() { } /** - * In a HA cluster, go through all the RMs and find the Active RM. In a + * In an HA cluster, go through all the RMs and find the Active RM. In a * non-HA cluster, return the index of the only RM. * * @return index of the active RM or -1 if none of them turn active