From b8a446ba57d89c0896ec2d56dd919b0101e69f44 Mon Sep 17 00:00:00 2001 From: Masatake Iwasaki Date: Tue, 16 Aug 2016 13:30:40 +0900 Subject: [PATCH] HADOOP-13419. Fix javadoc warnings by JDK8 in hadoop-common package. Contributed by Kai Sasaki. --- .../org/apache/hadoop/fs/FileContext.java | 4 +- .../apache/hadoop/io/retry/package-info.java | 22 +++++++++ .../org/apache/hadoop/io/retry/package.html | 48 ------------------- .../org/apache/hadoop/ipc/package-info.java | 4 ++ .../java/org/apache/hadoop/ipc/package.html | 23 --------- 5 files changed, 28 insertions(+), 73 deletions(-) delete mode 100644 hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/retry/package.html delete mode 100644 hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/package.html diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java index e6a4cf406c..f2357730d4 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java @@ -303,7 +303,7 @@ static void processDeleteOnExit() { * * @throws UnsupportedFileSystemException If the file system for * absOrFqPath is not supported. - * @throws IOExcepton If the file system for absOrFqPath could + * @throws IOException If the file system for absOrFqPath could * not be instantiated. */ protected AbstractFileSystem getFSofPath(final Path absOrFqPath) @@ -2713,7 +2713,7 @@ public Void next(final AbstractFileSystem fs, final Path p) /** * Query the effective storage policy ID for the given file or directory. * - * @param src file or directory path. + * @param path file or directory path. * @return storage policy for give file. * @throws IOException */ diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/retry/package-info.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/retry/package-info.java index 693065f35b..089cf6f54c 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/retry/package-info.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/retry/package-info.java @@ -15,6 +15,28 @@ * See the License for the specific language governing permissions and * limitations under the License. */ + +/** + * A mechanism for selectively retrying methods that throw exceptions under + * certain circumstances. + * Typical usage is + * UnreliableImplementation unreliableImpl = new UnreliableImplementation(); + * UnreliableInterface unreliable = (UnreliableInterface) + * RetryProxy.create(UnreliableInterface.class, unreliableImpl, + * RetryPolicies.retryUpToMaximumCountWithFixedSleep(4, 10, + * TimeUnit.SECONDS)); + * unreliable.call(); + * + * This will retry any method called on unreliable four times - + * in this case the call() method - sleeping 10 seconds between + * each retry. There are a number of + * {@link org.apache.hadoop.io.retry.RetryPolicies retry policies} + * available, or you can implement a custom one by implementing + * {@link org.apache.hadoop.io.retry.RetryPolicy}. + * It is also possible to specify retry policies on a + * {@link org.apache.hadoop.io.retry.RetryProxy#create(Class, Object, Map) + * per-method basis}. + */ @InterfaceAudience.LimitedPrivate({"HBase", "HDFS", "MapReduce"}) @InterfaceStability.Evolving package org.apache.hadoop.io.retry; diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/retry/package.html b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/retry/package.html deleted file mode 100644 index ae553fc7a6..0000000000 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/retry/package.html +++ /dev/null @@ -1,48 +0,0 @@ - - - - - - -

-A mechanism for selectively retrying methods that throw exceptions under certain circumstances. -

- -

-Typical usage is -

- -
-UnreliableImplementation unreliableImpl = new UnreliableImplementation();
-UnreliableInterface unreliable = (UnreliableInterface)
-  RetryProxy.create(UnreliableInterface.class, unreliableImpl,
-    RetryPolicies.retryUpToMaximumCountWithFixedSleep(4, 10, TimeUnit.SECONDS));
-unreliable.call();
-
- -

-This will retry any method called on unreliable four times - in this case the call() -method - sleeping 10 seconds between -each retry. There are a number of {@link org.apache.hadoop.io.retry.RetryPolicies retry policies} -available, or you can implement a custom one by implementing {@link org.apache.hadoop.io.retry.RetryPolicy}. -It is also possible to specify retry policies on a -{@link org.apache.hadoop.io.retry.RetryProxy#create(Class, Object, Map) per-method basis}. -

- - - diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/package-info.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/package-info.java index 525ef1f1b6..cb35e93848 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/package-info.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/package-info.java @@ -15,6 +15,10 @@ * See the License for the specific language governing permissions and * limitations under the License. */ + +/** + * Tools to help define network clients and servers. + */ @InterfaceAudience.LimitedPrivate({"HBase", "HDFS", "MapReduce"}) @InterfaceStability.Evolving package org.apache.hadoop.ipc; diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/package.html b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/package.html deleted file mode 100644 index 3efd81a297..0000000000 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/package.html +++ /dev/null @@ -1,23 +0,0 @@ - - - - - -Tools to help define network clients and servers. - -