HADOOP-12078. The default retry policy does not handle RetriableException correctly. (Contributed by Arpit Agarwal)

This commit is contained in:
Arpit Agarwal 2015-06-09 20:58:39 -07:00
parent 3c2397cb04
commit 3107434031
5 changed files with 113 additions and 3 deletions

View File

@ -887,6 +887,9 @@ Release 2.7.1 - UNRELEASED
HADOOP-12058. Fix dead links to DistCp and Hadoop Archives pages. HADOOP-12058. Fix dead links to DistCp and Hadoop Archives pages.
(Kazuho Fujii via aajisaka) (Kazuho Fujii via aajisaka)
HADOOP-12078. The default retry policy does not handle RetriableException
correctly. (Arpit Agarwal)
Release 2.7.0 - 2015-04-20 Release 2.7.0 - 2015-04-20
INCOMPATIBLE CHANGES INCOMPATIBLE CHANGES

View File

@ -624,7 +624,7 @@ private static boolean isWrappedStandbyException(Exception e) {
return unwrapped instanceof StandbyException; return unwrapped instanceof StandbyException;
} }
private static RetriableException getWrappedRetriableException(Exception e) { static RetriableException getWrappedRetriableException(Exception e) {
if (!(e instanceof RemoteException)) { if (!(e instanceof RemoteException)) {
return null; return null;
} }

View File

@ -25,6 +25,7 @@
import org.apache.hadoop.ipc.RemoteException; import org.apache.hadoop.ipc.RemoteException;
import com.google.protobuf.ServiceException; import com.google.protobuf.ServiceException;
import org.apache.hadoop.ipc.RetriableException;
public class RetryUtils { public class RetryUtils {
public static final Log LOG = LogFactory.getLog(RetryUtils.class); public static final Log LOG = LogFactory.getLog(RetryUtils.class);
@ -92,7 +93,11 @@ public RetryAction shouldRetry(Exception e, int retries, int failovers,
//see (1) and (2) in the javadoc of this method. //see (1) and (2) in the javadoc of this method.
final RetryPolicy p; final RetryPolicy p;
if (e instanceof RemoteException) { if (e instanceof RetriableException
|| RetryPolicies.getWrappedRetriableException(e) != null) {
// RetriableException or RetriableException wrapped
p = multipleLinearRandomRetry;
} else if (e instanceof RemoteException) {
final RemoteException re = (RemoteException)e; final RemoteException re = (RemoteException)e;
p = remoteExceptionToRetry.equals(re.getClassName())? p = remoteExceptionToRetry.equals(re.getClassName())?
multipleLinearRandomRetry: RetryPolicies.TRY_ONCE_THEN_FAIL; multipleLinearRandomRetry: RetryPolicies.TRY_ONCE_THEN_FAIL;

View File

@ -0,0 +1,101 @@
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
* <p/>
* http://www.apache.org/licenses/LICENSE-2.0
* <p/>
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.io.retry;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.ipc.RemoteException;
import org.apache.hadoop.ipc.RetriableException;
import org.junit.Rule;
import org.junit.Test;
import org.junit.rules.Timeout;
import java.io.IOException;
import static org.hamcrest.core.Is.is;
import static org.junit.Assert.assertThat;
/**
* Test the behavior of the default retry policy.
*/
public class TestDefaultRetryPolicy {
@Rule
public Timeout timeout = new Timeout(300000);
/**
* Verify that the default retry policy correctly retries
* RetriableException when defaultRetryPolicyEnabled is enabled.
*
* @throws IOException
*/
@Test
public void testWithRetriable() throws Exception {
Configuration conf = new Configuration();
RetryPolicy policy = RetryUtils.getDefaultRetryPolicy(
conf, "Test.No.Such.Key",
true, // defaultRetryPolicyEnabled = true
"Test.No.Such.Key", "10000,6",
null);
RetryPolicy.RetryAction action = policy.shouldRetry(
new RetriableException("Dummy exception"), 0, 0, true);
assertThat(action.action,
is(RetryPolicy.RetryAction.RetryDecision.RETRY));
}
/**
* Verify that the default retry policy correctly retries
* a RetriableException wrapped in a RemoteException when
* defaultRetryPolicyEnabled is enabled.
*
* @throws IOException
*/
@Test
public void testWithWrappedRetriable() throws Exception {
Configuration conf = new Configuration();
RetryPolicy policy = RetryUtils.getDefaultRetryPolicy(
conf, "Test.No.Such.Key",
true, // defaultRetryPolicyEnabled = true
"Test.No.Such.Key", "10000,6",
null);
RetryPolicy.RetryAction action = policy.shouldRetry(
new RemoteException(RetriableException.class.getName(),
"Dummy exception"), 0, 0, true);
assertThat(action.action,
is(RetryPolicy.RetryAction.RetryDecision.RETRY));
}
/**
* Verify that the default retry policy does *not* retry
* RetriableException when defaultRetryPolicyEnabled is disabled.
*
* @throws IOException
*/
@Test
public void testWithRetriableAndRetryDisabled() throws Exception {
Configuration conf = new Configuration();
RetryPolicy policy = RetryUtils.getDefaultRetryPolicy(
conf, "Test.No.Such.Key",
false, // defaultRetryPolicyEnabled = false
"Test.No.Such.Key", "10000,6",
null);
RetryPolicy.RetryAction action = policy.shouldRetry(
new RetriableException("Dummy exception"), 0, 0, true);
assertThat(action.action,
is(RetryPolicy.RetryAction.RetryDecision.FAIL));
}
}

View File

@ -144,6 +144,7 @@
import org.apache.hadoop.io.Text; import org.apache.hadoop.io.Text;
import org.apache.hadoop.ipc.ProtobufRpcEngine; import org.apache.hadoop.ipc.ProtobufRpcEngine;
import org.apache.hadoop.ipc.RPC; import org.apache.hadoop.ipc.RPC;
import org.apache.hadoop.ipc.RetriableException;
import org.apache.hadoop.ipc.RetryCache; import org.apache.hadoop.ipc.RetryCache;
import org.apache.hadoop.ipc.RetryCache.CacheEntry; import org.apache.hadoop.ipc.RetryCache.CacheEntry;
import org.apache.hadoop.ipc.RetryCache.CacheEntryWithPayload; import org.apache.hadoop.ipc.RetryCache.CacheEntryWithPayload;
@ -1886,7 +1887,7 @@ public void removeXAttr(String src, XAttr xAttr) throws IOException {
private void checkNNStartup() throws IOException { private void checkNNStartup() throws IOException {
if (!this.nn.isStarted()) { if (!this.nn.isStarted()) {
throw new IOException(this.nn.getRole() + " still not started"); throw new RetriableException(this.nn.getRole() + " still not started");
} }
} }