HDFS-6079. Timeout for getFileBlockStorageLocations does not work. Contributed by Andrew Wang.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1576979 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
135700a611
commit
b849dd5df9
@ -601,6 +601,8 @@ Release 2.4.0 - UNRELEASED
|
|||||||
HDFS-6086. Fix a case where zero-copy or no-checksum reads were not allowed
|
HDFS-6086. Fix a case where zero-copy or no-checksum reads were not allowed
|
||||||
even when the block was cached (cmccabe)
|
even when the block was cached (cmccabe)
|
||||||
|
|
||||||
|
HDFS-6079. Timeout for getFileBlockStorageLocations does not work. (wang)
|
||||||
|
|
||||||
BREAKDOWN OF HDFS-5698 SUBTASKS AND RELATED JIRAS
|
BREAKDOWN OF HDFS-5698 SUBTASKS AND RELATED JIRAS
|
||||||
|
|
||||||
HDFS-5717. Save FSImage header in protobuf. (Haohui Mai via jing9)
|
HDFS-5717. Save FSImage header in protobuf. (Haohui Mai via jing9)
|
||||||
|
@ -24,6 +24,7 @@
|
|||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.Map;
|
import java.util.Map;
|
||||||
import java.util.concurrent.Callable;
|
import java.util.concurrent.Callable;
|
||||||
|
import java.util.concurrent.CancellationException;
|
||||||
import java.util.concurrent.ExecutionException;
|
import java.util.concurrent.ExecutionException;
|
||||||
import java.util.concurrent.ExecutorService;
|
import java.util.concurrent.ExecutorService;
|
||||||
import java.util.concurrent.Future;
|
import java.util.concurrent.Future;
|
||||||
@ -158,6 +159,9 @@ static Map<DatanodeInfo, HdfsBlocksMetadata> queryDatanodesForHdfsBlocksMetadata
|
|||||||
try {
|
try {
|
||||||
HdfsBlocksMetadata metadata = future.get();
|
HdfsBlocksMetadata metadata = future.get();
|
||||||
metadatas.put(callable.getDatanodeInfo(), metadata);
|
metadatas.put(callable.getDatanodeInfo(), metadata);
|
||||||
|
} catch (CancellationException e) {
|
||||||
|
LOG.info("Cancelled while waiting for datanode "
|
||||||
|
+ datanode.getIpcAddr(false) + ": " + e.toString());
|
||||||
} catch (ExecutionException e) {
|
} catch (ExecutionException e) {
|
||||||
Throwable t = e.getCause();
|
Throwable t = e.getCause();
|
||||||
if (t instanceof InvalidBlockTokenException) {
|
if (t instanceof InvalidBlockTokenException) {
|
||||||
|
@ -1224,7 +1224,8 @@ public BlockStorageLocation[] getBlockStorageLocations(
|
|||||||
getConf().connectToDnViaHostname);
|
getConf().connectToDnViaHostname);
|
||||||
|
|
||||||
if (LOG.isTraceEnabled()) {
|
if (LOG.isTraceEnabled()) {
|
||||||
LOG.trace("metadata returned: " + Joiner.on("\n").withKeyValueSeparator("=").join(metadatas));
|
LOG.trace("metadata returned: "
|
||||||
|
+ Joiner.on("\n").withKeyValueSeparator("=").join(metadatas));
|
||||||
}
|
}
|
||||||
|
|
||||||
// Regroup the returned VolumeId metadata to again be grouped by
|
// Regroup the returned VolumeId metadata to again be grouped by
|
||||||
|
@ -1159,6 +1159,9 @@ public HdfsBlocksMetadata getHdfsBlocksMetadata(
|
|||||||
checkBlockToken(new ExtendedBlock(bpId, blockIds[i]),
|
checkBlockToken(new ExtendedBlock(bpId, blockIds[i]),
|
||||||
tokens.get(i), BlockTokenSecretManager.AccessMode.READ);
|
tokens.get(i), BlockTokenSecretManager.AccessMode.READ);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
DataNodeFaultInjector.get().getHdfsBlocksMetadata();
|
||||||
|
|
||||||
return data.getHdfsBlocksMetadata(bpId, blockIds);
|
return data.getHdfsBlocksMetadata(bpId, blockIds);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -0,0 +1,38 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.hdfs.server.datanode;
|
||||||
|
|
||||||
|
import com.google.common.annotations.VisibleForTesting;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Used for injecting faults in DFSClient and DFSOutputStream tests.
|
||||||
|
* Calls into this are a no-op in production code.
|
||||||
|
*/
|
||||||
|
@VisibleForTesting
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
public class DataNodeFaultInjector {
|
||||||
|
public static DataNodeFaultInjector instance = new DataNodeFaultInjector();
|
||||||
|
|
||||||
|
public static DataNodeFaultInjector get() {
|
||||||
|
return instance;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void getHdfsBlocksMetadata() {}
|
||||||
|
}
|
@ -38,6 +38,7 @@
|
|||||||
import java.util.EnumSet;
|
import java.util.EnumSet;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.Random;
|
import java.util.Random;
|
||||||
|
import java.util.concurrent.CancellationException;
|
||||||
|
|
||||||
import org.apache.commons.lang.ArrayUtils;
|
import org.apache.commons.lang.ArrayUtils;
|
||||||
import org.apache.commons.logging.impl.Log4JLogger;
|
import org.apache.commons.logging.impl.Log4JLogger;
|
||||||
@ -60,6 +61,7 @@
|
|||||||
import org.apache.hadoop.fs.VolumeId;
|
import org.apache.hadoop.fs.VolumeId;
|
||||||
import org.apache.hadoop.fs.permission.FsPermission;
|
import org.apache.hadoop.fs.permission.FsPermission;
|
||||||
import org.apache.hadoop.hdfs.MiniDFSCluster.DataNodeProperties;
|
import org.apache.hadoop.hdfs.MiniDFSCluster.DataNodeProperties;
|
||||||
|
import org.apache.hadoop.hdfs.server.datanode.DataNodeFaultInjector;
|
||||||
import org.apache.hadoop.hdfs.server.namenode.ha.HATestUtil;
|
import org.apache.hadoop.hdfs.server.namenode.ha.HATestUtil;
|
||||||
import org.apache.hadoop.hdfs.web.HftpFileSystem;
|
import org.apache.hadoop.hdfs.web.HftpFileSystem;
|
||||||
import org.apache.hadoop.hdfs.web.WebHdfsFileSystem;
|
import org.apache.hadoop.hdfs.web.WebHdfsFileSystem;
|
||||||
@ -71,6 +73,9 @@
|
|||||||
import org.apache.log4j.Level;
|
import org.apache.log4j.Level;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
import org.mockito.InOrder;
|
import org.mockito.InOrder;
|
||||||
|
import org.mockito.Mockito;
|
||||||
|
import org.mockito.invocation.InvocationOnMock;
|
||||||
|
import org.mockito.stubbing.Answer;
|
||||||
|
|
||||||
import com.google.common.base.Supplier;
|
import com.google.common.base.Supplier;
|
||||||
import com.google.common.collect.Lists;
|
import com.google.common.collect.Lists;
|
||||||
@ -782,8 +787,11 @@ public void testGetFileBlockStorageLocationsError() throws Exception {
|
|||||||
final Configuration conf = getTestConfiguration();
|
final Configuration conf = getTestConfiguration();
|
||||||
conf.setBoolean(DFSConfigKeys.DFS_HDFS_BLOCKS_METADATA_ENABLED,
|
conf.setBoolean(DFSConfigKeys.DFS_HDFS_BLOCKS_METADATA_ENABLED,
|
||||||
true);
|
true);
|
||||||
|
conf.setInt(
|
||||||
|
DFSConfigKeys.DFS_CLIENT_FILE_BLOCK_STORAGE_LOCATIONS_TIMEOUT_MS, 1500);
|
||||||
conf.setInt(
|
conf.setInt(
|
||||||
CommonConfigurationKeysPublic.IPC_CLIENT_CONNECT_MAX_RETRIES_KEY, 0);
|
CommonConfigurationKeysPublic.IPC_CLIENT_CONNECT_MAX_RETRIES_KEY, 0);
|
||||||
|
|
||||||
MiniDFSCluster cluster = null;
|
MiniDFSCluster cluster = null;
|
||||||
try {
|
try {
|
||||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||||
@ -828,12 +836,33 @@ public Boolean get() {
|
|||||||
List<BlockLocation> allLocs = Lists.newArrayList();
|
List<BlockLocation> allLocs = Lists.newArrayList();
|
||||||
allLocs.addAll(Arrays.asList(blockLocs1));
|
allLocs.addAll(Arrays.asList(blockLocs1));
|
||||||
allLocs.addAll(Arrays.asList(blockLocs2));
|
allLocs.addAll(Arrays.asList(blockLocs2));
|
||||||
|
|
||||||
|
// Stall on the DN to test the timeout
|
||||||
|
DataNodeFaultInjector injector = Mockito.mock(DataNodeFaultInjector.class);
|
||||||
|
Mockito.doAnswer(new Answer<Void>() {
|
||||||
|
@Override
|
||||||
|
public Void answer(InvocationOnMock invocation) throws Throwable {
|
||||||
|
Thread.sleep(3000);
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
}).when(injector).getHdfsBlocksMetadata();
|
||||||
|
DataNodeFaultInjector.instance = injector;
|
||||||
|
|
||||||
|
BlockStorageLocation[] locs = fs.getFileBlockStorageLocations(allLocs);
|
||||||
|
for (BlockStorageLocation loc: locs) {
|
||||||
|
assertEquals(
|
||||||
|
"Found more than 0 cached hosts although RPCs supposedly timed out",
|
||||||
|
0, loc.getCachedHosts().length);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Restore a default injector
|
||||||
|
DataNodeFaultInjector.instance = new DataNodeFaultInjector();
|
||||||
|
|
||||||
// Stop a datanode to simulate a failure.
|
// Stop a datanode to simulate a failure.
|
||||||
DataNodeProperties stoppedNode = cluster.stopDataNode(0);
|
DataNodeProperties stoppedNode = cluster.stopDataNode(0);
|
||||||
|
|
||||||
// Fetch VolumeBlockLocations
|
// Fetch VolumeBlockLocations
|
||||||
BlockStorageLocation[] locs = fs.getFileBlockStorageLocations(allLocs);
|
locs = fs.getFileBlockStorageLocations(allLocs);
|
||||||
assertEquals("Expected two HdfsBlockLocation for two 1-block files", 2,
|
assertEquals("Expected two HdfsBlockLocation for two 1-block files", 2,
|
||||||
locs.length);
|
locs.length);
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user