HADOOP-15729. [s3a] Allow core threads to time out. (#1075)
This commit is contained in:
parent
7adc4a327a
commit
5672efa5c7
@ -42,7 +42,6 @@
|
|||||||
import java.util.Set;
|
import java.util.Set;
|
||||||
import java.util.Objects;
|
import java.util.Objects;
|
||||||
import java.util.concurrent.CompletableFuture;
|
import java.util.concurrent.CompletableFuture;
|
||||||
import java.util.concurrent.ExecutorService;
|
|
||||||
import java.util.concurrent.LinkedBlockingQueue;
|
import java.util.concurrent.LinkedBlockingQueue;
|
||||||
import java.util.concurrent.ThreadPoolExecutor;
|
import java.util.concurrent.ThreadPoolExecutor;
|
||||||
import java.util.concurrent.TimeUnit;
|
import java.util.concurrent.TimeUnit;
|
||||||
@ -213,7 +212,7 @@ public class S3AFileSystem extends FileSystem implements StreamCapabilities,
|
|||||||
private boolean enableMultiObjectsDelete;
|
private boolean enableMultiObjectsDelete;
|
||||||
private TransferManager transfers;
|
private TransferManager transfers;
|
||||||
private ListeningExecutorService boundedThreadPool;
|
private ListeningExecutorService boundedThreadPool;
|
||||||
private ExecutorService unboundedThreadPool;
|
private ThreadPoolExecutor unboundedThreadPool;
|
||||||
private int executorCapacity;
|
private int executorCapacity;
|
||||||
private long multiPartThreshold;
|
private long multiPartThreshold;
|
||||||
public static final Logger LOG = LoggerFactory.getLogger(S3AFileSystem.class);
|
public static final Logger LOG = LoggerFactory.getLogger(S3AFileSystem.class);
|
||||||
@ -440,6 +439,7 @@ private void initThreadPools(Configuration conf) {
|
|||||||
new LinkedBlockingQueue<>(),
|
new LinkedBlockingQueue<>(),
|
||||||
BlockingThreadPoolExecutorService.newDaemonThreadFactory(
|
BlockingThreadPoolExecutorService.newDaemonThreadFactory(
|
||||||
"s3a-transfer-unbounded"));
|
"s3a-transfer-unbounded"));
|
||||||
|
unboundedThreadPool.allowCoreThreadTimeOut(true);
|
||||||
executorCapacity = intOption(conf,
|
executorCapacity = intOption(conf,
|
||||||
EXECUTOR_CAPACITY, DEFAULT_EXECUTOR_CAPACITY, 1);
|
EXECUTOR_CAPACITY, DEFAULT_EXECUTOR_CAPACITY, 1);
|
||||||
}
|
}
|
||||||
|
@ -35,6 +35,7 @@
|
|||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.fs.contract.ContractTestUtils;
|
import org.apache.hadoop.fs.contract.ContractTestUtils;
|
||||||
import org.apache.hadoop.fs.contract.ContractTestUtils.NanoTimer;
|
import org.apache.hadoop.fs.contract.ContractTestUtils.NanoTimer;
|
||||||
|
import org.apache.hadoop.fs.s3a.Constants;
|
||||||
import org.apache.hadoop.fs.s3a.S3AFileSystem;
|
import org.apache.hadoop.fs.s3a.S3AFileSystem;
|
||||||
import org.apache.hadoop.fs.s3a.S3ATestUtils;
|
import org.apache.hadoop.fs.s3a.S3ATestUtils;
|
||||||
|
|
||||||
@ -52,10 +53,8 @@ public class ITestS3AConcurrentOps extends S3AScaleTestBase {
|
|||||||
private static final Logger LOG = LoggerFactory.getLogger(
|
private static final Logger LOG = LoggerFactory.getLogger(
|
||||||
ITestS3AConcurrentOps.class);
|
ITestS3AConcurrentOps.class);
|
||||||
private final int concurrentRenames = 10;
|
private final int concurrentRenames = 10;
|
||||||
|
private final int fileSize = 1024 * 1024;
|
||||||
private Path testRoot;
|
private Path testRoot;
|
||||||
private Path[] source = new Path[concurrentRenames];
|
|
||||||
private Path[] target = new Path[concurrentRenames];
|
|
||||||
private S3AFileSystem fs;
|
|
||||||
private S3AFileSystem auxFs;
|
private S3AFileSystem auxFs;
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
@ -66,43 +65,10 @@ protected int getTestTimeoutSeconds() {
|
|||||||
@Override
|
@Override
|
||||||
public void setup() throws Exception {
|
public void setup() throws Exception {
|
||||||
super.setup();
|
super.setup();
|
||||||
fs = getRestrictedFileSystem();
|
|
||||||
auxFs = getNormalFileSystem();
|
auxFs = getNormalFileSystem();
|
||||||
|
|
||||||
testRoot = path("/ITestS3AConcurrentOps");
|
testRoot = path("/ITestS3AConcurrentOps");
|
||||||
testRoot = S3ATestUtils.createTestPath(testRoot);
|
testRoot = S3ATestUtils.createTestPath(testRoot);
|
||||||
|
|
||||||
for (int i = 0; i < concurrentRenames; i++){
|
|
||||||
source[i] = new Path(testRoot, "source" + i);
|
|
||||||
target[i] = new Path(testRoot, "target" + i);
|
|
||||||
}
|
|
||||||
|
|
||||||
LOG.info("Generating data...");
|
|
||||||
auxFs.mkdirs(testRoot);
|
|
||||||
byte[] zeroes = ContractTestUtils.dataset(1024*1024, 0, Integer.MAX_VALUE);
|
|
||||||
for (Path aSource : source) {
|
|
||||||
try(FSDataOutputStream out = auxFs.create(aSource)) {
|
|
||||||
for (int mb = 0; mb < 20; mb++) {
|
|
||||||
LOG.debug("{}: Block {}...", aSource, mb);
|
|
||||||
out.write(zeroes);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
LOG.info("Data generated...");
|
|
||||||
}
|
|
||||||
|
|
||||||
private S3AFileSystem getRestrictedFileSystem() throws Exception {
|
|
||||||
Configuration conf = getConfiguration();
|
|
||||||
conf.setInt(MAX_THREADS, 2);
|
|
||||||
conf.setInt(MAX_TOTAL_TASKS, 1);
|
|
||||||
|
|
||||||
conf.set(MIN_MULTIPART_THRESHOLD, "10M");
|
|
||||||
conf.set(MULTIPART_SIZE, "5M");
|
|
||||||
|
|
||||||
S3AFileSystem s3a = getFileSystem();
|
|
||||||
URI rootURI = new URI(conf.get(TEST_FS_S3A_NAME));
|
|
||||||
s3a.initialize(rootURI, conf);
|
|
||||||
return s3a;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
private S3AFileSystem getNormalFileSystem() throws Exception {
|
private S3AFileSystem getNormalFileSystem() throws Exception {
|
||||||
@ -118,18 +84,35 @@ public void teardown() throws Exception {
|
|||||||
super.teardown();
|
super.teardown();
|
||||||
if (auxFs != null) {
|
if (auxFs != null) {
|
||||||
auxFs.delete(testRoot, true);
|
auxFs.delete(testRoot, true);
|
||||||
|
auxFs.close();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
private void parallelRenames(int concurrentRenames, final S3AFileSystem fs,
|
||||||
* Attempts to trigger a deadlock that would happen if any bounded resource
|
String sourceNameBase, String targetNameBase) throws ExecutionException,
|
||||||
* pool became saturated with control tasks that depended on other tasks
|
InterruptedException, IOException {
|
||||||
* that now can't enter the resource pool to get completed.
|
|
||||||
*/
|
Path[] source = new Path[concurrentRenames];
|
||||||
@Test
|
Path[] target = new Path[concurrentRenames];
|
||||||
@SuppressWarnings("unchecked")
|
|
||||||
public void testParallelRename() throws InterruptedException,
|
for (int i = 0; i < concurrentRenames; i++){
|
||||||
ExecutionException, IOException {
|
source[i] = new Path(testRoot, sourceNameBase + i);
|
||||||
|
target[i] = new Path(testRoot, targetNameBase + i);
|
||||||
|
}
|
||||||
|
|
||||||
|
LOG.info("Generating data...");
|
||||||
|
auxFs.mkdirs(testRoot);
|
||||||
|
byte[] zeroes = ContractTestUtils.dataset(fileSize, 0, Integer.MAX_VALUE);
|
||||||
|
for (Path aSource : source) {
|
||||||
|
try(FSDataOutputStream out = auxFs.create(aSource)) {
|
||||||
|
for (int mb = 0; mb < 20; mb++) {
|
||||||
|
LOG.debug("{}: Block {}...", aSource, mb);
|
||||||
|
out.write(zeroes);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
LOG.info("Data generated...");
|
||||||
|
|
||||||
ExecutorService executor = Executors.newFixedThreadPool(
|
ExecutorService executor = Executors.newFixedThreadPool(
|
||||||
concurrentRenames, new ThreadFactory() {
|
concurrentRenames, new ThreadFactory() {
|
||||||
private AtomicInteger count = new AtomicInteger(0);
|
private AtomicInteger count = new AtomicInteger(0);
|
||||||
@ -139,6 +122,7 @@ public Thread newThread(Runnable r) {
|
|||||||
"testParallelRename" + count.getAndIncrement());
|
"testParallelRename" + count.getAndIncrement());
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
try {
|
||||||
((ThreadPoolExecutor)executor).prestartAllCoreThreads();
|
((ThreadPoolExecutor)executor).prestartAllCoreThreads();
|
||||||
Future<Boolean>[] futures = new Future[concurrentRenames];
|
Future<Boolean>[] futures = new Future[concurrentRenames];
|
||||||
for (int i = 0; i < concurrentRenames; i++) {
|
for (int i = 0; i < concurrentRenames; i++) {
|
||||||
@ -164,5 +148,66 @@ public Boolean call() throws Exception {
|
|||||||
assertPathDoesNotExist("source path", source[i]);
|
assertPathDoesNotExist("source path", source[i]);
|
||||||
}
|
}
|
||||||
LOG.info("All tasks have completed successfully");
|
LOG.info("All tasks have completed successfully");
|
||||||
|
} finally {
|
||||||
|
executor.shutdown();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Attempts to trigger a deadlock that would happen if any bounded resource
|
||||||
|
* pool became saturated with control tasks that depended on other tasks
|
||||||
|
* that now can't enter the resource pool to get completed.
|
||||||
|
*/
|
||||||
|
@Test
|
||||||
|
@SuppressWarnings("unchecked")
|
||||||
|
public void testParallelRename() throws InterruptedException,
|
||||||
|
ExecutionException, IOException {
|
||||||
|
|
||||||
|
Configuration conf = getConfiguration();
|
||||||
|
conf.setInt(MAX_THREADS, 2);
|
||||||
|
conf.setInt(MAX_TOTAL_TASKS, 1);
|
||||||
|
|
||||||
|
conf.set(MIN_MULTIPART_THRESHOLD, "10K");
|
||||||
|
conf.set(MULTIPART_SIZE, "5K");
|
||||||
|
|
||||||
|
try (S3AFileSystem tinyThreadPoolFs = new S3AFileSystem()) {
|
||||||
|
tinyThreadPoolFs.initialize(auxFs.getUri(), conf);
|
||||||
|
|
||||||
|
parallelRenames(concurrentRenames, tinyThreadPoolFs,
|
||||||
|
"testParallelRename-source", "testParallelRename-target");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testThreadPoolCoolDown() throws InterruptedException,
|
||||||
|
ExecutionException, IOException {
|
||||||
|
|
||||||
|
int hotThreads = 0;
|
||||||
|
int coldThreads = 0;
|
||||||
|
|
||||||
|
parallelRenames(concurrentRenames, auxFs,
|
||||||
|
"testThreadPoolCoolDown-source", "testThreadPoolCoolDown-target");
|
||||||
|
|
||||||
|
for (Thread t : Thread.getAllStackTraces().keySet()) {
|
||||||
|
if (t.getName().startsWith("s3a-transfer")) {
|
||||||
|
hotThreads++;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
int timeoutMs = Constants.DEFAULT_KEEPALIVE_TIME * 1000;
|
||||||
|
Thread.sleep((int)(1.1 * timeoutMs));
|
||||||
|
|
||||||
|
for (Thread t : Thread.getAllStackTraces().keySet()) {
|
||||||
|
if (t.getName().startsWith("s3a-transfer")) {
|
||||||
|
coldThreads++;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
assertNotEquals("Failed to find threads in active FS - test is flawed",
|
||||||
|
hotThreads, 0);
|
||||||
|
assertTrue("s3a-transfer threads went from " + hotThreads + " to " +
|
||||||
|
coldThreads + ", should have gone to 0", 0 == coldThreads);
|
||||||
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user