HADOOP-18458: AliyunOSSBlockOutputStream to support heap/off-heap buffer before uploading data to OSS (#4912)

This commit is contained in:
Jinhu Wu 2023-03-28 14:27:01 +08:00 committed by GitHub
parent 926993cb73
commit b5e8269d9b
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
11 changed files with 1830 additions and 116 deletions

View File

@ -19,6 +19,7 @@
package org.apache.hadoop.fs.aliyun.oss; package org.apache.hadoop.fs.aliyun.oss;
import com.aliyun.oss.model.PartETag; import com.aliyun.oss.model.PartETag;
import org.apache.hadoop.fs.aliyun.oss.statistics.BlockOutputStreamStatistics;
import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.Futures; import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.Futures;
import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.ListenableFuture; import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.ListenableFuture;
import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.ListeningExecutorService; import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.ListeningExecutorService;
@ -27,17 +28,15 @@
import org.slf4j.Logger; import org.slf4j.Logger;
import org.slf4j.LoggerFactory; import org.slf4j.LoggerFactory;
import java.io.BufferedOutputStream;
import java.io.File;
import java.io.FileOutputStream;
import java.io.IOException; import java.io.IOException;
import java.io.OutputStream; import java.io.OutputStream;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.HashMap;
import java.util.List; import java.util.List;
import java.util.Map;
import java.util.concurrent.ExecutionException; import java.util.concurrent.ExecutionException;
import java.util.concurrent.ExecutorService; import java.util.concurrent.ExecutorService;
import java.util.concurrent.atomic.AtomicBoolean;
import static org.apache.hadoop.io.IOUtils.cleanupWithLogger;
/** /**
* Asynchronous multi-part based uploading mechanism to support huge file * Asynchronous multi-part based uploading mechanism to support huge file
@ -49,71 +48,103 @@ public class AliyunOSSBlockOutputStream extends OutputStream {
LoggerFactory.getLogger(AliyunOSSBlockOutputStream.class); LoggerFactory.getLogger(AliyunOSSBlockOutputStream.class);
private AliyunOSSFileSystemStore store; private AliyunOSSFileSystemStore store;
private Configuration conf; private Configuration conf;
private boolean closed; private final AtomicBoolean closed = new AtomicBoolean(false);
private String key; private String key;
private File blockFile; private int blockSize;
private Map<Integer, File> blockFiles = new HashMap<>();
private long blockSize;
private int blockId = 0; private int blockId = 0;
private long blockWritten = 0L; private long blockWritten = 0L;
private String uploadId = null; private String uploadId = null;
private final List<ListenableFuture<PartETag>> partETagsFutures; private final List<ListenableFuture<PartETag>> partETagsFutures;
private final OSSDataBlocks.BlockFactory blockFactory;
private final BlockOutputStreamStatistics statistics;
private OSSDataBlocks.DataBlock activeBlock;
private final ListeningExecutorService executorService; private final ListeningExecutorService executorService;
private OutputStream blockStream;
private final byte[] singleByte = new byte[1]; private final byte[] singleByte = new byte[1];
public AliyunOSSBlockOutputStream(Configuration conf, public AliyunOSSBlockOutputStream(Configuration conf,
AliyunOSSFileSystemStore store, AliyunOSSFileSystemStore store,
String key, String key,
Long blockSize, int blockSize,
OSSDataBlocks.BlockFactory blockFactory,
BlockOutputStreamStatistics statistics,
ExecutorService executorService) throws IOException { ExecutorService executorService) throws IOException {
this.store = store; this.store = store;
this.conf = conf; this.conf = conf;
this.key = key; this.key = key;
this.blockSize = blockSize; this.blockSize = blockSize;
this.blockFile = newBlockFile(); this.blockFactory = blockFactory;
this.blockStream = this.statistics = statistics;
new BufferedOutputStream(new FileOutputStream(blockFile));
this.partETagsFutures = new ArrayList<>(2); this.partETagsFutures = new ArrayList<>(2);
this.executorService = MoreExecutors.listeningDecorator(executorService); this.executorService = MoreExecutors.listeningDecorator(executorService);
} }
private File newBlockFile() throws IOException { /**
return AliyunOSSUtils.createTmpFileForWrite( * Demand create a destination block.
String.format("oss-block-%04d-", blockId), blockSize, conf); * @return the active block; null if there isn't one.
* @throws IOException on any failure to create
*/
private synchronized OSSDataBlocks.DataBlock createBlockIfNeeded()
throws IOException {
if (activeBlock == null) {
blockId++;
activeBlock = blockFactory.create(blockId, blockSize, statistics);
}
return activeBlock;
} }
/**
* Check for the filesystem being open.
* @throws IOException if the filesystem is closed.
*/
void checkOpen() throws IOException {
if (closed.get()) {
throw new IOException("Stream closed.");
}
}
/**
* The flush operation does not trigger an upload; that awaits
* the next block being full. What it does do is call {@code flush() }
* on the current block, leaving it to choose how to react.
* @throws IOException Any IO problem.
*/
@Override @Override
public synchronized void flush() throws IOException { public synchronized void flush() throws IOException {
blockStream.flush(); checkOpen();
OSSDataBlocks.DataBlock dataBlock = getActiveBlock();
if (dataBlock != null) {
dataBlock.flush();
}
} }
@Override @Override
public synchronized void close() throws IOException { public synchronized void close() throws IOException {
if (closed) { if (closed.get()) {
// already closed
LOG.debug("Ignoring close() as stream is already closed");
return; return;
} }
blockStream.flush();
blockStream.close();
if (!blockFiles.values().contains(blockFile)) {
blockId++;
blockFiles.put(blockId, blockFile);
}
try { try {
if (blockFiles.size() == 1) { if (uploadId == null) {
// just upload it directly // just upload it directly
store.uploadObject(key, blockFile); OSSDataBlocks.DataBlock dataBlock = getActiveBlock();
if (dataBlock == null) {
// zero size file
store.storeEmptyFile(key);
} else {
OSSDataBlocks.BlockUploadData uploadData = dataBlock.startUpload();
if (uploadData.hasFile()) {
store.uploadObject(key, uploadData.getFile());
} else {
store.uploadObject(key,
uploadData.getUploadStream(), dataBlock.dataSize());
}
}
} else { } else {
if (blockWritten > 0) { if (blockWritten > 0) {
ListenableFuture<PartETag> partETagFuture = uploadCurrentBlock();
executorService.submit(() -> {
PartETag partETag = store.uploadPart(blockFile, key, uploadId,
blockId);
return partETag;
});
partETagsFutures.add(partETagFuture);
} }
// wait for the partial uploads to finish // wait for the partial uploads to finish
final List<PartETag> partETags = waitForAllPartUploads(); final List<PartETag> partETags = waitForAllPartUploads();
@ -124,8 +155,8 @@ public synchronized void close() throws IOException {
new ArrayList<>(partETags)); new ArrayList<>(partETags));
} }
} finally { } finally {
removeTemporaryFiles(); cleanupWithLogger(LOG, getActiveBlock(), blockFactory);
closed = true; closed.set(true);
} }
} }
@ -138,64 +169,82 @@ public synchronized void write(int b) throws IOException {
@Override @Override
public synchronized void write(byte[] b, int off, int len) public synchronized void write(byte[] b, int off, int len)
throws IOException { throws IOException {
if (closed) { int totalWritten = 0;
throw new IOException("Stream closed."); while (totalWritten < len) {
int written = internalWrite(b, off + totalWritten, len - totalWritten);
totalWritten += written;
LOG.debug("Buffer len {}, written {}, total written {}",
len, written, totalWritten);
} }
blockStream.write(b, off, len); }
blockWritten += len; private synchronized int internalWrite(byte[] b, int off, int len)
if (blockWritten >= blockSize) { throws IOException {
uploadCurrentPart(); OSSDataBlocks.validateWriteArgs(b, off, len);
blockWritten = 0L; checkOpen();
if (len == 0) {
return 0;
}
OSSDataBlocks.DataBlock block = createBlockIfNeeded();
int written = block.write(b, off, len);
blockWritten += written;
int remainingCapacity = block.remainingCapacity();
if (written < len) {
// not everything was written the block has run out
// of capacity
// Trigger an upload then process the remainder.
LOG.debug("writing more data than block has capacity -triggering upload");
uploadCurrentBlock();
} else {
if (remainingCapacity == 0) {
// the whole buffer is done, trigger an upload
uploadCurrentBlock();
}
}
return written;
}
/**
* Clear the active block.
*/
private void clearActiveBlock() {
if (activeBlock != null) {
LOG.debug("Clearing active block");
}
synchronized (this) {
activeBlock = null;
} }
} }
private void removeTemporaryFiles() { private synchronized OSSDataBlocks.DataBlock getActiveBlock() {
for (File file : blockFiles.values()) { return activeBlock;
if (file != null && file.exists() && !file.delete()) {
LOG.warn("Failed to delete temporary file {}", file);
}
}
} }
private void removePartFiles() throws IOException { private void uploadCurrentBlock()
for (ListenableFuture<PartETag> partETagFuture : partETagsFutures) { throws IOException {
if (!partETagFuture.isDone()) { if (uploadId == null) {
continue;
}
try {
File blockFile = blockFiles.get(partETagFuture.get().getPartNumber());
if (blockFile != null && blockFile.exists() && !blockFile.delete()) {
LOG.warn("Failed to delete temporary file {}", blockFile);
}
} catch (InterruptedException | ExecutionException e) {
throw new IOException(e);
}
}
}
private void uploadCurrentPart() throws IOException {
blockStream.flush();
blockStream.close();
if (blockId == 0) {
uploadId = store.getUploadId(key); uploadId = store.getUploadId(key);
} }
blockId++;
blockFiles.put(blockId, blockFile);
File currentFile = blockFile;
int currentBlockId = blockId; int currentBlockId = blockId;
ListenableFuture<PartETag> partETagFuture = OSSDataBlocks.DataBlock dataBlock = getActiveBlock();
executorService.submit(() -> { long size = dataBlock.dataSize();
PartETag partETag = store.uploadPart(currentFile, key, uploadId, OSSDataBlocks.BlockUploadData uploadData = dataBlock.startUpload();
currentBlockId); try {
return partETag; ListenableFuture<PartETag> partETagFuture =
}); executorService.submit(() -> {
partETagsFutures.add(partETagFuture); try {
removePartFiles(); PartETag partETag = store.uploadPart(uploadData, size, key,
blockFile = newBlockFile(); uploadId, currentBlockId);
blockStream = new BufferedOutputStream(new FileOutputStream(blockFile)); return partETag;
} finally {
cleanupWithLogger(LOG, uploadData, dataBlock);
}
});
partETagsFutures.add(partETagFuture);
} finally {
blockWritten = 0;
clearActiveBlock();
}
} }
/** /**

View File

@ -27,6 +27,9 @@
import java.util.concurrent.ExecutorService; import java.util.concurrent.ExecutorService;
import java.util.concurrent.TimeUnit; import java.util.concurrent.TimeUnit;
import org.apache.hadoop.classification.VisibleForTesting;
import org.apache.hadoop.fs.aliyun.oss.statistics.BlockOutputStreamStatistics;
import org.apache.hadoop.fs.aliyun.oss.statistics.impl.OutputStreamStatistics;
import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.MoreExecutors; import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.MoreExecutors;
import org.apache.commons.collections.CollectionUtils; import org.apache.commons.collections.CollectionUtils;
import org.apache.commons.lang3.StringUtils; import org.apache.commons.lang3.StringUtils;
@ -71,6 +74,9 @@ public class AliyunOSSFileSystem extends FileSystem {
private String bucket; private String bucket;
private String username; private String username;
private Path workingDir; private Path workingDir;
private OSSDataBlocks.BlockFactory blockFactory;
private BlockOutputStreamStatistics blockOutputStreamStatistics;
private int uploadPartSize;
private int blockOutputActiveBlocks; private int blockOutputActiveBlocks;
private AliyunOSSFileSystemStore store; private AliyunOSSFileSystemStore store;
private int maxKeys; private int maxKeys;
@ -128,13 +134,13 @@ public FSDataOutputStream create(Path path, FsPermission permission,
// this means the file is not found // this means the file is not found
} }
long uploadPartSize = AliyunOSSUtils.getMultipartSizeProperty(getConf(),
MULTIPART_UPLOAD_PART_SIZE_KEY, MULTIPART_UPLOAD_PART_SIZE_DEFAULT);
return new FSDataOutputStream( return new FSDataOutputStream(
new AliyunOSSBlockOutputStream(getConf(), new AliyunOSSBlockOutputStream(getConf(),
store, store,
key, key,
uploadPartSize, uploadPartSize,
blockFactory,
blockOutputStreamStatistics,
new SemaphoredDelegatingExecutor(boundedThreadPool, new SemaphoredDelegatingExecutor(boundedThreadPool,
blockOutputActiveBlocks, true)), statistics); blockOutputActiveBlocks, true)), statistics);
} }
@ -334,6 +340,7 @@ public String getCanonicalServiceName() {
*/ */
public void initialize(URI name, Configuration conf) throws IOException { public void initialize(URI name, Configuration conf) throws IOException {
super.initialize(name, conf); super.initialize(name, conf);
setConf(conf);
bucket = name.getHost(); bucket = name.getHost();
uri = java.net.URI.create(name.getScheme() + "://" + name.getAuthority()); uri = java.net.URI.create(name.getScheme() + "://" + name.getAuthority());
@ -345,6 +352,16 @@ public void initialize(URI name, Configuration conf) throws IOException {
blockOutputActiveBlocks = intOption(conf, blockOutputActiveBlocks = intOption(conf,
UPLOAD_ACTIVE_BLOCKS_KEY, UPLOAD_ACTIVE_BLOCKS_DEFAULT, 1); UPLOAD_ACTIVE_BLOCKS_KEY, UPLOAD_ACTIVE_BLOCKS_DEFAULT, 1);
uploadPartSize = (int)AliyunOSSUtils.getMultipartSizeProperty(conf,
MULTIPART_UPLOAD_PART_SIZE_KEY, MULTIPART_UPLOAD_PART_SIZE_DEFAULT);
String uploadBuffer = conf.getTrimmed(FAST_UPLOAD_BUFFER,
DEFAULT_FAST_UPLOAD_BUFFER);
blockOutputStreamStatistics = new OutputStreamStatistics();
blockFactory = OSSDataBlocks.createFactory(this, uploadBuffer);
LOG.debug("Using OSSBlockOutputStream with buffer = {}; block={};" +
" queue limit={}",
uploadBuffer, uploadPartSize, blockOutputActiveBlocks);
store = new AliyunOSSFileSystemStore(); store = new AliyunOSSFileSystemStore();
store.initialize(name, conf, username, statistics); store.initialize(name, conf, username, statistics);
maxKeys = conf.getInt(MAX_PAGING_KEYS_KEY, MAX_PAGING_KEYS_DEFAULT); maxKeys = conf.getInt(MAX_PAGING_KEYS_KEY, MAX_PAGING_KEYS_DEFAULT);
@ -379,8 +396,6 @@ public void initialize(URI name, Configuration conf) throws IOException {
this.boundedCopyThreadPool = BlockingThreadPoolExecutorService.newInstance( this.boundedCopyThreadPool = BlockingThreadPoolExecutorService.newInstance(
maxCopyThreads, maxCopyTasks, 60L, maxCopyThreads, maxCopyTasks, 60L,
TimeUnit.SECONDS, "oss-copy-unbounded"); TimeUnit.SECONDS, "oss-copy-unbounded");
setConf(conf);
} }
/** /**
@ -757,4 +772,14 @@ public void setWorkingDirectory(Path dir) {
public AliyunOSSFileSystemStore getStore() { public AliyunOSSFileSystemStore getStore() {
return store; return store;
} }
@VisibleForTesting
OSSDataBlocks.BlockFactory getBlockFactory() {
return blockFactory;
}
@VisibleForTesting
BlockOutputStreamStatistics getBlockOutputStreamStatistics() {
return blockOutputStreamStatistics;
}
} }

View File

@ -422,6 +422,27 @@ public void uploadObject(String key, File file) throws IOException {
} }
} }
/**
* Upload an input stream as an OSS object, using single upload.
* @param key object key.
* @param in input stream to upload.
* @param size size of the input stream.
* @throws IOException if failed to upload object.
*/
public void uploadObject(String key, InputStream in, long size)
throws IOException {
ObjectMetadata meta = new ObjectMetadata();
meta.setContentLength(size);
if (StringUtils.isNotEmpty(serverSideEncryptionAlgorithm)) {
meta.setServerSideEncryption(serverSideEncryptionAlgorithm);
}
PutObjectResult result = ossClient.putObject(bucketName, key, in, meta);
LOG.debug(result.getETag());
statistics.incrementWriteOps(1);
}
/** /**
* list objects. * list objects.
* *
@ -652,44 +673,58 @@ private boolean continueListStatus() {
}; };
} }
public PartETag uploadPart(OSSDataBlocks.BlockUploadData partData,
long size, String key, String uploadId, int idx) throws IOException {
if (partData.hasFile()) {
return uploadPart(partData.getFile(), key, uploadId, idx);
} else {
return uploadPart(partData.getUploadStream(), size, key, uploadId, idx);
}
}
public PartETag uploadPart(File file, String key, String uploadId, int idx) public PartETag uploadPart(File file, String key, String uploadId, int idx)
throws IOException { throws IOException {
InputStream instream = null; InputStream in = new FileInputStream(file);
try {
return uploadPart(in, file.length(), key, uploadId, idx);
} finally {
in.close();
}
}
public PartETag uploadPart(InputStream in, long size, String key,
String uploadId, int idx) throws IOException {
Exception caught = null; Exception caught = null;
int tries = 3; int tries = 3;
while (tries > 0) { while (tries > 0) {
try { try {
instream = new FileInputStream(file);
UploadPartRequest uploadRequest = new UploadPartRequest(); UploadPartRequest uploadRequest = new UploadPartRequest();
uploadRequest.setBucketName(bucketName); uploadRequest.setBucketName(bucketName);
uploadRequest.setKey(key); uploadRequest.setKey(key);
uploadRequest.setUploadId(uploadId); uploadRequest.setUploadId(uploadId);
uploadRequest.setInputStream(instream); uploadRequest.setInputStream(in);
uploadRequest.setPartSize(file.length()); uploadRequest.setPartSize(size);
uploadRequest.setPartNumber(idx); uploadRequest.setPartNumber(idx);
UploadPartResult uploadResult = ossClient.uploadPart(uploadRequest); UploadPartResult uploadResult = ossClient.uploadPart(uploadRequest);
statistics.incrementWriteOps(1); statistics.incrementWriteOps(1);
return uploadResult.getPartETag(); return uploadResult.getPartETag();
} catch (Exception e) { } catch (Exception e) {
LOG.debug("Failed to upload "+ file.getPath() +", " + LOG.debug("Failed to upload " + key + ", part " + idx +
"try again.", e); "try again.", e);
caught = e; caught = e;
} finally {
if (instream != null) {
instream.close();
instream = null;
}
} }
tries--; tries--;
} }
assert (caught != null); assert (caught != null);
throw new IOException("Failed to upload " + file.getPath() + throw new IOException("Failed to upload " + key + ", part " + idx +
" for 3 times.", caught); " for 3 times.", caught);
} }
/** /**
* Initiate multipart upload. * Initiate multipart upload.
* @param key object key.
* @return upload id.
*/ */
public String getUploadId(String key) { public String getUploadId(String key) {
InitiateMultipartUploadRequest initiateMultipartUploadRequest = InitiateMultipartUploadRequest initiateMultipartUploadRequest =
@ -701,6 +736,10 @@ public String getUploadId(String key) {
/** /**
* Complete the specific multipart upload. * Complete the specific multipart upload.
* @param key object key.
* @param uploadId upload id of this multipart upload.
* @param partETags part etags need to be completed.
* @return CompleteMultipartUploadResult.
*/ */
public CompleteMultipartUploadResult completeMultipartUpload(String key, public CompleteMultipartUploadResult completeMultipartUpload(String key,
String uploadId, List<PartETag> partETags) { String uploadId, List<PartETag> partETags) {
@ -713,6 +752,8 @@ public CompleteMultipartUploadResult completeMultipartUpload(String key,
/** /**
* Abort the specific multipart upload. * Abort the specific multipart upload.
* @param key object key.
* @param uploadId upload id of this multipart upload.
*/ */
public void abortMultipartUpload(String key, String uploadId) { public void abortMultipartUpload(String key, String uploadId) {
AbortMultipartUploadRequest request = new AbortMultipartUploadRequest( AbortMultipartUploadRequest request = new AbortMultipartUploadRequest(

View File

@ -134,6 +134,59 @@ private Constants() {
// Comma separated list of directories // Comma separated list of directories
public static final String BUFFER_DIR_KEY = "fs.oss.buffer.dir"; public static final String BUFFER_DIR_KEY = "fs.oss.buffer.dir";
/**
* What buffer to use.
* Default is {@link #FAST_UPLOAD_BUFFER_DISK}
* Value: {@value}
*/
public static final String FAST_UPLOAD_BUFFER =
"fs.oss.fast.upload.buffer";
/**
* Buffer blocks to disk: {@value}.
* Capacity is limited to available disk space.
*/
public static final String FAST_UPLOAD_BUFFER_DISK = "disk";
/**
* Use an in-memory array. Fast but will run of heap rapidly: {@value}.
*/
public static final String FAST_UPLOAD_BUFFER_ARRAY = "array";
/**
* Use a byte buffer. May be more memory efficient than the
* {@link #FAST_UPLOAD_BUFFER_ARRAY}: {@value}.
*/
public static final String FAST_UPLOAD_BYTEBUFFER = "bytebuffer";
/**
* Use an in-memory array and fallback to disk if
* used memory exceed the quota.
*/
public static final String FAST_UPLOAD_BUFFER_ARRAY_DISK = "array_disk";
/**
* Use a byte buffer and fallback to disk if
* used memory exceed the quota.
*/
public static final String FAST_UPLOAD_BYTEBUFFER_DISK = "bytebuffer_disk";
/**
* Memory limit of {@link #FAST_UPLOAD_BUFFER_ARRAY_DISK} or
* {@link #FAST_UPLOAD_BYTEBUFFER_DISK}.
*/
public static final String FAST_UPLOAD_BUFFER_MEMORY_LIMIT =
"fs.oss.fast.upload.memory.limit";
public static final long FAST_UPLOAD_BUFFER_MEMORY_LIMIT_DEFAULT =
1024 * 1024 * 1024; // 1GB
/**
* Default buffer option: {@value}.
*/
public static final String DEFAULT_FAST_UPLOAD_BUFFER =
FAST_UPLOAD_BUFFER_DISK;
// private | public-read | public-read-write // private | public-read | public-read-write
public static final String CANNED_ACL_KEY = "fs.oss.acl.default"; public static final String CANNED_ACL_KEY = "fs.oss.acl.default";
public static final String CANNED_ACL_DEFAULT = ""; public static final String CANNED_ACL_DEFAULT = "";

View File

@ -0,0 +1,72 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.fs.aliyun.oss.statistics;
import org.apache.hadoop.classification.InterfaceStability;
/**
* Block output stream statistics.
*/
@InterfaceStability.Unstable
public interface BlockOutputStreamStatistics {
/**
* A block has been allocated.
*/
void blockAllocated();
/**
* A block has been released.
*/
void blockReleased();
/**
* A disk block has been allocated.
*/
void diskBlockAllocated();
/**
* A disk block has been released.
*/
void diskBlockReleased();
/**
* Memory bytes has been allocated.
* @param size allocated size.
*/
void bytesAllocated(long size);
/**
* Memory bytes has been released.
* @param size released size.
*/
void bytesReleased(long size);
int getBlocksAllocated();
int getBlocksReleased();
int getDiskBlocksAllocated();
int getDiskBlocksReleased();
long getBytesAllocated();
long getBytesReleased();
}

View File

@ -0,0 +1,98 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.fs.aliyun.oss.statistics.impl;
import org.apache.hadoop.fs.aliyun.oss.statistics.BlockOutputStreamStatistics;
import java.util.concurrent.atomic.AtomicInteger;
import java.util.concurrent.atomic.AtomicLong;
/**
* Implementation of {@link BlockOutputStreamStatistics}.
*/
public class OutputStreamStatistics implements BlockOutputStreamStatistics {
private final AtomicInteger blocksAllocated = new AtomicInteger(0);
private final AtomicInteger blocksReleased = new AtomicInteger(0);
private final AtomicInteger diskBlocksAllocated = new AtomicInteger(0);
private final AtomicInteger diskBlocksReleased = new AtomicInteger(0);
private final AtomicLong bytesAllocated = new AtomicLong(0);
private final AtomicLong bytesReleased = new AtomicLong(0);
@Override
public void blockAllocated() {
blocksAllocated.incrementAndGet();
}
@Override
public void blockReleased() {
blocksReleased.incrementAndGet();
}
@Override
public void diskBlockAllocated() {
diskBlocksAllocated.incrementAndGet();
}
@Override
public void diskBlockReleased() {
diskBlocksReleased.incrementAndGet();
}
@Override
public int getBlocksAllocated() {
return blocksAllocated.get();
}
@Override
public int getBlocksReleased() {
return blocksReleased.get();
}
@Override
public int getDiskBlocksAllocated() {
return diskBlocksAllocated.get();
}
@Override
public int getDiskBlocksReleased() {
return diskBlocksReleased.get();
}
@Override
public void bytesAllocated(long size) {
bytesAllocated.getAndAdd(size);
}
@Override
public void bytesReleased(long size) {
bytesReleased.getAndAdd(size);
}
@Override
public long getBytesAllocated() {
return bytesAllocated.get();
}
@Override
public long getBytesReleased() {
return bytesReleased.get();
}
}

View File

@ -0,0 +1,29 @@
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/**
* Statistics collection for the OSS connector: implementation.
* Not for use by anything outside the hadoop-aliyun source tree.
*/
@InterfaceAudience.Private
@InterfaceStability.Unstable
package org.apache.hadoop.fs.aliyun.oss.statistics.impl;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.classification.InterfaceStability;

View File

@ -0,0 +1,27 @@
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/**
* Statistics collection for the OSS connector: interfaces.
*/
@InterfaceAudience.Private
@InterfaceStability.Unstable
package org.apache.hadoop.fs.aliyun.oss.statistics;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.classification.InterfaceStability;

View File

@ -164,7 +164,7 @@ please raise your issues with them.
<property> <property>
<name>fs.oss.attempts.maximum</name> <name>fs.oss.attempts.maximum</name>
<value>20</value> <value>10</value>
<description>How many times we should retry commands on transient errors.</description> <description>How many times we should retry commands on transient errors.</description>
</property> </property>
@ -239,7 +239,7 @@ please raise your issues with them.
<property> <property>
<name>fs.oss.multipart.download.size</name> <name>fs.oss.multipart.download.size</name>
<value>102400/value> <value>524288/value>
<description>Size in bytes in each request from ALiyun OSS.</description> <description>Size in bytes in each request from ALiyun OSS.</description>
</property> </property>
@ -251,9 +251,53 @@ please raise your issues with them.
</description> </description>
</property> </property>
<property>
<name>fs.oss.fast.upload.buffer</name>
<value>disk</value>
<description>
The buffering mechanism to use.
Values: disk, array, bytebuffer, array_disk, bytebuffer_disk.
"disk" will use the directories listed in fs.oss.buffer.dir as
the location(s) to save data prior to being uploaded.
"array" uses arrays in the JVM heap
"bytebuffer" uses off-heap memory within the JVM.
Both "array" and "bytebuffer" will consume memory in a single stream up to the number
of blocks set by:
fs.oss.multipart.upload.size * fs.oss.upload.active.blocks.
If using either of these mechanisms, keep this value low
The total number of threads performing work across all threads is set by
fs.oss.multipart.download.threads(Currently fast upload shares the same thread tool with download.
The thread pool size is specified in "fs.oss.multipart.download.threads"),
with fs.oss.max.total.tasks values setting the number of queued work items.
"array_disk" and "bytebuffer_disk" support fallback to disk.
</description>
</property>
<property>
<name>fs.oss.fast.upload.memory.limit</name>
<value>1073741824</value>
<description>
Memory limit of "array_disk" and "bytebuffer_disk" upload buffers.
Will fallback to disk buffers if used memory reaches the limit.
</description>
</property>
<property> <property>
<name>fs.oss.buffer.dir</name> <name>fs.oss.buffer.dir</name>
<description>Comma separated list of directories to buffer OSS data before uploading to Aliyun OSS</description> <value>${env.LOCAL_DIRS:-${hadoop.tmp.dir}}/oss</value>
<description>Comma separated list of directories to buffer
OSS data before uploading to Aliyun OSS.
Yarn container path will be used as default value on yarn applications,
otherwise fall back to hadoop.tmp.dir
</description>
</property> </property>
<property> <property>

View File

@ -22,6 +22,8 @@
import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.aliyun.oss.OSSDataBlocks.ByteBufferBlockFactory;
import org.apache.hadoop.fs.aliyun.oss.statistics.BlockOutputStreamStatistics;
import org.apache.hadoop.fs.contract.ContractTestUtils; import org.apache.hadoop.fs.contract.ContractTestUtils;
import org.apache.hadoop.test.GenericTestUtils; import org.apache.hadoop.test.GenericTestUtils;
import org.junit.After; import org.junit.After;
@ -37,12 +39,19 @@
import java.util.LinkedHashSet; import java.util.LinkedHashSet;
import static org.apache.hadoop.fs.aliyun.oss.Constants.BUFFER_DIR_KEY; import static org.apache.hadoop.fs.aliyun.oss.Constants.BUFFER_DIR_KEY;
import static org.apache.hadoop.fs.aliyun.oss.Constants.FAST_UPLOAD_BUFFER;
import static org.apache.hadoop.fs.aliyun.oss.Constants.FAST_UPLOAD_BUFFER_ARRAY_DISK;
import static org.apache.hadoop.fs.aliyun.oss.Constants.FAST_UPLOAD_BUFFER_DISK;
import static org.apache.hadoop.fs.aliyun.oss.Constants.FAST_UPLOAD_BUFFER_MEMORY_LIMIT;
import static org.apache.hadoop.fs.aliyun.oss.Constants.FAST_UPLOAD_BYTEBUFFER;
import static org.apache.hadoop.fs.aliyun.oss.Constants.FAST_UPLOAD_BYTEBUFFER_DISK;
import static org.apache.hadoop.fs.aliyun.oss.Constants.MULTIPART_UPLOAD_PART_SIZE_DEFAULT; import static org.apache.hadoop.fs.aliyun.oss.Constants.MULTIPART_UPLOAD_PART_SIZE_DEFAULT;
import static org.apache.hadoop.fs.aliyun.oss.Constants.MULTIPART_UPLOAD_PART_SIZE_KEY; import static org.apache.hadoop.fs.aliyun.oss.Constants.MULTIPART_UPLOAD_PART_SIZE_KEY;
import static org.apache.hadoop.fs.contract.ContractTestUtils.IO_CHUNK_BUFFER_SIZE; import static org.apache.hadoop.fs.contract.ContractTestUtils.IO_CHUNK_BUFFER_SIZE;
import static org.junit.Assert.assertEquals; import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertFalse; import static org.junit.Assert.assertFalse;
import static org.junit.Assert.assertNotEquals; import static org.junit.Assert.assertNotEquals;
import static org.junit.Assert.assertNotNull;
import static org.junit.Assert.assertTrue; import static org.junit.Assert.assertTrue;
/** /**
@ -54,6 +63,7 @@ public class TestAliyunOSSBlockOutputStream {
private static final int PART_SIZE = 1024 * 1024; private static final int PART_SIZE = 1024 * 1024;
private static String testRootPath = private static String testRootPath =
AliyunOSSTestUtils.generateUniqueTestPath(); AliyunOSSTestUtils.generateUniqueTestPath();
private static final long MEMORY_LIMIT = 10 * 1024 * 1024;
@Rule @Rule
public Timeout testTimeout = new Timeout(30 * 60 * 1000); public Timeout testTimeout = new Timeout(30 * 60 * 1000);
@ -65,6 +75,7 @@ public void setUp() throws Exception {
conf.setInt(IO_CHUNK_BUFFER_SIZE, conf.setInt(IO_CHUNK_BUFFER_SIZE,
conf.getInt(MULTIPART_UPLOAD_PART_SIZE_KEY, 0)); conf.getInt(MULTIPART_UPLOAD_PART_SIZE_KEY, 0));
conf.setInt(Constants.UPLOAD_ACTIVE_BLOCKS_KEY, 20); conf.setInt(Constants.UPLOAD_ACTIVE_BLOCKS_KEY, 20);
conf.setLong(FAST_UPLOAD_BUFFER_MEMORY_LIMIT, MEMORY_LIMIT);
fs = AliyunOSSTestUtils.createTestFileSystem(conf); fs = AliyunOSSTestUtils.createTestFileSystem(conf);
} }
@ -82,7 +93,7 @@ private Path getTestPath() {
@Test @Test
public void testZeroByteUpload() throws IOException { public void testZeroByteUpload() throws IOException {
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), 0); ContractTestUtils.createAndVerifyFile(fs, getTestPath(), 0);
bufferDirShouldEmpty(); bufferShouldReleased(true);
} }
@Test @Test
@ -106,20 +117,21 @@ public void testRegularUpload() throws IOException {
assertEquals(size - 1, statistics.getBytesRead()); assertEquals(size - 1, statistics.getBytesRead());
assertEquals(3, statistics.getWriteOps()); assertEquals(3, statistics.getWriteOps());
assertEquals(size - 1, statistics.getBytesWritten()); assertEquals(size - 1, statistics.getBytesWritten());
bufferShouldReleased();
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size); ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size);
assertEquals(14, statistics.getReadOps()); assertEquals(14, statistics.getReadOps());
assertEquals(2 * size - 1, statistics.getBytesRead()); assertEquals(2 * size - 1, statistics.getBytesRead());
assertEquals(6, statistics.getWriteOps()); assertEquals(6, statistics.getWriteOps());
assertEquals(2 * size - 1, statistics.getBytesWritten()); assertEquals(2 * size - 1, statistics.getBytesWritten());
bufferShouldReleased();
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size + 1); ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size + 1);
assertEquals(22, statistics.getReadOps()); assertEquals(22, statistics.getReadOps());
assertEquals(3 * size, statistics.getBytesRead()); assertEquals(3 * size, statistics.getBytesRead());
assertEquals(10, statistics.getWriteOps()); assertEquals(10, statistics.getWriteOps());
assertEquals(3 * size, statistics.getBytesWritten()); assertEquals(3 * size, statistics.getBytesWritten());
bufferDirShouldEmpty(); bufferShouldReleased();
} }
@Test @Test
@ -133,19 +145,21 @@ public void testMultiPartUpload() throws IOException {
assertEquals(size - 1, statistics.getBytesRead()); assertEquals(size - 1, statistics.getBytesRead());
assertEquals(8, statistics.getWriteOps()); assertEquals(8, statistics.getWriteOps());
assertEquals(size - 1, statistics.getBytesWritten()); assertEquals(size - 1, statistics.getBytesWritten());
bufferShouldReleased();
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size); ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size);
assertEquals(34, statistics.getReadOps()); assertEquals(34, statistics.getReadOps());
assertEquals(2 * size - 1, statistics.getBytesRead()); assertEquals(2 * size - 1, statistics.getBytesRead());
assertEquals(16, statistics.getWriteOps()); assertEquals(16, statistics.getWriteOps());
assertEquals(2 * size - 1, statistics.getBytesWritten()); assertEquals(2 * size - 1, statistics.getBytesWritten());
bufferShouldReleased();
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size + 1); ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size + 1);
assertEquals(52, statistics.getReadOps()); assertEquals(52, statistics.getReadOps());
assertEquals(3 * size, statistics.getBytesRead()); assertEquals(3 * size, statistics.getBytesRead());
assertEquals(25, statistics.getWriteOps()); assertEquals(25, statistics.getWriteOps());
assertEquals(3 * size, statistics.getBytesWritten()); assertEquals(3 * size, statistics.getBytesWritten());
bufferDirShouldEmpty(); bufferShouldReleased();
} }
@Test @Test
@ -159,16 +173,18 @@ public void testMultiPartUploadConcurrent() throws IOException {
assertEquals(size, statistics.getBytesRead()); assertEquals(size, statistics.getBytesRead());
assertEquals(52, statistics.getWriteOps()); assertEquals(52, statistics.getWriteOps());
assertEquals(size, statistics.getBytesWritten()); assertEquals(size, statistics.getBytesWritten());
bufferDirShouldEmpty(); bufferShouldReleased();
} }
@Test @Test
public void testHugeUpload() throws IOException { public void testHugeUpload() throws IOException {
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), PART_SIZE - 1); ContractTestUtils.createAndVerifyFile(fs, getTestPath(), PART_SIZE - 1);
bufferShouldReleased();
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), PART_SIZE); ContractTestUtils.createAndVerifyFile(fs, getTestPath(), PART_SIZE);
bufferShouldReleased();
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), ContractTestUtils.createAndVerifyFile(fs, getTestPath(),
MULTIPART_UPLOAD_PART_SIZE_DEFAULT + 1); MULTIPART_UPLOAD_PART_SIZE_DEFAULT + 1);
bufferDirShouldEmpty(); bufferShouldReleased();
} }
@Test @Test
@ -199,15 +215,43 @@ public void testMultiPartUploadLimit() throws IOException {
public void testSmallUpload() throws IOException { public void testSmallUpload() throws IOException {
long size = fs.getConf().getInt(MULTIPART_UPLOAD_PART_SIZE_KEY, 1024); long size = fs.getConf().getInt(MULTIPART_UPLOAD_PART_SIZE_KEY, 1024);
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size - 1); ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size - 1);
bufferDirShouldEmpty(); bufferShouldReleased();
} }
private void bufferDirShouldEmpty() throws IOException { private void bufferShouldReleased() throws IOException {
Path bufferPath = new Path(fs.getConf().get(BUFFER_DIR_KEY)); bufferShouldReleased(false);
FileStatus[] files = bufferPath.getFileSystem( }
fs.getConf()).listStatus(bufferPath);
// Temporary file should be deleted private void bufferShouldReleased(boolean zeroSizeFile) throws IOException {
assertEquals(0, files.length); String bufferDir = fs.getConf().get(BUFFER_DIR_KEY);
String bufferType = fs.getConf().get(FAST_UPLOAD_BUFFER);
if (bufferType.equals(FAST_UPLOAD_BUFFER_DISK)) {
assertNotNull(bufferDir);
Path bufferPath = new Path(fs.getConf().get(BUFFER_DIR_KEY));
FileStatus[] files = bufferPath.getFileSystem(
fs.getConf()).listStatus(bufferPath);
// Temporary file should be deleted
assertEquals(0, files.length);
} else {
if (bufferType.equals(FAST_UPLOAD_BYTEBUFFER)) {
OSSDataBlocks.ByteBufferBlockFactory
blockFactory = (OSSDataBlocks.ByteBufferBlockFactory)
((AliyunOSSFileSystem)fs).getBlockFactory();
assertEquals("outstanding buffers in " + blockFactory,
0, blockFactory.getOutstandingBufferCount());
}
}
BlockOutputStreamStatistics statistics =
((AliyunOSSFileSystem)fs).getBlockOutputStreamStatistics();
assertEquals(statistics.getBlocksAllocated(),
statistics.getBlocksReleased());
if (zeroSizeFile) {
assertEquals(statistics.getBlocksAllocated(), 0);
} else {
assertTrue(statistics.getBlocksAllocated() >= 1);
}
assertEquals(statistics.getBytesReleased(),
statistics.getBytesAllocated());
} }
@Test @Test
@ -249,4 +293,127 @@ public void testDirectoryAllocatorRR() throws Throwable {
assertNotEquals("round robin not working", assertNotEquals("round robin not working",
tmp1.getParent(), tmp2.getParent()); tmp1.getParent(), tmp2.getParent());
} }
@Test
public void testByteBufferIO() throws IOException {
try (OSSDataBlocks.ByteBufferBlockFactory factory =
new OSSDataBlocks.ByteBufferBlockFactory((AliyunOSSFileSystem)fs)) {
int limit = 128;
OSSDataBlocks.ByteBufferBlockFactory.ByteBufferBlock block
= factory.create(1, limit, null);
assertEquals("outstanding buffers in " + factory,
1, factory.getOutstandingBufferCount());
byte[] buffer = ContractTestUtils.toAsciiByteArray("test data");
int bufferLen = buffer.length;
block.write(buffer, 0, bufferLen);
assertEquals(bufferLen, block.dataSize());
assertEquals("capacity in " + block,
limit - bufferLen, block.remainingCapacity());
assertTrue("hasCapacity(64) in " + block, block.hasCapacity(64));
assertTrue("No capacity in " + block,
block.hasCapacity(limit - bufferLen));
// now start the write
OSSDataBlocks.BlockUploadData blockUploadData = block.startUpload();
ByteBufferBlockFactory.ByteBufferBlock.ByteBufferInputStream
stream =
(ByteBufferBlockFactory.ByteBufferBlock.ByteBufferInputStream)
blockUploadData.getUploadStream();
assertTrue("Mark not supported in " + stream, stream.markSupported());
assertTrue("!hasRemaining() in " + stream, stream.hasRemaining());
int expected = bufferLen;
assertEquals("wrong available() in " + stream,
expected, stream.available());
assertEquals('t', stream.read());
stream.mark(limit);
expected--;
assertEquals("wrong available() in " + stream,
expected, stream.available());
// read into a byte array with an offset
int offset = 5;
byte[] in = new byte[limit];
assertEquals(2, stream.read(in, offset, 2));
assertEquals('e', in[offset]);
assertEquals('s', in[offset + 1]);
expected -= 2;
assertEquals("wrong available() in " + stream,
expected, stream.available());
// read to end
byte[] remainder = new byte[limit];
int c;
int index = 0;
while ((c = stream.read()) >= 0) {
remainder[index++] = (byte) c;
}
assertEquals(expected, index);
assertEquals('a', remainder[--index]);
assertEquals("wrong available() in " + stream,
0, stream.available());
assertTrue("hasRemaining() in " + stream, !stream.hasRemaining());
// go the mark point
stream.reset();
assertEquals('e', stream.read());
// when the stream is closed, the data should be returned
stream.close();
assertEquals("outstanding buffers in " + factory,
1, factory.getOutstandingBufferCount());
block.close();
assertEquals("outstanding buffers in " + factory,
0, factory.getOutstandingBufferCount());
stream.close();
assertEquals("outstanding buffers in " + factory,
0, factory.getOutstandingBufferCount());
}
}
@Test
public void testFastUploadArrayDisk() throws IOException {
testFastUploadFallback(FAST_UPLOAD_BUFFER_ARRAY_DISK);
}
@Test
public void testFastUploadByteBufferDisk() throws IOException {
testFastUploadFallback(FAST_UPLOAD_BYTEBUFFER_DISK);
}
private void testFastUploadFallback(String name) throws IOException {
Configuration conf = fs.getConf();
fs.close();
conf.set(FAST_UPLOAD_BUFFER, name);
fs = AliyunOSSTestUtils.createTestFileSystem(conf);
long size = 5 * MEMORY_LIMIT;
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size);
OSSDataBlocks.MemoryBlockFactory
blockFactory = ((OSSDataBlocks.MemoryAndDiskBlockFactory)
((AliyunOSSFileSystem)fs).getBlockFactory()).getMemoryFactory();
assertEquals(blockFactory.getMemoryUsed(), 0);
Path bufferPath = new Path(fs.getConf().get(BUFFER_DIR_KEY));
FileStatus[] files = bufferPath.getFileSystem(
fs.getConf()).listStatus(bufferPath);
// Temporary file should be deleted
assertEquals(0, files.length);
BlockOutputStreamStatistics statistics =
((AliyunOSSFileSystem)fs).getBlockOutputStreamStatistics();
assertEquals(statistics.getBlocksAllocated(),
statistics.getBlocksReleased());
assertTrue(statistics.getBlocksAllocated() > 1);
assertEquals(statistics.getBytesReleased(),
statistics.getBytesAllocated());
assertTrue(statistics.getBytesAllocated() >= MEMORY_LIMIT);
assertTrue(statistics.getDiskBlocksAllocated() > 0);
assertEquals(statistics.getDiskBlocksAllocated(),
statistics.getDiskBlocksReleased());
}
} }