HADOOP-18458: AliyunOSSBlockOutputStream to support heap/off-heap buffer before uploading data to OSS (#4912)
This commit is contained in:
parent
926993cb73
commit
b5e8269d9b
@ -19,6 +19,7 @@
|
|||||||
package org.apache.hadoop.fs.aliyun.oss;
|
package org.apache.hadoop.fs.aliyun.oss;
|
||||||
|
|
||||||
import com.aliyun.oss.model.PartETag;
|
import com.aliyun.oss.model.PartETag;
|
||||||
|
import org.apache.hadoop.fs.aliyun.oss.statistics.BlockOutputStreamStatistics;
|
||||||
import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.Futures;
|
import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.Futures;
|
||||||
import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.ListenableFuture;
|
import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.ListenableFuture;
|
||||||
import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.ListeningExecutorService;
|
import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.ListeningExecutorService;
|
||||||
@ -27,17 +28,15 @@
|
|||||||
import org.slf4j.Logger;
|
import org.slf4j.Logger;
|
||||||
import org.slf4j.LoggerFactory;
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
import java.io.BufferedOutputStream;
|
|
||||||
import java.io.File;
|
|
||||||
import java.io.FileOutputStream;
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.io.OutputStream;
|
import java.io.OutputStream;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.HashMap;
|
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.Map;
|
|
||||||
import java.util.concurrent.ExecutionException;
|
import java.util.concurrent.ExecutionException;
|
||||||
import java.util.concurrent.ExecutorService;
|
import java.util.concurrent.ExecutorService;
|
||||||
|
import java.util.concurrent.atomic.AtomicBoolean;
|
||||||
|
|
||||||
|
import static org.apache.hadoop.io.IOUtils.cleanupWithLogger;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Asynchronous multi-part based uploading mechanism to support huge file
|
* Asynchronous multi-part based uploading mechanism to support huge file
|
||||||
@ -49,71 +48,103 @@ public class AliyunOSSBlockOutputStream extends OutputStream {
|
|||||||
LoggerFactory.getLogger(AliyunOSSBlockOutputStream.class);
|
LoggerFactory.getLogger(AliyunOSSBlockOutputStream.class);
|
||||||
private AliyunOSSFileSystemStore store;
|
private AliyunOSSFileSystemStore store;
|
||||||
private Configuration conf;
|
private Configuration conf;
|
||||||
private boolean closed;
|
private final AtomicBoolean closed = new AtomicBoolean(false);
|
||||||
private String key;
|
private String key;
|
||||||
private File blockFile;
|
private int blockSize;
|
||||||
private Map<Integer, File> blockFiles = new HashMap<>();
|
|
||||||
private long blockSize;
|
|
||||||
private int blockId = 0;
|
private int blockId = 0;
|
||||||
private long blockWritten = 0L;
|
private long blockWritten = 0L;
|
||||||
private String uploadId = null;
|
private String uploadId = null;
|
||||||
private final List<ListenableFuture<PartETag>> partETagsFutures;
|
private final List<ListenableFuture<PartETag>> partETagsFutures;
|
||||||
|
private final OSSDataBlocks.BlockFactory blockFactory;
|
||||||
|
private final BlockOutputStreamStatistics statistics;
|
||||||
|
private OSSDataBlocks.DataBlock activeBlock;
|
||||||
private final ListeningExecutorService executorService;
|
private final ListeningExecutorService executorService;
|
||||||
private OutputStream blockStream;
|
|
||||||
private final byte[] singleByte = new byte[1];
|
private final byte[] singleByte = new byte[1];
|
||||||
|
|
||||||
public AliyunOSSBlockOutputStream(Configuration conf,
|
public AliyunOSSBlockOutputStream(Configuration conf,
|
||||||
AliyunOSSFileSystemStore store,
|
AliyunOSSFileSystemStore store,
|
||||||
String key,
|
String key,
|
||||||
Long blockSize,
|
int blockSize,
|
||||||
|
OSSDataBlocks.BlockFactory blockFactory,
|
||||||
|
BlockOutputStreamStatistics statistics,
|
||||||
ExecutorService executorService) throws IOException {
|
ExecutorService executorService) throws IOException {
|
||||||
this.store = store;
|
this.store = store;
|
||||||
this.conf = conf;
|
this.conf = conf;
|
||||||
this.key = key;
|
this.key = key;
|
||||||
this.blockSize = blockSize;
|
this.blockSize = blockSize;
|
||||||
this.blockFile = newBlockFile();
|
this.blockFactory = blockFactory;
|
||||||
this.blockStream =
|
this.statistics = statistics;
|
||||||
new BufferedOutputStream(new FileOutputStream(blockFile));
|
|
||||||
this.partETagsFutures = new ArrayList<>(2);
|
this.partETagsFutures = new ArrayList<>(2);
|
||||||
this.executorService = MoreExecutors.listeningDecorator(executorService);
|
this.executorService = MoreExecutors.listeningDecorator(executorService);
|
||||||
}
|
}
|
||||||
|
|
||||||
private File newBlockFile() throws IOException {
|
/**
|
||||||
return AliyunOSSUtils.createTmpFileForWrite(
|
* Demand create a destination block.
|
||||||
String.format("oss-block-%04d-", blockId), blockSize, conf);
|
* @return the active block; null if there isn't one.
|
||||||
|
* @throws IOException on any failure to create
|
||||||
|
*/
|
||||||
|
private synchronized OSSDataBlocks.DataBlock createBlockIfNeeded()
|
||||||
|
throws IOException {
|
||||||
|
if (activeBlock == null) {
|
||||||
|
blockId++;
|
||||||
|
activeBlock = blockFactory.create(blockId, blockSize, statistics);
|
||||||
|
}
|
||||||
|
return activeBlock;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Check for the filesystem being open.
|
||||||
|
* @throws IOException if the filesystem is closed.
|
||||||
|
*/
|
||||||
|
void checkOpen() throws IOException {
|
||||||
|
if (closed.get()) {
|
||||||
|
throw new IOException("Stream closed.");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* The flush operation does not trigger an upload; that awaits
|
||||||
|
* the next block being full. What it does do is call {@code flush() }
|
||||||
|
* on the current block, leaving it to choose how to react.
|
||||||
|
* @throws IOException Any IO problem.
|
||||||
|
*/
|
||||||
@Override
|
@Override
|
||||||
public synchronized void flush() throws IOException {
|
public synchronized void flush() throws IOException {
|
||||||
blockStream.flush();
|
checkOpen();
|
||||||
|
|
||||||
|
OSSDataBlocks.DataBlock dataBlock = getActiveBlock();
|
||||||
|
if (dataBlock != null) {
|
||||||
|
dataBlock.flush();
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public synchronized void close() throws IOException {
|
public synchronized void close() throws IOException {
|
||||||
if (closed) {
|
if (closed.get()) {
|
||||||
|
// already closed
|
||||||
|
LOG.debug("Ignoring close() as stream is already closed");
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
blockStream.flush();
|
|
||||||
blockStream.close();
|
|
||||||
if (!blockFiles.values().contains(blockFile)) {
|
|
||||||
blockId++;
|
|
||||||
blockFiles.put(blockId, blockFile);
|
|
||||||
}
|
|
||||||
|
|
||||||
try {
|
try {
|
||||||
if (blockFiles.size() == 1) {
|
if (uploadId == null) {
|
||||||
// just upload it directly
|
// just upload it directly
|
||||||
store.uploadObject(key, blockFile);
|
OSSDataBlocks.DataBlock dataBlock = getActiveBlock();
|
||||||
|
if (dataBlock == null) {
|
||||||
|
// zero size file
|
||||||
|
store.storeEmptyFile(key);
|
||||||
|
} else {
|
||||||
|
OSSDataBlocks.BlockUploadData uploadData = dataBlock.startUpload();
|
||||||
|
if (uploadData.hasFile()) {
|
||||||
|
store.uploadObject(key, uploadData.getFile());
|
||||||
|
} else {
|
||||||
|
store.uploadObject(key,
|
||||||
|
uploadData.getUploadStream(), dataBlock.dataSize());
|
||||||
|
}
|
||||||
|
}
|
||||||
} else {
|
} else {
|
||||||
if (blockWritten > 0) {
|
if (blockWritten > 0) {
|
||||||
ListenableFuture<PartETag> partETagFuture =
|
uploadCurrentBlock();
|
||||||
executorService.submit(() -> {
|
|
||||||
PartETag partETag = store.uploadPart(blockFile, key, uploadId,
|
|
||||||
blockId);
|
|
||||||
return partETag;
|
|
||||||
});
|
|
||||||
partETagsFutures.add(partETagFuture);
|
|
||||||
}
|
}
|
||||||
// wait for the partial uploads to finish
|
// wait for the partial uploads to finish
|
||||||
final List<PartETag> partETags = waitForAllPartUploads();
|
final List<PartETag> partETags = waitForAllPartUploads();
|
||||||
@ -124,8 +155,8 @@ public synchronized void close() throws IOException {
|
|||||||
new ArrayList<>(partETags));
|
new ArrayList<>(partETags));
|
||||||
}
|
}
|
||||||
} finally {
|
} finally {
|
||||||
removeTemporaryFiles();
|
cleanupWithLogger(LOG, getActiveBlock(), blockFactory);
|
||||||
closed = true;
|
closed.set(true);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -138,64 +169,82 @@ public synchronized void write(int b) throws IOException {
|
|||||||
@Override
|
@Override
|
||||||
public synchronized void write(byte[] b, int off, int len)
|
public synchronized void write(byte[] b, int off, int len)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
if (closed) {
|
int totalWritten = 0;
|
||||||
throw new IOException("Stream closed.");
|
while (totalWritten < len) {
|
||||||
|
int written = internalWrite(b, off + totalWritten, len - totalWritten);
|
||||||
|
totalWritten += written;
|
||||||
|
LOG.debug("Buffer len {}, written {}, total written {}",
|
||||||
|
len, written, totalWritten);
|
||||||
}
|
}
|
||||||
blockStream.write(b, off, len);
|
}
|
||||||
blockWritten += len;
|
private synchronized int internalWrite(byte[] b, int off, int len)
|
||||||
if (blockWritten >= blockSize) {
|
throws IOException {
|
||||||
uploadCurrentPart();
|
OSSDataBlocks.validateWriteArgs(b, off, len);
|
||||||
blockWritten = 0L;
|
checkOpen();
|
||||||
|
if (len == 0) {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
OSSDataBlocks.DataBlock block = createBlockIfNeeded();
|
||||||
|
int written = block.write(b, off, len);
|
||||||
|
blockWritten += written;
|
||||||
|
int remainingCapacity = block.remainingCapacity();
|
||||||
|
if (written < len) {
|
||||||
|
// not everything was written — the block has run out
|
||||||
|
// of capacity
|
||||||
|
// Trigger an upload then process the remainder.
|
||||||
|
LOG.debug("writing more data than block has capacity -triggering upload");
|
||||||
|
uploadCurrentBlock();
|
||||||
|
} else {
|
||||||
|
if (remainingCapacity == 0) {
|
||||||
|
// the whole buffer is done, trigger an upload
|
||||||
|
uploadCurrentBlock();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return written;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Clear the active block.
|
||||||
|
*/
|
||||||
|
private void clearActiveBlock() {
|
||||||
|
if (activeBlock != null) {
|
||||||
|
LOG.debug("Clearing active block");
|
||||||
|
}
|
||||||
|
synchronized (this) {
|
||||||
|
activeBlock = null;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
private void removeTemporaryFiles() {
|
private synchronized OSSDataBlocks.DataBlock getActiveBlock() {
|
||||||
for (File file : blockFiles.values()) {
|
return activeBlock;
|
||||||
if (file != null && file.exists() && !file.delete()) {
|
|
||||||
LOG.warn("Failed to delete temporary file {}", file);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
private void removePartFiles() throws IOException {
|
private void uploadCurrentBlock()
|
||||||
for (ListenableFuture<PartETag> partETagFuture : partETagsFutures) {
|
throws IOException {
|
||||||
if (!partETagFuture.isDone()) {
|
if (uploadId == null) {
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
try {
|
|
||||||
File blockFile = blockFiles.get(partETagFuture.get().getPartNumber());
|
|
||||||
if (blockFile != null && blockFile.exists() && !blockFile.delete()) {
|
|
||||||
LOG.warn("Failed to delete temporary file {}", blockFile);
|
|
||||||
}
|
|
||||||
} catch (InterruptedException | ExecutionException e) {
|
|
||||||
throw new IOException(e);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
private void uploadCurrentPart() throws IOException {
|
|
||||||
blockStream.flush();
|
|
||||||
blockStream.close();
|
|
||||||
if (blockId == 0) {
|
|
||||||
uploadId = store.getUploadId(key);
|
uploadId = store.getUploadId(key);
|
||||||
}
|
}
|
||||||
|
|
||||||
blockId++;
|
|
||||||
blockFiles.put(blockId, blockFile);
|
|
||||||
|
|
||||||
File currentFile = blockFile;
|
|
||||||
int currentBlockId = blockId;
|
int currentBlockId = blockId;
|
||||||
ListenableFuture<PartETag> partETagFuture =
|
OSSDataBlocks.DataBlock dataBlock = getActiveBlock();
|
||||||
executorService.submit(() -> {
|
long size = dataBlock.dataSize();
|
||||||
PartETag partETag = store.uploadPart(currentFile, key, uploadId,
|
OSSDataBlocks.BlockUploadData uploadData = dataBlock.startUpload();
|
||||||
currentBlockId);
|
try {
|
||||||
return partETag;
|
ListenableFuture<PartETag> partETagFuture =
|
||||||
});
|
executorService.submit(() -> {
|
||||||
partETagsFutures.add(partETagFuture);
|
try {
|
||||||
removePartFiles();
|
PartETag partETag = store.uploadPart(uploadData, size, key,
|
||||||
blockFile = newBlockFile();
|
uploadId, currentBlockId);
|
||||||
blockStream = new BufferedOutputStream(new FileOutputStream(blockFile));
|
return partETag;
|
||||||
|
} finally {
|
||||||
|
cleanupWithLogger(LOG, uploadData, dataBlock);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
partETagsFutures.add(partETagFuture);
|
||||||
|
} finally {
|
||||||
|
blockWritten = 0;
|
||||||
|
clearActiveBlock();
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -27,6 +27,9 @@
|
|||||||
import java.util.concurrent.ExecutorService;
|
import java.util.concurrent.ExecutorService;
|
||||||
import java.util.concurrent.TimeUnit;
|
import java.util.concurrent.TimeUnit;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.VisibleForTesting;
|
||||||
|
import org.apache.hadoop.fs.aliyun.oss.statistics.BlockOutputStreamStatistics;
|
||||||
|
import org.apache.hadoop.fs.aliyun.oss.statistics.impl.OutputStreamStatistics;
|
||||||
import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.MoreExecutors;
|
import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.MoreExecutors;
|
||||||
import org.apache.commons.collections.CollectionUtils;
|
import org.apache.commons.collections.CollectionUtils;
|
||||||
import org.apache.commons.lang3.StringUtils;
|
import org.apache.commons.lang3.StringUtils;
|
||||||
@ -71,6 +74,9 @@ public class AliyunOSSFileSystem extends FileSystem {
|
|||||||
private String bucket;
|
private String bucket;
|
||||||
private String username;
|
private String username;
|
||||||
private Path workingDir;
|
private Path workingDir;
|
||||||
|
private OSSDataBlocks.BlockFactory blockFactory;
|
||||||
|
private BlockOutputStreamStatistics blockOutputStreamStatistics;
|
||||||
|
private int uploadPartSize;
|
||||||
private int blockOutputActiveBlocks;
|
private int blockOutputActiveBlocks;
|
||||||
private AliyunOSSFileSystemStore store;
|
private AliyunOSSFileSystemStore store;
|
||||||
private int maxKeys;
|
private int maxKeys;
|
||||||
@ -128,13 +134,13 @@ public FSDataOutputStream create(Path path, FsPermission permission,
|
|||||||
// this means the file is not found
|
// this means the file is not found
|
||||||
}
|
}
|
||||||
|
|
||||||
long uploadPartSize = AliyunOSSUtils.getMultipartSizeProperty(getConf(),
|
|
||||||
MULTIPART_UPLOAD_PART_SIZE_KEY, MULTIPART_UPLOAD_PART_SIZE_DEFAULT);
|
|
||||||
return new FSDataOutputStream(
|
return new FSDataOutputStream(
|
||||||
new AliyunOSSBlockOutputStream(getConf(),
|
new AliyunOSSBlockOutputStream(getConf(),
|
||||||
store,
|
store,
|
||||||
key,
|
key,
|
||||||
uploadPartSize,
|
uploadPartSize,
|
||||||
|
blockFactory,
|
||||||
|
blockOutputStreamStatistics,
|
||||||
new SemaphoredDelegatingExecutor(boundedThreadPool,
|
new SemaphoredDelegatingExecutor(boundedThreadPool,
|
||||||
blockOutputActiveBlocks, true)), statistics);
|
blockOutputActiveBlocks, true)), statistics);
|
||||||
}
|
}
|
||||||
@ -334,6 +340,7 @@ public String getCanonicalServiceName() {
|
|||||||
*/
|
*/
|
||||||
public void initialize(URI name, Configuration conf) throws IOException {
|
public void initialize(URI name, Configuration conf) throws IOException {
|
||||||
super.initialize(name, conf);
|
super.initialize(name, conf);
|
||||||
|
setConf(conf);
|
||||||
|
|
||||||
bucket = name.getHost();
|
bucket = name.getHost();
|
||||||
uri = java.net.URI.create(name.getScheme() + "://" + name.getAuthority());
|
uri = java.net.URI.create(name.getScheme() + "://" + name.getAuthority());
|
||||||
@ -345,6 +352,16 @@ public void initialize(URI name, Configuration conf) throws IOException {
|
|||||||
blockOutputActiveBlocks = intOption(conf,
|
blockOutputActiveBlocks = intOption(conf,
|
||||||
UPLOAD_ACTIVE_BLOCKS_KEY, UPLOAD_ACTIVE_BLOCKS_DEFAULT, 1);
|
UPLOAD_ACTIVE_BLOCKS_KEY, UPLOAD_ACTIVE_BLOCKS_DEFAULT, 1);
|
||||||
|
|
||||||
|
uploadPartSize = (int)AliyunOSSUtils.getMultipartSizeProperty(conf,
|
||||||
|
MULTIPART_UPLOAD_PART_SIZE_KEY, MULTIPART_UPLOAD_PART_SIZE_DEFAULT);
|
||||||
|
String uploadBuffer = conf.getTrimmed(FAST_UPLOAD_BUFFER,
|
||||||
|
DEFAULT_FAST_UPLOAD_BUFFER);
|
||||||
|
|
||||||
|
blockOutputStreamStatistics = new OutputStreamStatistics();
|
||||||
|
blockFactory = OSSDataBlocks.createFactory(this, uploadBuffer);
|
||||||
|
LOG.debug("Using OSSBlockOutputStream with buffer = {}; block={};" +
|
||||||
|
" queue limit={}",
|
||||||
|
uploadBuffer, uploadPartSize, blockOutputActiveBlocks);
|
||||||
store = new AliyunOSSFileSystemStore();
|
store = new AliyunOSSFileSystemStore();
|
||||||
store.initialize(name, conf, username, statistics);
|
store.initialize(name, conf, username, statistics);
|
||||||
maxKeys = conf.getInt(MAX_PAGING_KEYS_KEY, MAX_PAGING_KEYS_DEFAULT);
|
maxKeys = conf.getInt(MAX_PAGING_KEYS_KEY, MAX_PAGING_KEYS_DEFAULT);
|
||||||
@ -379,8 +396,6 @@ public void initialize(URI name, Configuration conf) throws IOException {
|
|||||||
this.boundedCopyThreadPool = BlockingThreadPoolExecutorService.newInstance(
|
this.boundedCopyThreadPool = BlockingThreadPoolExecutorService.newInstance(
|
||||||
maxCopyThreads, maxCopyTasks, 60L,
|
maxCopyThreads, maxCopyTasks, 60L,
|
||||||
TimeUnit.SECONDS, "oss-copy-unbounded");
|
TimeUnit.SECONDS, "oss-copy-unbounded");
|
||||||
|
|
||||||
setConf(conf);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -757,4 +772,14 @@ public void setWorkingDirectory(Path dir) {
|
|||||||
public AliyunOSSFileSystemStore getStore() {
|
public AliyunOSSFileSystemStore getStore() {
|
||||||
return store;
|
return store;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@VisibleForTesting
|
||||||
|
OSSDataBlocks.BlockFactory getBlockFactory() {
|
||||||
|
return blockFactory;
|
||||||
|
}
|
||||||
|
|
||||||
|
@VisibleForTesting
|
||||||
|
BlockOutputStreamStatistics getBlockOutputStreamStatistics() {
|
||||||
|
return blockOutputStreamStatistics;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
@ -422,6 +422,27 @@ public void uploadObject(String key, File file) throws IOException {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Upload an input stream as an OSS object, using single upload.
|
||||||
|
* @param key object key.
|
||||||
|
* @param in input stream to upload.
|
||||||
|
* @param size size of the input stream.
|
||||||
|
* @throws IOException if failed to upload object.
|
||||||
|
*/
|
||||||
|
public void uploadObject(String key, InputStream in, long size)
|
||||||
|
throws IOException {
|
||||||
|
ObjectMetadata meta = new ObjectMetadata();
|
||||||
|
meta.setContentLength(size);
|
||||||
|
|
||||||
|
if (StringUtils.isNotEmpty(serverSideEncryptionAlgorithm)) {
|
||||||
|
meta.setServerSideEncryption(serverSideEncryptionAlgorithm);
|
||||||
|
}
|
||||||
|
|
||||||
|
PutObjectResult result = ossClient.putObject(bucketName, key, in, meta);
|
||||||
|
LOG.debug(result.getETag());
|
||||||
|
statistics.incrementWriteOps(1);
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* list objects.
|
* list objects.
|
||||||
*
|
*
|
||||||
@ -652,44 +673,58 @@ private boolean continueListStatus() {
|
|||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public PartETag uploadPart(OSSDataBlocks.BlockUploadData partData,
|
||||||
|
long size, String key, String uploadId, int idx) throws IOException {
|
||||||
|
if (partData.hasFile()) {
|
||||||
|
return uploadPart(partData.getFile(), key, uploadId, idx);
|
||||||
|
} else {
|
||||||
|
return uploadPart(partData.getUploadStream(), size, key, uploadId, idx);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
public PartETag uploadPart(File file, String key, String uploadId, int idx)
|
public PartETag uploadPart(File file, String key, String uploadId, int idx)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
InputStream instream = null;
|
InputStream in = new FileInputStream(file);
|
||||||
|
try {
|
||||||
|
return uploadPart(in, file.length(), key, uploadId, idx);
|
||||||
|
} finally {
|
||||||
|
in.close();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public PartETag uploadPart(InputStream in, long size, String key,
|
||||||
|
String uploadId, int idx) throws IOException {
|
||||||
Exception caught = null;
|
Exception caught = null;
|
||||||
int tries = 3;
|
int tries = 3;
|
||||||
while (tries > 0) {
|
while (tries > 0) {
|
||||||
try {
|
try {
|
||||||
instream = new FileInputStream(file);
|
|
||||||
UploadPartRequest uploadRequest = new UploadPartRequest();
|
UploadPartRequest uploadRequest = new UploadPartRequest();
|
||||||
uploadRequest.setBucketName(bucketName);
|
uploadRequest.setBucketName(bucketName);
|
||||||
uploadRequest.setKey(key);
|
uploadRequest.setKey(key);
|
||||||
uploadRequest.setUploadId(uploadId);
|
uploadRequest.setUploadId(uploadId);
|
||||||
uploadRequest.setInputStream(instream);
|
uploadRequest.setInputStream(in);
|
||||||
uploadRequest.setPartSize(file.length());
|
uploadRequest.setPartSize(size);
|
||||||
uploadRequest.setPartNumber(idx);
|
uploadRequest.setPartNumber(idx);
|
||||||
UploadPartResult uploadResult = ossClient.uploadPart(uploadRequest);
|
UploadPartResult uploadResult = ossClient.uploadPart(uploadRequest);
|
||||||
statistics.incrementWriteOps(1);
|
statistics.incrementWriteOps(1);
|
||||||
return uploadResult.getPartETag();
|
return uploadResult.getPartETag();
|
||||||
} catch (Exception e) {
|
} catch (Exception e) {
|
||||||
LOG.debug("Failed to upload "+ file.getPath() +", " +
|
LOG.debug("Failed to upload " + key + ", part " + idx +
|
||||||
"try again.", e);
|
"try again.", e);
|
||||||
caught = e;
|
caught = e;
|
||||||
} finally {
|
|
||||||
if (instream != null) {
|
|
||||||
instream.close();
|
|
||||||
instream = null;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
tries--;
|
tries--;
|
||||||
}
|
}
|
||||||
|
|
||||||
assert (caught != null);
|
assert (caught != null);
|
||||||
throw new IOException("Failed to upload " + file.getPath() +
|
throw new IOException("Failed to upload " + key + ", part " + idx +
|
||||||
" for 3 times.", caught);
|
" for 3 times.", caught);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Initiate multipart upload.
|
* Initiate multipart upload.
|
||||||
|
* @param key object key.
|
||||||
|
* @return upload id.
|
||||||
*/
|
*/
|
||||||
public String getUploadId(String key) {
|
public String getUploadId(String key) {
|
||||||
InitiateMultipartUploadRequest initiateMultipartUploadRequest =
|
InitiateMultipartUploadRequest initiateMultipartUploadRequest =
|
||||||
@ -701,6 +736,10 @@ public String getUploadId(String key) {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Complete the specific multipart upload.
|
* Complete the specific multipart upload.
|
||||||
|
* @param key object key.
|
||||||
|
* @param uploadId upload id of this multipart upload.
|
||||||
|
* @param partETags part etags need to be completed.
|
||||||
|
* @return CompleteMultipartUploadResult.
|
||||||
*/
|
*/
|
||||||
public CompleteMultipartUploadResult completeMultipartUpload(String key,
|
public CompleteMultipartUploadResult completeMultipartUpload(String key,
|
||||||
String uploadId, List<PartETag> partETags) {
|
String uploadId, List<PartETag> partETags) {
|
||||||
@ -713,6 +752,8 @@ public CompleteMultipartUploadResult completeMultipartUpload(String key,
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Abort the specific multipart upload.
|
* Abort the specific multipart upload.
|
||||||
|
* @param key object key.
|
||||||
|
* @param uploadId upload id of this multipart upload.
|
||||||
*/
|
*/
|
||||||
public void abortMultipartUpload(String key, String uploadId) {
|
public void abortMultipartUpload(String key, String uploadId) {
|
||||||
AbortMultipartUploadRequest request = new AbortMultipartUploadRequest(
|
AbortMultipartUploadRequest request = new AbortMultipartUploadRequest(
|
||||||
|
@ -134,6 +134,59 @@ private Constants() {
|
|||||||
// Comma separated list of directories
|
// Comma separated list of directories
|
||||||
public static final String BUFFER_DIR_KEY = "fs.oss.buffer.dir";
|
public static final String BUFFER_DIR_KEY = "fs.oss.buffer.dir";
|
||||||
|
|
||||||
|
/**
|
||||||
|
* What buffer to use.
|
||||||
|
* Default is {@link #FAST_UPLOAD_BUFFER_DISK}
|
||||||
|
* Value: {@value}
|
||||||
|
*/
|
||||||
|
public static final String FAST_UPLOAD_BUFFER =
|
||||||
|
"fs.oss.fast.upload.buffer";
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Buffer blocks to disk: {@value}.
|
||||||
|
* Capacity is limited to available disk space.
|
||||||
|
*/
|
||||||
|
public static final String FAST_UPLOAD_BUFFER_DISK = "disk";
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Use an in-memory array. Fast but will run of heap rapidly: {@value}.
|
||||||
|
*/
|
||||||
|
public static final String FAST_UPLOAD_BUFFER_ARRAY = "array";
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Use a byte buffer. May be more memory efficient than the
|
||||||
|
* {@link #FAST_UPLOAD_BUFFER_ARRAY}: {@value}.
|
||||||
|
*/
|
||||||
|
public static final String FAST_UPLOAD_BYTEBUFFER = "bytebuffer";
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Use an in-memory array and fallback to disk if
|
||||||
|
* used memory exceed the quota.
|
||||||
|
*/
|
||||||
|
public static final String FAST_UPLOAD_BUFFER_ARRAY_DISK = "array_disk";
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Use a byte buffer and fallback to disk if
|
||||||
|
* used memory exceed the quota.
|
||||||
|
*/
|
||||||
|
public static final String FAST_UPLOAD_BYTEBUFFER_DISK = "bytebuffer_disk";
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Memory limit of {@link #FAST_UPLOAD_BUFFER_ARRAY_DISK} or
|
||||||
|
* {@link #FAST_UPLOAD_BYTEBUFFER_DISK}.
|
||||||
|
*/
|
||||||
|
public static final String FAST_UPLOAD_BUFFER_MEMORY_LIMIT =
|
||||||
|
"fs.oss.fast.upload.memory.limit";
|
||||||
|
|
||||||
|
public static final long FAST_UPLOAD_BUFFER_MEMORY_LIMIT_DEFAULT =
|
||||||
|
1024 * 1024 * 1024; // 1GB
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Default buffer option: {@value}.
|
||||||
|
*/
|
||||||
|
public static final String DEFAULT_FAST_UPLOAD_BUFFER =
|
||||||
|
FAST_UPLOAD_BUFFER_DISK;
|
||||||
|
|
||||||
// private | public-read | public-read-write
|
// private | public-read | public-read-write
|
||||||
public static final String CANNED_ACL_KEY = "fs.oss.acl.default";
|
public static final String CANNED_ACL_KEY = "fs.oss.acl.default";
|
||||||
public static final String CANNED_ACL_DEFAULT = "";
|
public static final String CANNED_ACL_DEFAULT = "";
|
||||||
|
File diff suppressed because it is too large
Load Diff
@ -0,0 +1,72 @@
|
|||||||
|
/*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.fs.aliyun.oss.statistics;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Block output stream statistics.
|
||||||
|
*/
|
||||||
|
@InterfaceStability.Unstable
|
||||||
|
public interface BlockOutputStreamStatistics {
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A block has been allocated.
|
||||||
|
*/
|
||||||
|
void blockAllocated();
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A block has been released.
|
||||||
|
*/
|
||||||
|
void blockReleased();
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A disk block has been allocated.
|
||||||
|
*/
|
||||||
|
void diskBlockAllocated();
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A disk block has been released.
|
||||||
|
*/
|
||||||
|
void diskBlockReleased();
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Memory bytes has been allocated.
|
||||||
|
* @param size allocated size.
|
||||||
|
*/
|
||||||
|
void bytesAllocated(long size);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Memory bytes has been released.
|
||||||
|
* @param size released size.
|
||||||
|
*/
|
||||||
|
void bytesReleased(long size);
|
||||||
|
|
||||||
|
int getBlocksAllocated();
|
||||||
|
|
||||||
|
int getBlocksReleased();
|
||||||
|
|
||||||
|
int getDiskBlocksAllocated();
|
||||||
|
|
||||||
|
int getDiskBlocksReleased();
|
||||||
|
|
||||||
|
long getBytesAllocated();
|
||||||
|
|
||||||
|
long getBytesReleased();
|
||||||
|
}
|
@ -0,0 +1,98 @@
|
|||||||
|
/*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.fs.aliyun.oss.statistics.impl;
|
||||||
|
|
||||||
|
import org.apache.hadoop.fs.aliyun.oss.statistics.BlockOutputStreamStatistics;
|
||||||
|
|
||||||
|
import java.util.concurrent.atomic.AtomicInteger;
|
||||||
|
import java.util.concurrent.atomic.AtomicLong;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Implementation of {@link BlockOutputStreamStatistics}.
|
||||||
|
*/
|
||||||
|
public class OutputStreamStatistics implements BlockOutputStreamStatistics {
|
||||||
|
private final AtomicInteger blocksAllocated = new AtomicInteger(0);
|
||||||
|
private final AtomicInteger blocksReleased = new AtomicInteger(0);
|
||||||
|
|
||||||
|
private final AtomicInteger diskBlocksAllocated = new AtomicInteger(0);
|
||||||
|
private final AtomicInteger diskBlocksReleased = new AtomicInteger(0);
|
||||||
|
|
||||||
|
private final AtomicLong bytesAllocated = new AtomicLong(0);
|
||||||
|
private final AtomicLong bytesReleased = new AtomicLong(0);
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void blockAllocated() {
|
||||||
|
blocksAllocated.incrementAndGet();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void blockReleased() {
|
||||||
|
blocksReleased.incrementAndGet();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void diskBlockAllocated() {
|
||||||
|
diskBlocksAllocated.incrementAndGet();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void diskBlockReleased() {
|
||||||
|
diskBlocksReleased.incrementAndGet();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int getBlocksAllocated() {
|
||||||
|
return blocksAllocated.get();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int getBlocksReleased() {
|
||||||
|
return blocksReleased.get();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int getDiskBlocksAllocated() {
|
||||||
|
return diskBlocksAllocated.get();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int getDiskBlocksReleased() {
|
||||||
|
return diskBlocksReleased.get();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void bytesAllocated(long size) {
|
||||||
|
bytesAllocated.getAndAdd(size);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void bytesReleased(long size) {
|
||||||
|
bytesReleased.getAndAdd(size);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public long getBytesAllocated() {
|
||||||
|
return bytesAllocated.get();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public long getBytesReleased() {
|
||||||
|
return bytesReleased.get();
|
||||||
|
}
|
||||||
|
}
|
@ -0,0 +1,29 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Statistics collection for the OSS connector: implementation.
|
||||||
|
* Not for use by anything outside the hadoop-aliyun source tree.
|
||||||
|
*/
|
||||||
|
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
@InterfaceStability.Unstable
|
||||||
|
package org.apache.hadoop.fs.aliyun.oss.statistics.impl;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
@ -0,0 +1,27 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Statistics collection for the OSS connector: interfaces.
|
||||||
|
*/
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
@InterfaceStability.Unstable
|
||||||
|
package org.apache.hadoop.fs.aliyun.oss.statistics;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
@ -164,7 +164,7 @@ please raise your issues with them.
|
|||||||
|
|
||||||
<property>
|
<property>
|
||||||
<name>fs.oss.attempts.maximum</name>
|
<name>fs.oss.attempts.maximum</name>
|
||||||
<value>20</value>
|
<value>10</value>
|
||||||
<description>How many times we should retry commands on transient errors.</description>
|
<description>How many times we should retry commands on transient errors.</description>
|
||||||
</property>
|
</property>
|
||||||
|
|
||||||
@ -239,7 +239,7 @@ please raise your issues with them.
|
|||||||
|
|
||||||
<property>
|
<property>
|
||||||
<name>fs.oss.multipart.download.size</name>
|
<name>fs.oss.multipart.download.size</name>
|
||||||
<value>102400/value>
|
<value>524288/value>
|
||||||
<description>Size in bytes in each request from ALiyun OSS.</description>
|
<description>Size in bytes in each request from ALiyun OSS.</description>
|
||||||
</property>
|
</property>
|
||||||
|
|
||||||
@ -251,9 +251,53 @@ please raise your issues with them.
|
|||||||
</description>
|
</description>
|
||||||
</property>
|
</property>
|
||||||
|
|
||||||
|
<property>
|
||||||
|
<name>fs.oss.fast.upload.buffer</name>
|
||||||
|
<value>disk</value>
|
||||||
|
<description>
|
||||||
|
The buffering mechanism to use.
|
||||||
|
Values: disk, array, bytebuffer, array_disk, bytebuffer_disk.
|
||||||
|
|
||||||
|
"disk" will use the directories listed in fs.oss.buffer.dir as
|
||||||
|
the location(s) to save data prior to being uploaded.
|
||||||
|
|
||||||
|
"array" uses arrays in the JVM heap
|
||||||
|
|
||||||
|
"bytebuffer" uses off-heap memory within the JVM.
|
||||||
|
|
||||||
|
Both "array" and "bytebuffer" will consume memory in a single stream up to the number
|
||||||
|
of blocks set by:
|
||||||
|
|
||||||
|
fs.oss.multipart.upload.size * fs.oss.upload.active.blocks.
|
||||||
|
|
||||||
|
If using either of these mechanisms, keep this value low
|
||||||
|
|
||||||
|
The total number of threads performing work across all threads is set by
|
||||||
|
fs.oss.multipart.download.threads(Currently fast upload shares the same thread tool with download.
|
||||||
|
The thread pool size is specified in "fs.oss.multipart.download.threads"),
|
||||||
|
with fs.oss.max.total.tasks values setting the number of queued work items.
|
||||||
|
|
||||||
|
"array_disk" and "bytebuffer_disk" support fallback to disk.
|
||||||
|
</description>
|
||||||
|
</property>
|
||||||
|
|
||||||
|
<property>
|
||||||
|
<name>fs.oss.fast.upload.memory.limit</name>
|
||||||
|
<value>1073741824</value>
|
||||||
|
<description>
|
||||||
|
Memory limit of "array_disk" and "bytebuffer_disk" upload buffers.
|
||||||
|
Will fallback to disk buffers if used memory reaches the limit.
|
||||||
|
</description>
|
||||||
|
</property>
|
||||||
|
|
||||||
<property>
|
<property>
|
||||||
<name>fs.oss.buffer.dir</name>
|
<name>fs.oss.buffer.dir</name>
|
||||||
<description>Comma separated list of directories to buffer OSS data before uploading to Aliyun OSS</description>
|
<value>${env.LOCAL_DIRS:-${hadoop.tmp.dir}}/oss</value>
|
||||||
|
<description>Comma separated list of directories to buffer
|
||||||
|
OSS data before uploading to Aliyun OSS.
|
||||||
|
Yarn container path will be used as default value on yarn applications,
|
||||||
|
otherwise fall back to hadoop.tmp.dir
|
||||||
|
</description>
|
||||||
</property>
|
</property>
|
||||||
|
|
||||||
<property>
|
<property>
|
||||||
|
@ -22,6 +22,8 @@
|
|||||||
import org.apache.hadoop.fs.FileStatus;
|
import org.apache.hadoop.fs.FileStatus;
|
||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
|
import org.apache.hadoop.fs.aliyun.oss.OSSDataBlocks.ByteBufferBlockFactory;
|
||||||
|
import org.apache.hadoop.fs.aliyun.oss.statistics.BlockOutputStreamStatistics;
|
||||||
import org.apache.hadoop.fs.contract.ContractTestUtils;
|
import org.apache.hadoop.fs.contract.ContractTestUtils;
|
||||||
import org.apache.hadoop.test.GenericTestUtils;
|
import org.apache.hadoop.test.GenericTestUtils;
|
||||||
import org.junit.After;
|
import org.junit.After;
|
||||||
@ -37,12 +39,19 @@
|
|||||||
import java.util.LinkedHashSet;
|
import java.util.LinkedHashSet;
|
||||||
|
|
||||||
import static org.apache.hadoop.fs.aliyun.oss.Constants.BUFFER_DIR_KEY;
|
import static org.apache.hadoop.fs.aliyun.oss.Constants.BUFFER_DIR_KEY;
|
||||||
|
import static org.apache.hadoop.fs.aliyun.oss.Constants.FAST_UPLOAD_BUFFER;
|
||||||
|
import static org.apache.hadoop.fs.aliyun.oss.Constants.FAST_UPLOAD_BUFFER_ARRAY_DISK;
|
||||||
|
import static org.apache.hadoop.fs.aliyun.oss.Constants.FAST_UPLOAD_BUFFER_DISK;
|
||||||
|
import static org.apache.hadoop.fs.aliyun.oss.Constants.FAST_UPLOAD_BUFFER_MEMORY_LIMIT;
|
||||||
|
import static org.apache.hadoop.fs.aliyun.oss.Constants.FAST_UPLOAD_BYTEBUFFER;
|
||||||
|
import static org.apache.hadoop.fs.aliyun.oss.Constants.FAST_UPLOAD_BYTEBUFFER_DISK;
|
||||||
import static org.apache.hadoop.fs.aliyun.oss.Constants.MULTIPART_UPLOAD_PART_SIZE_DEFAULT;
|
import static org.apache.hadoop.fs.aliyun.oss.Constants.MULTIPART_UPLOAD_PART_SIZE_DEFAULT;
|
||||||
import static org.apache.hadoop.fs.aliyun.oss.Constants.MULTIPART_UPLOAD_PART_SIZE_KEY;
|
import static org.apache.hadoop.fs.aliyun.oss.Constants.MULTIPART_UPLOAD_PART_SIZE_KEY;
|
||||||
import static org.apache.hadoop.fs.contract.ContractTestUtils.IO_CHUNK_BUFFER_SIZE;
|
import static org.apache.hadoop.fs.contract.ContractTestUtils.IO_CHUNK_BUFFER_SIZE;
|
||||||
import static org.junit.Assert.assertEquals;
|
import static org.junit.Assert.assertEquals;
|
||||||
import static org.junit.Assert.assertFalse;
|
import static org.junit.Assert.assertFalse;
|
||||||
import static org.junit.Assert.assertNotEquals;
|
import static org.junit.Assert.assertNotEquals;
|
||||||
|
import static org.junit.Assert.assertNotNull;
|
||||||
import static org.junit.Assert.assertTrue;
|
import static org.junit.Assert.assertTrue;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -54,6 +63,7 @@ public class TestAliyunOSSBlockOutputStream {
|
|||||||
private static final int PART_SIZE = 1024 * 1024;
|
private static final int PART_SIZE = 1024 * 1024;
|
||||||
private static String testRootPath =
|
private static String testRootPath =
|
||||||
AliyunOSSTestUtils.generateUniqueTestPath();
|
AliyunOSSTestUtils.generateUniqueTestPath();
|
||||||
|
private static final long MEMORY_LIMIT = 10 * 1024 * 1024;
|
||||||
|
|
||||||
@Rule
|
@Rule
|
||||||
public Timeout testTimeout = new Timeout(30 * 60 * 1000);
|
public Timeout testTimeout = new Timeout(30 * 60 * 1000);
|
||||||
@ -65,6 +75,7 @@ public void setUp() throws Exception {
|
|||||||
conf.setInt(IO_CHUNK_BUFFER_SIZE,
|
conf.setInt(IO_CHUNK_BUFFER_SIZE,
|
||||||
conf.getInt(MULTIPART_UPLOAD_PART_SIZE_KEY, 0));
|
conf.getInt(MULTIPART_UPLOAD_PART_SIZE_KEY, 0));
|
||||||
conf.setInt(Constants.UPLOAD_ACTIVE_BLOCKS_KEY, 20);
|
conf.setInt(Constants.UPLOAD_ACTIVE_BLOCKS_KEY, 20);
|
||||||
|
conf.setLong(FAST_UPLOAD_BUFFER_MEMORY_LIMIT, MEMORY_LIMIT);
|
||||||
fs = AliyunOSSTestUtils.createTestFileSystem(conf);
|
fs = AliyunOSSTestUtils.createTestFileSystem(conf);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -82,7 +93,7 @@ private Path getTestPath() {
|
|||||||
@Test
|
@Test
|
||||||
public void testZeroByteUpload() throws IOException {
|
public void testZeroByteUpload() throws IOException {
|
||||||
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), 0);
|
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), 0);
|
||||||
bufferDirShouldEmpty();
|
bufferShouldReleased(true);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
@ -106,20 +117,21 @@ public void testRegularUpload() throws IOException {
|
|||||||
assertEquals(size - 1, statistics.getBytesRead());
|
assertEquals(size - 1, statistics.getBytesRead());
|
||||||
assertEquals(3, statistics.getWriteOps());
|
assertEquals(3, statistics.getWriteOps());
|
||||||
assertEquals(size - 1, statistics.getBytesWritten());
|
assertEquals(size - 1, statistics.getBytesWritten());
|
||||||
|
bufferShouldReleased();
|
||||||
|
|
||||||
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size);
|
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size);
|
||||||
assertEquals(14, statistics.getReadOps());
|
assertEquals(14, statistics.getReadOps());
|
||||||
assertEquals(2 * size - 1, statistics.getBytesRead());
|
assertEquals(2 * size - 1, statistics.getBytesRead());
|
||||||
assertEquals(6, statistics.getWriteOps());
|
assertEquals(6, statistics.getWriteOps());
|
||||||
assertEquals(2 * size - 1, statistics.getBytesWritten());
|
assertEquals(2 * size - 1, statistics.getBytesWritten());
|
||||||
|
bufferShouldReleased();
|
||||||
|
|
||||||
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size + 1);
|
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size + 1);
|
||||||
|
|
||||||
assertEquals(22, statistics.getReadOps());
|
assertEquals(22, statistics.getReadOps());
|
||||||
assertEquals(3 * size, statistics.getBytesRead());
|
assertEquals(3 * size, statistics.getBytesRead());
|
||||||
assertEquals(10, statistics.getWriteOps());
|
assertEquals(10, statistics.getWriteOps());
|
||||||
assertEquals(3 * size, statistics.getBytesWritten());
|
assertEquals(3 * size, statistics.getBytesWritten());
|
||||||
bufferDirShouldEmpty();
|
bufferShouldReleased();
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
@ -133,19 +145,21 @@ public void testMultiPartUpload() throws IOException {
|
|||||||
assertEquals(size - 1, statistics.getBytesRead());
|
assertEquals(size - 1, statistics.getBytesRead());
|
||||||
assertEquals(8, statistics.getWriteOps());
|
assertEquals(8, statistics.getWriteOps());
|
||||||
assertEquals(size - 1, statistics.getBytesWritten());
|
assertEquals(size - 1, statistics.getBytesWritten());
|
||||||
|
bufferShouldReleased();
|
||||||
|
|
||||||
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size);
|
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size);
|
||||||
assertEquals(34, statistics.getReadOps());
|
assertEquals(34, statistics.getReadOps());
|
||||||
assertEquals(2 * size - 1, statistics.getBytesRead());
|
assertEquals(2 * size - 1, statistics.getBytesRead());
|
||||||
assertEquals(16, statistics.getWriteOps());
|
assertEquals(16, statistics.getWriteOps());
|
||||||
assertEquals(2 * size - 1, statistics.getBytesWritten());
|
assertEquals(2 * size - 1, statistics.getBytesWritten());
|
||||||
|
bufferShouldReleased();
|
||||||
|
|
||||||
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size + 1);
|
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size + 1);
|
||||||
assertEquals(52, statistics.getReadOps());
|
assertEquals(52, statistics.getReadOps());
|
||||||
assertEquals(3 * size, statistics.getBytesRead());
|
assertEquals(3 * size, statistics.getBytesRead());
|
||||||
assertEquals(25, statistics.getWriteOps());
|
assertEquals(25, statistics.getWriteOps());
|
||||||
assertEquals(3 * size, statistics.getBytesWritten());
|
assertEquals(3 * size, statistics.getBytesWritten());
|
||||||
bufferDirShouldEmpty();
|
bufferShouldReleased();
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
@ -159,16 +173,18 @@ public void testMultiPartUploadConcurrent() throws IOException {
|
|||||||
assertEquals(size, statistics.getBytesRead());
|
assertEquals(size, statistics.getBytesRead());
|
||||||
assertEquals(52, statistics.getWriteOps());
|
assertEquals(52, statistics.getWriteOps());
|
||||||
assertEquals(size, statistics.getBytesWritten());
|
assertEquals(size, statistics.getBytesWritten());
|
||||||
bufferDirShouldEmpty();
|
bufferShouldReleased();
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testHugeUpload() throws IOException {
|
public void testHugeUpload() throws IOException {
|
||||||
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), PART_SIZE - 1);
|
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), PART_SIZE - 1);
|
||||||
|
bufferShouldReleased();
|
||||||
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), PART_SIZE);
|
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), PART_SIZE);
|
||||||
|
bufferShouldReleased();
|
||||||
ContractTestUtils.createAndVerifyFile(fs, getTestPath(),
|
ContractTestUtils.createAndVerifyFile(fs, getTestPath(),
|
||||||
MULTIPART_UPLOAD_PART_SIZE_DEFAULT + 1);
|
MULTIPART_UPLOAD_PART_SIZE_DEFAULT + 1);
|
||||||
bufferDirShouldEmpty();
|
bufferShouldReleased();
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
@ -199,15 +215,43 @@ public void testMultiPartUploadLimit() throws IOException {
|
|||||||
public void testSmallUpload() throws IOException {
|
public void testSmallUpload() throws IOException {
|
||||||
long size = fs.getConf().getInt(MULTIPART_UPLOAD_PART_SIZE_KEY, 1024);
|
long size = fs.getConf().getInt(MULTIPART_UPLOAD_PART_SIZE_KEY, 1024);
|
||||||
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size - 1);
|
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size - 1);
|
||||||
bufferDirShouldEmpty();
|
bufferShouldReleased();
|
||||||
}
|
}
|
||||||
|
|
||||||
private void bufferDirShouldEmpty() throws IOException {
|
private void bufferShouldReleased() throws IOException {
|
||||||
Path bufferPath = new Path(fs.getConf().get(BUFFER_DIR_KEY));
|
bufferShouldReleased(false);
|
||||||
FileStatus[] files = bufferPath.getFileSystem(
|
}
|
||||||
fs.getConf()).listStatus(bufferPath);
|
|
||||||
// Temporary file should be deleted
|
private void bufferShouldReleased(boolean zeroSizeFile) throws IOException {
|
||||||
assertEquals(0, files.length);
|
String bufferDir = fs.getConf().get(BUFFER_DIR_KEY);
|
||||||
|
String bufferType = fs.getConf().get(FAST_UPLOAD_BUFFER);
|
||||||
|
if (bufferType.equals(FAST_UPLOAD_BUFFER_DISK)) {
|
||||||
|
assertNotNull(bufferDir);
|
||||||
|
Path bufferPath = new Path(fs.getConf().get(BUFFER_DIR_KEY));
|
||||||
|
FileStatus[] files = bufferPath.getFileSystem(
|
||||||
|
fs.getConf()).listStatus(bufferPath);
|
||||||
|
// Temporary file should be deleted
|
||||||
|
assertEquals(0, files.length);
|
||||||
|
} else {
|
||||||
|
if (bufferType.equals(FAST_UPLOAD_BYTEBUFFER)) {
|
||||||
|
OSSDataBlocks.ByteBufferBlockFactory
|
||||||
|
blockFactory = (OSSDataBlocks.ByteBufferBlockFactory)
|
||||||
|
((AliyunOSSFileSystem)fs).getBlockFactory();
|
||||||
|
assertEquals("outstanding buffers in " + blockFactory,
|
||||||
|
0, blockFactory.getOutstandingBufferCount());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
BlockOutputStreamStatistics statistics =
|
||||||
|
((AliyunOSSFileSystem)fs).getBlockOutputStreamStatistics();
|
||||||
|
assertEquals(statistics.getBlocksAllocated(),
|
||||||
|
statistics.getBlocksReleased());
|
||||||
|
if (zeroSizeFile) {
|
||||||
|
assertEquals(statistics.getBlocksAllocated(), 0);
|
||||||
|
} else {
|
||||||
|
assertTrue(statistics.getBlocksAllocated() >= 1);
|
||||||
|
}
|
||||||
|
assertEquals(statistics.getBytesReleased(),
|
||||||
|
statistics.getBytesAllocated());
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
@ -249,4 +293,127 @@ public void testDirectoryAllocatorRR() throws Throwable {
|
|||||||
assertNotEquals("round robin not working",
|
assertNotEquals("round robin not working",
|
||||||
tmp1.getParent(), tmp2.getParent());
|
tmp1.getParent(), tmp2.getParent());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testByteBufferIO() throws IOException {
|
||||||
|
try (OSSDataBlocks.ByteBufferBlockFactory factory =
|
||||||
|
new OSSDataBlocks.ByteBufferBlockFactory((AliyunOSSFileSystem)fs)) {
|
||||||
|
int limit = 128;
|
||||||
|
OSSDataBlocks.ByteBufferBlockFactory.ByteBufferBlock block
|
||||||
|
= factory.create(1, limit, null);
|
||||||
|
assertEquals("outstanding buffers in " + factory,
|
||||||
|
1, factory.getOutstandingBufferCount());
|
||||||
|
|
||||||
|
byte[] buffer = ContractTestUtils.toAsciiByteArray("test data");
|
||||||
|
int bufferLen = buffer.length;
|
||||||
|
block.write(buffer, 0, bufferLen);
|
||||||
|
assertEquals(bufferLen, block.dataSize());
|
||||||
|
assertEquals("capacity in " + block,
|
||||||
|
limit - bufferLen, block.remainingCapacity());
|
||||||
|
assertTrue("hasCapacity(64) in " + block, block.hasCapacity(64));
|
||||||
|
assertTrue("No capacity in " + block,
|
||||||
|
block.hasCapacity(limit - bufferLen));
|
||||||
|
|
||||||
|
// now start the write
|
||||||
|
OSSDataBlocks.BlockUploadData blockUploadData = block.startUpload();
|
||||||
|
ByteBufferBlockFactory.ByteBufferBlock.ByteBufferInputStream
|
||||||
|
stream =
|
||||||
|
(ByteBufferBlockFactory.ByteBufferBlock.ByteBufferInputStream)
|
||||||
|
blockUploadData.getUploadStream();
|
||||||
|
assertTrue("Mark not supported in " + stream, stream.markSupported());
|
||||||
|
assertTrue("!hasRemaining() in " + stream, stream.hasRemaining());
|
||||||
|
|
||||||
|
int expected = bufferLen;
|
||||||
|
assertEquals("wrong available() in " + stream,
|
||||||
|
expected, stream.available());
|
||||||
|
|
||||||
|
assertEquals('t', stream.read());
|
||||||
|
stream.mark(limit);
|
||||||
|
expected--;
|
||||||
|
assertEquals("wrong available() in " + stream,
|
||||||
|
expected, stream.available());
|
||||||
|
|
||||||
|
// read into a byte array with an offset
|
||||||
|
int offset = 5;
|
||||||
|
byte[] in = new byte[limit];
|
||||||
|
assertEquals(2, stream.read(in, offset, 2));
|
||||||
|
assertEquals('e', in[offset]);
|
||||||
|
assertEquals('s', in[offset + 1]);
|
||||||
|
expected -= 2;
|
||||||
|
assertEquals("wrong available() in " + stream,
|
||||||
|
expected, stream.available());
|
||||||
|
|
||||||
|
// read to end
|
||||||
|
byte[] remainder = new byte[limit];
|
||||||
|
int c;
|
||||||
|
int index = 0;
|
||||||
|
while ((c = stream.read()) >= 0) {
|
||||||
|
remainder[index++] = (byte) c;
|
||||||
|
}
|
||||||
|
assertEquals(expected, index);
|
||||||
|
assertEquals('a', remainder[--index]);
|
||||||
|
|
||||||
|
assertEquals("wrong available() in " + stream,
|
||||||
|
0, stream.available());
|
||||||
|
assertTrue("hasRemaining() in " + stream, !stream.hasRemaining());
|
||||||
|
|
||||||
|
// go the mark point
|
||||||
|
stream.reset();
|
||||||
|
assertEquals('e', stream.read());
|
||||||
|
|
||||||
|
// when the stream is closed, the data should be returned
|
||||||
|
stream.close();
|
||||||
|
assertEquals("outstanding buffers in " + factory,
|
||||||
|
1, factory.getOutstandingBufferCount());
|
||||||
|
block.close();
|
||||||
|
assertEquals("outstanding buffers in " + factory,
|
||||||
|
0, factory.getOutstandingBufferCount());
|
||||||
|
stream.close();
|
||||||
|
assertEquals("outstanding buffers in " + factory,
|
||||||
|
0, factory.getOutstandingBufferCount());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testFastUploadArrayDisk() throws IOException {
|
||||||
|
testFastUploadFallback(FAST_UPLOAD_BUFFER_ARRAY_DISK);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testFastUploadByteBufferDisk() throws IOException {
|
||||||
|
testFastUploadFallback(FAST_UPLOAD_BYTEBUFFER_DISK);
|
||||||
|
}
|
||||||
|
|
||||||
|
private void testFastUploadFallback(String name) throws IOException {
|
||||||
|
Configuration conf = fs.getConf();
|
||||||
|
fs.close();
|
||||||
|
|
||||||
|
conf.set(FAST_UPLOAD_BUFFER, name);
|
||||||
|
|
||||||
|
fs = AliyunOSSTestUtils.createTestFileSystem(conf);
|
||||||
|
long size = 5 * MEMORY_LIMIT;
|
||||||
|
ContractTestUtils.createAndVerifyFile(fs, getTestPath(), size);
|
||||||
|
OSSDataBlocks.MemoryBlockFactory
|
||||||
|
blockFactory = ((OSSDataBlocks.MemoryAndDiskBlockFactory)
|
||||||
|
((AliyunOSSFileSystem)fs).getBlockFactory()).getMemoryFactory();
|
||||||
|
assertEquals(blockFactory.getMemoryUsed(), 0);
|
||||||
|
|
||||||
|
Path bufferPath = new Path(fs.getConf().get(BUFFER_DIR_KEY));
|
||||||
|
FileStatus[] files = bufferPath.getFileSystem(
|
||||||
|
fs.getConf()).listStatus(bufferPath);
|
||||||
|
// Temporary file should be deleted
|
||||||
|
assertEquals(0, files.length);
|
||||||
|
|
||||||
|
BlockOutputStreamStatistics statistics =
|
||||||
|
((AliyunOSSFileSystem)fs).getBlockOutputStreamStatistics();
|
||||||
|
assertEquals(statistics.getBlocksAllocated(),
|
||||||
|
statistics.getBlocksReleased());
|
||||||
|
assertTrue(statistics.getBlocksAllocated() > 1);
|
||||||
|
assertEquals(statistics.getBytesReleased(),
|
||||||
|
statistics.getBytesAllocated());
|
||||||
|
assertTrue(statistics.getBytesAllocated() >= MEMORY_LIMIT);
|
||||||
|
assertTrue(statistics.getDiskBlocksAllocated() > 0);
|
||||||
|
assertEquals(statistics.getDiskBlocksAllocated(),
|
||||||
|
statistics.getDiskBlocksReleased());
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user