hadoop/hadoop-tools/hadoop-azure/src
Mehakmeet Singh 8e5620cd9e
HADOOP-17195. ABFS: OutOfMemory error while uploading huge files (#3446)
Addresses the problem of processes running out of memory when
there are many ABFS output streams queuing data to upload,
especially when the network upload bandwidth is less than the rate
data is generated.

ABFS Output streams now buffer their blocks of data to
"disk", "bytebuffer" or "array", as set in
"fs.azure.data.blocks.buffer"

When buffering via disk, the location for temporary storage
is set in "fs.azure.buffer.dir"

For safe scaling: use "disk" (default); for performance, when
confident that upload bandwidth will never be a bottleneck,
experiment with the memory options.

The number of blocks a single stream can have queued for uploading
is set in "fs.azure.block.upload.active.blocks".
The default value is 20.

Contributed by Mehakmeet Singh.
2021-09-22 11:19:16 +01:00
..
config HADOOP-17404. ABFS: Small write - Merge append and flush 2021-01-22 10:48:04 +00:00
main HADOOP-17195. ABFS: OutOfMemory error while uploading huge files (#3446) 2021-09-22 11:19:16 +01:00
site HADOOP-17290. ABFS: Add Identifiers to Client Request Header (#2520) 2021-09-21 16:45:51 +01:00
test HADOOP-17195. ABFS: OutOfMemory error while uploading huge files (#3446) 2021-09-22 11:19:16 +01:00