HDFS-4906. HDFS Output streams should not accept writes after being closed. Contributed by Aaron T. Myers.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1494303 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
52fe4e6be9
commit
1c309f763b
@ -19,6 +19,7 @@
|
||||
package org.apache.hadoop.fs;
|
||||
|
||||
import java.io.*;
|
||||
import java.nio.channels.ClosedChannelException;
|
||||
import java.util.Arrays;
|
||||
|
||||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
@ -368,6 +369,7 @@ private static class ChecksumFSOutputSummer extends FSOutputSummer {
|
||||
private FSDataOutputStream datas;
|
||||
private FSDataOutputStream sums;
|
||||
private static final float CHKSUM_AS_FRACTION = 0.01f;
|
||||
private boolean isClosed = false;
|
||||
|
||||
public ChecksumFSOutputSummer(ChecksumFileSystem fs,
|
||||
Path file,
|
||||
@ -391,9 +393,13 @@ public ChecksumFSOutputSummer(ChecksumFileSystem fs,
|
||||
|
||||
@Override
|
||||
public void close() throws IOException {
|
||||
flushBuffer();
|
||||
sums.close();
|
||||
datas.close();
|
||||
try {
|
||||
flushBuffer();
|
||||
sums.close();
|
||||
datas.close();
|
||||
} finally {
|
||||
isClosed = true;
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
@ -402,6 +408,13 @@ protected void writeChunk(byte[] b, int offset, int len, byte[] checksum)
|
||||
datas.write(b, offset, len);
|
||||
sums.write(checksum);
|
||||
}
|
||||
|
||||
@Override
|
||||
protected void checkClosed() throws IOException {
|
||||
if (isClosed) {
|
||||
throw new ClosedChannelException();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
|
@ -20,6 +20,7 @@
|
||||
|
||||
import java.io.*;
|
||||
import java.net.URISyntaxException;
|
||||
import java.nio.channels.ClosedChannelException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.EnumSet;
|
||||
@ -325,6 +326,7 @@ private static class ChecksumFSOutputSummer extends FSOutputSummer {
|
||||
private FSDataOutputStream datas;
|
||||
private FSDataOutputStream sums;
|
||||
private static final float CHKSUM_AS_FRACTION = 0.01f;
|
||||
private boolean isClosed = false;
|
||||
|
||||
|
||||
public ChecksumFSOutputSummer(final ChecksumFs fs, final Path file,
|
||||
@ -356,9 +358,13 @@ public ChecksumFSOutputSummer(final ChecksumFs fs, final Path file,
|
||||
|
||||
@Override
|
||||
public void close() throws IOException {
|
||||
flushBuffer();
|
||||
sums.close();
|
||||
datas.close();
|
||||
try {
|
||||
flushBuffer();
|
||||
sums.close();
|
||||
datas.close();
|
||||
} finally {
|
||||
isClosed = true;
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
@ -367,6 +373,13 @@ protected void writeChunk(byte[] b, int offset, int len, byte[] checksum)
|
||||
datas.write(b, offset, len);
|
||||
sums.write(checksum);
|
||||
}
|
||||
|
||||
@Override
|
||||
protected void checkClosed() throws IOException {
|
||||
if (isClosed) {
|
||||
throw new ClosedChannelException();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
|
@ -54,6 +54,15 @@ protected FSOutputSummer(Checksum sum, int maxChunkSize, int checksumSize) {
|
||||
protected abstract void writeChunk(byte[] b, int offset, int len, byte[] checksum)
|
||||
throws IOException;
|
||||
|
||||
/**
|
||||
* Check if the implementing OutputStream is closed and should no longer
|
||||
* accept writes. Implementations should do nothing if this stream is not
|
||||
* closed, and should throw an {@link IOException} if it is closed.
|
||||
*
|
||||
* @throws IOException if this stream is already closed.
|
||||
*/
|
||||
protected abstract void checkClosed() throws IOException;
|
||||
|
||||
/** Write one byte */
|
||||
@Override
|
||||
public synchronized void write(int b) throws IOException {
|
||||
@ -84,7 +93,10 @@ public synchronized void write(int b) throws IOException {
|
||||
*/
|
||||
@Override
|
||||
public synchronized void write(byte b[], int off, int len)
|
||||
throws IOException {
|
||||
throws IOException {
|
||||
|
||||
checkClosed();
|
||||
|
||||
if (off < 0 || len < 0 || off > b.length - len) {
|
||||
throw new ArrayIndexOutOfBoundsException();
|
||||
}
|
||||
|
@ -582,6 +582,9 @@ Release 2.1.0-beta - UNRELEASED
|
||||
HDFS-4845. FSNamesystem.deleteInternal should acquire write-lock before
|
||||
changing the inode map. (Arpit Agarwal via szetszwo)
|
||||
|
||||
HDFS-4906. HDFS Output streams should not accept writes after being
|
||||
closed. (atm)
|
||||
|
||||
BREAKDOWN OF HDFS-347 SUBTASKS AND RELATED JIRAS
|
||||
|
||||
HDFS-4353. Encapsulate connections to peers in Peer and PeerServer classes.
|
||||
|
@ -30,6 +30,7 @@
|
||||
import java.net.InetSocketAddress;
|
||||
import java.net.Socket;
|
||||
import java.nio.BufferOverflowException;
|
||||
import java.nio.channels.ClosedChannelException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.EnumSet;
|
||||
@ -1299,10 +1300,10 @@ static Socket createSocketForPipeline(final DatanodeInfo first,
|
||||
return sock;
|
||||
}
|
||||
|
||||
private void isClosed() throws IOException {
|
||||
protected void checkClosed() throws IOException {
|
||||
if (closed) {
|
||||
IOException e = lastException;
|
||||
throw e != null ? e : new IOException("DFSOutputStream is closed");
|
||||
throw e != null ? e : new ClosedChannelException();
|
||||
}
|
||||
}
|
||||
|
||||
@ -1471,7 +1472,7 @@ private void waitAndQueueCurrentPacket() throws IOException {
|
||||
break;
|
||||
}
|
||||
}
|
||||
isClosed();
|
||||
checkClosed();
|
||||
queueCurrentPacket();
|
||||
}
|
||||
}
|
||||
@ -1481,7 +1482,7 @@ private void waitAndQueueCurrentPacket() throws IOException {
|
||||
protected synchronized void writeChunk(byte[] b, int offset, int len, byte[] checksum)
|
||||
throws IOException {
|
||||
dfsClient.checkOpen();
|
||||
isClosed();
|
||||
checkClosed();
|
||||
|
||||
int cklen = checksum.length;
|
||||
int bytesPerChecksum = this.checksum.getBytesPerChecksum();
|
||||
@ -1608,7 +1609,7 @@ public void hsync(EnumSet<SyncFlag> syncFlags) throws IOException {
|
||||
private void flushOrSync(boolean isSync, EnumSet<SyncFlag> syncFlags)
|
||||
throws IOException {
|
||||
dfsClient.checkOpen();
|
||||
isClosed();
|
||||
checkClosed();
|
||||
try {
|
||||
long toWaitFor;
|
||||
long lastBlockLength = -1L;
|
||||
@ -1695,7 +1696,7 @@ private void flushOrSync(boolean isSync, EnumSet<SyncFlag> syncFlags)
|
||||
// If we got an error here, it might be because some other thread called
|
||||
// close before our hflush completed. In that case, we should throw an
|
||||
// exception that the stream is closed.
|
||||
isClosed();
|
||||
checkClosed();
|
||||
// If we aren't closed but failed to sync, we should expose that to the
|
||||
// caller.
|
||||
throw ioe;
|
||||
@ -1740,7 +1741,7 @@ public synchronized int getNumCurrentReplicas() throws IOException {
|
||||
*/
|
||||
public synchronized int getCurrentBlockReplication() throws IOException {
|
||||
dfsClient.checkOpen();
|
||||
isClosed();
|
||||
checkClosed();
|
||||
if (streamer == null) {
|
||||
return blockReplication; // no pipeline, return repl factor of file
|
||||
}
|
||||
@ -1759,7 +1760,7 @@ private void flushInternal() throws IOException {
|
||||
long toWaitFor;
|
||||
synchronized (this) {
|
||||
dfsClient.checkOpen();
|
||||
isClosed();
|
||||
checkClosed();
|
||||
//
|
||||
// If there is data in the current buffer, send it across
|
||||
//
|
||||
@ -1776,7 +1777,7 @@ private void waitForAckedSeqno(long seqno) throws IOException {
|
||||
}
|
||||
synchronized (dataQueue) {
|
||||
while (!closed) {
|
||||
isClosed();
|
||||
checkClosed();
|
||||
if (lastAckedSeqno >= seqno) {
|
||||
break;
|
||||
}
|
||||
@ -1788,7 +1789,7 @@ private void waitForAckedSeqno(long seqno) throws IOException {
|
||||
}
|
||||
}
|
||||
}
|
||||
isClosed();
|
||||
checkClosed();
|
||||
}
|
||||
|
||||
private synchronized void start() {
|
||||
|
@ -0,0 +1,64 @@
|
||||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.junit.Assert.fail;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.OutputStream;
|
||||
import java.nio.channels.ClosedChannelException;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.junit.Test;
|
||||
|
||||
public class TestClose {
|
||||
|
||||
@Test
|
||||
public void testWriteAfterClose() throws IOException {
|
||||
Configuration conf = new Configuration();
|
||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf)
|
||||
.build();
|
||||
|
||||
try {
|
||||
final byte[] data = "foo".getBytes();
|
||||
|
||||
FileSystem fs = FileSystem.get(conf);
|
||||
OutputStream out = fs.create(new Path("/test"));
|
||||
|
||||
out.write(data);
|
||||
out.close();
|
||||
try {
|
||||
// Should fail.
|
||||
out.write(data);
|
||||
fail("Should not have been able to write more data after file is closed.");
|
||||
} catch (ClosedChannelException cce) {
|
||||
// We got the correct exception. Ignoring.
|
||||
}
|
||||
// Should succeed. Double closes are OK.
|
||||
out.close();
|
||||
} finally {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
}
|
@ -18,6 +18,7 @@
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.nio.channels.ClosedChannelException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.concurrent.CountDownLatch;
|
||||
import java.util.concurrent.TimeUnit;
|
||||
@ -169,12 +170,9 @@ public void run() {
|
||||
while (true) {
|
||||
try {
|
||||
stm.hflush();
|
||||
} catch (IOException ioe) {
|
||||
if (!ioe.toString().contains("DFSOutputStream is closed")) {
|
||||
throw ioe;
|
||||
} else {
|
||||
return;
|
||||
}
|
||||
} catch (ClosedChannelException ioe) {
|
||||
// Expected exception caught. Ignoring.
|
||||
return;
|
||||
}
|
||||
}
|
||||
} catch (Throwable t) {
|
||||
|
Loading…
Reference in New Issue
Block a user