HADOOP-11470. Remove some uses of obsolete guava APIs from the hadoop codebase (Sangjin Lee via Colin P. McCabe)
This commit is contained in:
parent
7e2d9a3242
commit
ae91b13a4b
@ -482,6 +482,9 @@ Release 2.7.0 - UNRELEASED
|
|||||||
HADOOP-11032. Replace use of Guava's Stopwatch with Hadoop's StopWatch
|
HADOOP-11032. Replace use of Guava's Stopwatch with Hadoop's StopWatch
|
||||||
(ozawa)
|
(ozawa)
|
||||||
|
|
||||||
|
HADOOP-11470. Remove some uses of obsolete guava APIs from the hadoop
|
||||||
|
codebase. (Sangjin Lee via Colin P. McCabe)
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
HDFS-7454. Reduce memory footprint for AclEntries in NameNode.
|
HDFS-7454. Reduce memory footprint for AclEntries in NameNode.
|
||||||
|
@ -29,6 +29,7 @@
|
|||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.concurrent.TimeUnit;
|
import java.util.concurrent.TimeUnit;
|
||||||
|
|
||||||
|
import org.apache.commons.lang.math.LongRange;
|
||||||
import org.apache.commons.logging.Log;
|
import org.apache.commons.logging.Log;
|
||||||
import org.apache.commons.logging.LogFactory;
|
import org.apache.commons.logging.LogFactory;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
@ -61,15 +62,13 @@
|
|||||||
import org.apache.hadoop.ipc.Server;
|
import org.apache.hadoop.ipc.Server;
|
||||||
import org.apache.hadoop.security.SecurityUtil;
|
import org.apache.hadoop.security.SecurityUtil;
|
||||||
import org.apache.hadoop.security.UserGroupInformation;
|
import org.apache.hadoop.security.UserGroupInformation;
|
||||||
|
import org.apache.hadoop.util.StopWatch;
|
||||||
|
|
||||||
import com.google.common.annotations.VisibleForTesting;
|
import com.google.common.annotations.VisibleForTesting;
|
||||||
import com.google.common.base.Charsets;
|
import com.google.common.base.Charsets;
|
||||||
import com.google.common.base.Preconditions;
|
import com.google.common.base.Preconditions;
|
||||||
import com.google.common.collect.ImmutableList;
|
import com.google.common.collect.ImmutableList;
|
||||||
import com.google.common.collect.Range;
|
|
||||||
import com.google.common.collect.Ranges;
|
|
||||||
import com.google.protobuf.TextFormat;
|
import com.google.protobuf.TextFormat;
|
||||||
import org.apache.hadoop.util.StopWatch;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* A JournalNode can manage journals for several clusters at once.
|
* A JournalNode can manage journals for several clusters at once.
|
||||||
@ -793,8 +792,8 @@ public synchronized void acceptRecovery(RequestInfo reqInfo,
|
|||||||
// Paranoid sanity check: if the new log is shorter than the log we
|
// Paranoid sanity check: if the new log is shorter than the log we
|
||||||
// currently have, we should not end up discarding any transactions
|
// currently have, we should not end up discarding any transactions
|
||||||
// which are already Committed.
|
// which are already Committed.
|
||||||
if (txnRange(currentSegment).contains(committedTxnId.get()) &&
|
if (txnRange(currentSegment).containsLong(committedTxnId.get()) &&
|
||||||
!txnRange(segment).contains(committedTxnId.get())) {
|
!txnRange(segment).containsLong(committedTxnId.get())) {
|
||||||
throw new AssertionError(
|
throw new AssertionError(
|
||||||
"Cannot replace segment " +
|
"Cannot replace segment " +
|
||||||
TextFormat.shortDebugString(currentSegment) +
|
TextFormat.shortDebugString(currentSegment) +
|
||||||
@ -812,7 +811,7 @@ public synchronized void acceptRecovery(RequestInfo reqInfo,
|
|||||||
|
|
||||||
// If we're shortening the log, update our highest txid
|
// If we're shortening the log, update our highest txid
|
||||||
// used for lag metrics.
|
// used for lag metrics.
|
||||||
if (txnRange(currentSegment).contains(highestWrittenTxId)) {
|
if (txnRange(currentSegment).containsLong(highestWrittenTxId)) {
|
||||||
highestWrittenTxId = segment.getEndTxId();
|
highestWrittenTxId = segment.getEndTxId();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -856,10 +855,10 @@ public synchronized void acceptRecovery(RequestInfo reqInfo,
|
|||||||
TextFormat.shortDebugString(newData));
|
TextFormat.shortDebugString(newData));
|
||||||
}
|
}
|
||||||
|
|
||||||
private Range<Long> txnRange(SegmentStateProto seg) {
|
private LongRange txnRange(SegmentStateProto seg) {
|
||||||
Preconditions.checkArgument(seg.hasEndTxId(),
|
Preconditions.checkArgument(seg.hasEndTxId(),
|
||||||
"invalid segment: %s", seg);
|
"invalid segment: %s", seg);
|
||||||
return Ranges.closed(seg.getStartTxId(), seg.getEndTxId());
|
return new LongRange(seg.getStartTxId(), seg.getEndTxId());
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -29,7 +29,6 @@
|
|||||||
import static org.junit.Assert.fail;
|
import static org.junit.Assert.fail;
|
||||||
|
|
||||||
import java.io.InputStream;
|
import java.io.InputStream;
|
||||||
import java.io.PrintWriter;
|
|
||||||
|
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.fs.FSDataInputStream;
|
import org.apache.hadoop.fs.FSDataInputStream;
|
||||||
@ -46,7 +45,6 @@
|
|||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
|
|
||||||
import com.google.common.base.Supplier;
|
import com.google.common.base.Supplier;
|
||||||
import com.google.common.io.NullOutputStream;
|
|
||||||
|
|
||||||
public class TestDataTransferKeepalive {
|
public class TestDataTransferKeepalive {
|
||||||
final Configuration conf = new HdfsConfiguration();
|
final Configuration conf = new HdfsConfiguration();
|
||||||
@ -223,7 +221,7 @@ public void testManyClosedSocketsInCache() throws Exception {
|
|||||||
stms[i] = fs.open(TEST_FILE);
|
stms[i] = fs.open(TEST_FILE);
|
||||||
}
|
}
|
||||||
for (InputStream stm : stms) {
|
for (InputStream stm : stms) {
|
||||||
IOUtils.copyBytes(stm, new NullOutputStream(), 1024);
|
IOUtils.copyBytes(stm, new IOUtils.NullOutputStream(), 1024);
|
||||||
}
|
}
|
||||||
} finally {
|
} finally {
|
||||||
IOUtils.cleanup(null, stms);
|
IOUtils.cleanup(null, stms);
|
||||||
|
Loading…
Reference in New Issue
Block a user