HADOOP-11470. Remove some uses of obsolete guava APIs from the hadoop codebase (Sangjin Lee via Colin P. McCabe)

This commit is contained in:
Colin Patrick Mccabe 2015-01-08 16:09:44 -08:00
parent 7e2d9a3242
commit ae91b13a4b
3 changed files with 11 additions and 11 deletions

View File

@ -482,6 +482,9 @@ Release 2.7.0 - UNRELEASED
HADOOP-11032. Replace use of Guava's Stopwatch with Hadoop's StopWatch HADOOP-11032. Replace use of Guava's Stopwatch with Hadoop's StopWatch
(ozawa) (ozawa)
HADOOP-11470. Remove some uses of obsolete guava APIs from the hadoop
codebase. (Sangjin Lee via Colin P. McCabe)
OPTIMIZATIONS OPTIMIZATIONS
HDFS-7454. Reduce memory footprint for AclEntries in NameNode. HDFS-7454. Reduce memory footprint for AclEntries in NameNode.

View File

@ -29,6 +29,7 @@
import java.util.List; import java.util.List;
import java.util.concurrent.TimeUnit; import java.util.concurrent.TimeUnit;
import org.apache.commons.lang.math.LongRange;
import org.apache.commons.logging.Log; import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory; import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
@ -61,15 +62,13 @@
import org.apache.hadoop.ipc.Server; import org.apache.hadoop.ipc.Server;
import org.apache.hadoop.security.SecurityUtil; import org.apache.hadoop.security.SecurityUtil;
import org.apache.hadoop.security.UserGroupInformation; import org.apache.hadoop.security.UserGroupInformation;
import org.apache.hadoop.util.StopWatch;
import com.google.common.annotations.VisibleForTesting; import com.google.common.annotations.VisibleForTesting;
import com.google.common.base.Charsets; import com.google.common.base.Charsets;
import com.google.common.base.Preconditions; import com.google.common.base.Preconditions;
import com.google.common.collect.ImmutableList; import com.google.common.collect.ImmutableList;
import com.google.common.collect.Range;
import com.google.common.collect.Ranges;
import com.google.protobuf.TextFormat; import com.google.protobuf.TextFormat;
import org.apache.hadoop.util.StopWatch;
/** /**
* A JournalNode can manage journals for several clusters at once. * A JournalNode can manage journals for several clusters at once.
@ -793,8 +792,8 @@ public synchronized void acceptRecovery(RequestInfo reqInfo,
// Paranoid sanity check: if the new log is shorter than the log we // Paranoid sanity check: if the new log is shorter than the log we
// currently have, we should not end up discarding any transactions // currently have, we should not end up discarding any transactions
// which are already Committed. // which are already Committed.
if (txnRange(currentSegment).contains(committedTxnId.get()) && if (txnRange(currentSegment).containsLong(committedTxnId.get()) &&
!txnRange(segment).contains(committedTxnId.get())) { !txnRange(segment).containsLong(committedTxnId.get())) {
throw new AssertionError( throw new AssertionError(
"Cannot replace segment " + "Cannot replace segment " +
TextFormat.shortDebugString(currentSegment) + TextFormat.shortDebugString(currentSegment) +
@ -812,7 +811,7 @@ public synchronized void acceptRecovery(RequestInfo reqInfo,
// If we're shortening the log, update our highest txid // If we're shortening the log, update our highest txid
// used for lag metrics. // used for lag metrics.
if (txnRange(currentSegment).contains(highestWrittenTxId)) { if (txnRange(currentSegment).containsLong(highestWrittenTxId)) {
highestWrittenTxId = segment.getEndTxId(); highestWrittenTxId = segment.getEndTxId();
} }
} }
@ -856,10 +855,10 @@ public synchronized void acceptRecovery(RequestInfo reqInfo,
TextFormat.shortDebugString(newData)); TextFormat.shortDebugString(newData));
} }
private Range<Long> txnRange(SegmentStateProto seg) { private LongRange txnRange(SegmentStateProto seg) {
Preconditions.checkArgument(seg.hasEndTxId(), Preconditions.checkArgument(seg.hasEndTxId(),
"invalid segment: %s", seg); "invalid segment: %s", seg);
return Ranges.closed(seg.getStartTxId(), seg.getEndTxId()); return new LongRange(seg.getStartTxId(), seg.getEndTxId());
} }
/** /**

View File

@ -29,7 +29,6 @@
import static org.junit.Assert.fail; import static org.junit.Assert.fail;
import java.io.InputStream; import java.io.InputStream;
import java.io.PrintWriter;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FSDataInputStream;
@ -46,7 +45,6 @@
import org.junit.Test; import org.junit.Test;
import com.google.common.base.Supplier; import com.google.common.base.Supplier;
import com.google.common.io.NullOutputStream;
public class TestDataTransferKeepalive { public class TestDataTransferKeepalive {
final Configuration conf = new HdfsConfiguration(); final Configuration conf = new HdfsConfiguration();
@ -223,7 +221,7 @@ public void testManyClosedSocketsInCache() throws Exception {
stms[i] = fs.open(TEST_FILE); stms[i] = fs.open(TEST_FILE);
} }
for (InputStream stm : stms) { for (InputStream stm : stms) {
IOUtils.copyBytes(stm, new NullOutputStream(), 1024); IOUtils.copyBytes(stm, new IOUtils.NullOutputStream(), 1024);
} }
} finally { } finally {
IOUtils.cleanup(null, stms); IOUtils.cleanup(null, stms);