HDFS-7449. Add metrics to NFS gateway. Contributed by Brandon Li

This commit is contained in:
Brandon Li 2014-12-11 15:40:45 -08:00
parent 0bcea111e5
commit f6f2a3f1c7
10 changed files with 297 additions and 15 deletions

View File

@ -70,4 +70,7 @@ public class NfsConfigKeys {
public static final int NFS_HTTPS_PORT_DEFAULT = 50579;
public static final String NFS_HTTPS_ADDRESS_KEY = "nfs.https.address";
public static final String NFS_HTTPS_ADDRESS_DEFAULT = "0.0.0.0:" + NFS_HTTPS_PORT_DEFAULT;
public static final String NFS_METRICS_PERCENTILES_INTERVALS_KEY = "nfs.metrics.percentiles.intervals";
public static final String NFS_METRICS_PERCENTILES_INTERVALS_DEFAULT = "";
}

View File

@ -42,7 +42,8 @@ public Nfs3(NfsConfiguration conf) throws IOException {
public Nfs3(NfsConfiguration conf, DatagramSocket registrationSocket,
boolean allowInsecurePorts) throws IOException {
super(new RpcProgramNfs3(conf, registrationSocket, allowInsecurePorts), conf);
super(RpcProgramNfs3.createRpcProgramNfs3(conf, registrationSocket,
allowInsecurePorts), conf);
mountd = new Mountd(conf, registrationSocket, allowInsecurePorts);
}

View File

@ -0,0 +1,220 @@
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.hdfs.nfs.nfs3;
import static org.apache.hadoop.metrics2.impl.MsInfo.SessionId;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hdfs.DFSConfigKeys;
import org.apache.hadoop.hdfs.nfs.conf.NfsConfigKeys;
import org.apache.hadoop.metrics2.MetricsSystem;
import org.apache.hadoop.metrics2.annotation.Metric;
import org.apache.hadoop.metrics2.annotation.Metrics;
import org.apache.hadoop.metrics2.lib.DefaultMetricsSystem;
import org.apache.hadoop.metrics2.lib.MetricsRegistry;
import org.apache.hadoop.metrics2.lib.MutableCounterLong;
import org.apache.hadoop.metrics2.lib.MutableQuantiles;
import org.apache.hadoop.metrics2.lib.MutableRate;
import org.apache.hadoop.metrics2.source.JvmMetrics;
/**
* This class is for maintaining the various NFS gateway activity statistics and
* publishing them through the metrics interfaces.
*/
@InterfaceAudience.Private
@Metrics(about = "Nfs3 metrics", context = "dfs")
public class Nfs3Metrics {
// All mutable rates are in nanoseconds
// No metric for nullProcedure;
@Metric MutableRate getattr;
@Metric MutableRate setattr;
@Metric MutableRate lookup;
@Metric MutableRate access;
@Metric MutableRate readlink;
@Metric MutableRate read;
final MutableQuantiles[] readNanosQuantiles;
@Metric MutableRate write;
final MutableQuantiles[] writeNanosQuantiles;
@Metric MutableRate create;
@Metric MutableRate mkdir;
@Metric MutableRate symlink;
@Metric MutableRate mknod;
@Metric MutableRate remove;
@Metric MutableRate rmdir;
@Metric MutableRate rename;
@Metric MutableRate link;
@Metric MutableRate readdir;
@Metric MutableRate readdirplus;
@Metric MutableRate fsstat;
@Metric MutableRate fsinfo;
@Metric MutableRate pathconf;
@Metric MutableRate commit;
final MutableQuantiles[] commitNanosQuantiles;
@Metric MutableCounterLong bytesWritten;
@Metric MutableCounterLong bytesRead;
final MetricsRegistry registry = new MetricsRegistry("nfs3");
final String name;
JvmMetrics jvmMetrics = null;
public Nfs3Metrics(String name, String sessionId, int[] intervals,
final JvmMetrics jvmMetrics) {
this.name = name;
this.jvmMetrics = jvmMetrics;
registry.tag(SessionId, sessionId);
final int len = intervals.length;
readNanosQuantiles = new MutableQuantiles[len];
writeNanosQuantiles = new MutableQuantiles[len];
commitNanosQuantiles = new MutableQuantiles[len];
for (int i = 0; i < len; i++) {
int interval = intervals[i];
readNanosQuantiles[i] = registry.newQuantiles("readProcessNanos"
+ interval + "s", "Read process in ns", "ops", "latency", interval);
writeNanosQuantiles[i] = registry.newQuantiles("writeProcessNanos"
+ interval + "s", " process in ns", "ops", "latency", interval);
commitNanosQuantiles[i] = registry.newQuantiles("commitProcessNanos"
+ interval + "s", "Read process in ns", "ops", "latency", interval);
}
}
public static Nfs3Metrics create(Configuration conf, String gatewayName) {
String sessionId = conf.get(DFSConfigKeys.DFS_METRICS_SESSION_ID_KEY);
MetricsSystem ms = DefaultMetricsSystem.instance();
JvmMetrics jm = JvmMetrics.create(gatewayName, sessionId, ms);
// Percentile measurement is [,,,] by default
int[] intervals = conf.getInts(conf.get(
NfsConfigKeys.NFS_METRICS_PERCENTILES_INTERVALS_KEY,
NfsConfigKeys.NFS_METRICS_PERCENTILES_INTERVALS_DEFAULT));
return ms.register(new Nfs3Metrics(gatewayName, sessionId, intervals, jm));
}
public String name() {
return name;
}
public JvmMetrics getJvmMetrics() {
return jvmMetrics;
}
public void incrBytesWritten(long bytes) {
bytesWritten.incr(bytes);
}
public void incrBytesRead(long bytes) {
bytesRead.incr(bytes);
}
public void addGetattr(long latencyNanos) {
getattr.add(latencyNanos);
}
public void addSetattr(long latencyNanos) {
setattr.add(latencyNanos);
}
public void addLookup(long latencyNanos) {
lookup.add(latencyNanos);
}
public void addAccess(long latencyNanos) {
access.add(latencyNanos);
}
public void addReadlink(long latencyNanos) {
readlink.add(latencyNanos);
}
public void addRead(long latencyNanos) {
read.add(latencyNanos);
for (MutableQuantiles q : readNanosQuantiles) {
q.add(latencyNanos);
}
}
public void addWrite(long latencyNanos) {
write.add(latencyNanos);
for (MutableQuantiles q : writeNanosQuantiles) {
q.add(latencyNanos);
}
}
public void addCreate(long latencyNanos) {
create.add(latencyNanos);
}
public void addMkdir(long latencyNanos) {
mkdir.add(latencyNanos);
}
public void addSymlink(long latencyNanos) {
symlink.add(latencyNanos);
}
public void addMknod(long latencyNanos) {
mknod.add(latencyNanos);
}
public void addRemove(long latencyNanos) {
remove.add(latencyNanos);
}
public void addRmdir(long latencyNanos) {
rmdir.add(latencyNanos);
}
public void addRename(long latencyNanos) {
rename.add(latencyNanos);
}
public void addLink(long latencyNanos) {
link.add(latencyNanos);
}
public void addReaddir(long latencyNanos) {
readdir.add(latencyNanos);
}
public void addReaddirplus(long latencyNanos) {
readdirplus.add(latencyNanos);
}
public void addFsstat(long latencyNanos) {
fsstat.add(latencyNanos);
}
public void addFsinfo(long latencyNanos) {
fsinfo.add(latencyNanos);
}
public void addPathconf(long latencyNanos) {
pathconf.add(latencyNanos);
}
public void addCommit(long latencyNanos) {
commit.add(latencyNanos);
for (MutableQuantiles q : commitNanosQuantiles) {
q.add(latencyNanos);
}
}
}

View File

@ -213,4 +213,8 @@ public static byte[] longToByte(long v) {
data[7] = (byte) (v >>> 0);
return data;
}
public static long getElapsedTime(long startTimeNano) {
return System.nanoTime() - startTimeNano;
}
}

View File

@ -129,9 +129,8 @@ static class CommitCtx {
private final Channel channel;
private final int xid;
private final Nfs3FileAttributes preOpAttr;
// Remember time for debug purpose
private final long startTime;
public final long startTime;
long getOffset() {
return offset;
@ -159,7 +158,7 @@ long getStartTime() {
this.channel = channel;
this.xid = xid;
this.preOpAttr = preOpAttr;
this.startTime = Time.monotonicNow();
this.startTime = System.nanoTime();
}
@Override
@ -687,6 +686,8 @@ private void receivedNewWriteInternal(DFSClient dfsClient,
WccData fileWcc = new WccData(preOpAttr, latestAttr);
WRITE3Response response = new WRITE3Response(Nfs3Status.NFS3_OK,
fileWcc, count, stableHow, Nfs3Constant.WRITE_COMMIT_VERF);
RpcProgramNfs3.metrics.addWrite(Nfs3Utils
.getElapsedTime(writeCtx.startTime));
Nfs3Utils
.writeChannel(channel, response.serialize(new XDR(),
xid, new VerifierNone()), xid);
@ -1131,14 +1132,16 @@ private void processCommits(long offset) {
COMMIT3Response response = new COMMIT3Response(status, wccData,
Nfs3Constant.WRITE_COMMIT_VERF);
RpcProgramNfs3.metrics.addCommit(Nfs3Utils
.getElapsedTime(commit.startTime));
Nfs3Utils.writeChannelCommit(commit.getChannel(), response
.serialize(new XDR(), commit.getXid(),
new VerifierNone()), commit.getXid());
if (LOG.isDebugEnabled()) {
LOG.debug("FileId: " + latestAttr.getFileId() + " Service time:"
+ (Time.monotonicNow() - commit.getStartTime())
+ "ms. Sent response for commit:" + commit);
+ Nfs3Utils.getElapsedTime(commit.startTime)
+ "ns. Sent response for commit:" + commit);
}
entry = pendingCommits.firstEntry();
}
@ -1162,6 +1165,7 @@ private void doSingleWrite(final WriteCtx writeCtx) {
// The write is not protected by lock. asyncState is used to make sure
// there is one thread doing write back at any time
writeCtx.writeData(fos);
RpcProgramNfs3.metrics.incrBytesWritten(writeCtx.getCount());
long flushedOffset = getFlushedOffset();
if (flushedOffset != (offset + count)) {
@ -1213,6 +1217,7 @@ private void doSingleWrite(final WriteCtx writeCtx) {
}
WRITE3Response response = new WRITE3Response(Nfs3Status.NFS3_OK,
fileWcc, count, stableHow, Nfs3Constant.WRITE_COMMIT_VERF);
RpcProgramNfs3.metrics.addWrite(Nfs3Utils.getElapsedTime(writeCtx.startTime));
Nfs3Utils.writeChannel(channel, response.serialize(
new XDR(), xid, new VerifierNone()), xid);
}

View File

@ -48,6 +48,8 @@
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
import org.apache.hadoop.hdfs.protocol.HdfsFileStatus;
import org.apache.hadoop.ipc.RemoteException;
import org.apache.hadoop.metrics2.lib.DefaultMetricsSystem;
import org.apache.hadoop.net.DNS;
import org.apache.hadoop.nfs.AccessPrivilege;
import org.apache.hadoop.nfs.NfsExports;
import org.apache.hadoop.nfs.NfsFileType;
@ -164,6 +166,7 @@ public class RpcProgramNfs3 extends RpcProgram implements Nfs3Interface {
private final RpcCallCache rpcCallCache;
private JvmPauseMonitor pauseMonitor;
private Nfs3HttpServer infoServer = null;
static Nfs3Metrics metrics;
public RpcProgramNfs3(NfsConfiguration config, DatagramSocket registrationSocket,
boolean allowInsecurePorts) throws IOException {
@ -209,6 +212,17 @@ public RpcProgramNfs3(NfsConfiguration config, DatagramSocket registrationSocket
infoServer = new Nfs3HttpServer(config);
}
public static RpcProgramNfs3 createRpcProgramNfs3(NfsConfiguration config,
DatagramSocket registrationSocket, boolean allowInsecurePorts)
throws IOException {
DefaultMetricsSystem.initialize("Nfs3");
String displayName = DNS.getDefaultHost("default", "default")
+ config.getInt(NfsConfigKeys.DFS_NFS_SERVER_PORT_KEY,
NfsConfigKeys.DFS_NFS_SERVER_PORT_DEFAULT);
metrics = Nfs3Metrics.create(config, displayName);
return new RpcProgramNfs3(config, registrationSocket, allowInsecurePorts);
}
private void clearDirectory(String writeDumpDir) throws IOException {
File dumpDir = new File(writeDumpDir);
if (dumpDir.exists()) {
@ -225,10 +239,11 @@ private void clearDirectory(String writeDumpDir) throws IOException {
}
@Override
public void startDaemons() {
public void startDaemons() {
if (pauseMonitor == null) {
pauseMonitor = new JvmPauseMonitor(config);
pauseMonitor.start();
metrics.getJvmMetrics().setPauseMonitor(pauseMonitor);
}
writeManager.startAsyncDataSerivce();
try {
@ -770,6 +785,7 @@ READ3Response read(XDR xdr, SecurityHandler securityHandler,
try {
readCount = fis.read(offset, readbuffer, 0, count);
metrics.incrBytesRead(readCount);
} catch (IOException e) {
// TODO: A cleaner way is to throw a new type of exception
// which requires incompatible changes.
@ -2049,8 +2065,8 @@ COMMIT3Response commit(XDR xdr, Channel channel, int xid,
: (request.getOffset() + request.getCount());
// Insert commit as an async request
writeManager.handleCommit(dfsClient, handle, commitOffset,
channel, xid, preOpAttr);
writeManager.handleCommit(dfsClient, handle, commitOffset, channel, xid,
preOpAttr);
return null;
} catch (IOException e) {
LOG.warn("Exception ", e);
@ -2132,20 +2148,29 @@ public void handleInternal(ChannelHandlerContext ctx, RpcInfo info) {
}
}
}
// Since write and commit could be async, they use their own startTime and
// only record success requests.
final long startTime = System.nanoTime();
NFS3Response response = null;
if (nfsproc3 == NFSPROC3.NULL) {
response = nullProcedure();
} else if (nfsproc3 == NFSPROC3.GETATTR) {
response = getattr(xdr, info);
metrics.addGetattr(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.SETATTR) {
response = setattr(xdr, info);
metrics.addSetattr(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.LOOKUP) {
response = lookup(xdr, info);
metrics.addLookup(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.ACCESS) {
response = access(xdr, info);
metrics.addAccess(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.READLINK) {
response = readlink(xdr, info);
metrics.addReadlink(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.READ) {
if (LOG.isDebugEnabled()) {
LOG.debug(Nfs3Utils.READ_RPC_START + xid);
@ -2154,6 +2179,7 @@ public void handleInternal(ChannelHandlerContext ctx, RpcInfo info) {
if (LOG.isDebugEnabled() && (nfsproc3 == NFSPROC3.READ)) {
LOG.debug(Nfs3Utils.READ_RPC_END + xid);
}
metrics.addRead(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.WRITE) {
if (LOG.isDebugEnabled()) {
LOG.debug(Nfs3Utils.WRITE_RPC_START + xid);
@ -2162,30 +2188,43 @@ public void handleInternal(ChannelHandlerContext ctx, RpcInfo info) {
// Write end debug trace is in Nfs3Utils.writeChannel
} else if (nfsproc3 == NFSPROC3.CREATE) {
response = create(xdr, info);
metrics.addCreate(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.MKDIR) {
response = mkdir(xdr, info);
metrics.addMkdir(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.SYMLINK) {
response = symlink(xdr, info);
metrics.addSymlink(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.MKNOD) {
response = mknod(xdr, info);
metrics.addMknod(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.REMOVE) {
response = remove(xdr, info);
metrics.addRemove(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.RMDIR) {
response = rmdir(xdr, info);
metrics.addRmdir(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.RENAME) {
response = rename(xdr, info);
metrics.addRename(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.LINK) {
response = link(xdr, info);
metrics.addLink(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.READDIR) {
response = readdir(xdr, info);
metrics.addReaddir(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.READDIRPLUS) {
response = readdirplus(xdr, info);
metrics.addReaddirplus(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.FSSTAT) {
response = fsstat(xdr, info);
metrics.addFsstat(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.FSINFO) {
response = fsinfo(xdr, info);
metrics.addFsinfo(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.PATHCONF) {
response = pathconf(xdr,info);
response = pathconf(xdr, info);
metrics.addPathconf(Nfs3Utils.getElapsedTime(startTime));
} else if (nfsproc3 == NFSPROC3.COMMIT) {
response = commit(xdr, info);
} else {

View File

@ -84,7 +84,8 @@ public int getOriginalCount() {
private long dumpFileOffset;
private volatile DataState dataState;
public final long startTime;
public DataState getDataState() {
return dataState;
}
@ -235,6 +236,7 @@ void setReplied(boolean replied) {
this.replied = replied;
this.dataState = dataState;
raf = null;
this.startTime = System.nanoTime();
}
@Override

View File

@ -224,6 +224,7 @@ int commitBeforeRead(DFSClient dfsClient, FileHandle fileHandle,
status = Nfs3Status.NFS3_OK;
} else {
// commit request triggered by read won't create pending comment obj
COMMIT_STATUS ret = openFileCtx.checkCommit(dfsClient, commitOffset,
null, 0, null, true);
switch (ret) {
@ -260,6 +261,7 @@ int commitBeforeRead(DFSClient dfsClient, FileHandle fileHandle,
void handleCommit(DFSClient dfsClient, FileHandle fileHandle,
long commitOffset, Channel channel, int xid, Nfs3FileAttributes preOpAttr) {
long startTime = System.nanoTime();
int status;
OpenFileCtx openFileCtx = fileContextCache.get(fileHandle);
@ -306,9 +308,9 @@ void handleCommit(DFSClient dfsClient, FileHandle fileHandle,
WccData fileWcc = new WccData(Nfs3Utils.getWccAttr(preOpAttr), postOpAttr);
COMMIT3Response response = new COMMIT3Response(status, fileWcc,
Nfs3Constant.WRITE_COMMIT_VERF);
RpcProgramNfs3.metrics.addCommit(Nfs3Utils.getElapsedTime(startTime));
Nfs3Utils.writeChannelCommit(channel,
response.serialize(new XDR(), xid, new VerifierNone()),
xid);
response.serialize(new XDR(), xid, new VerifierNone()), xid);
}
/**

View File

@ -48,6 +48,10 @@ public static void setUp() throws Exception {
HttpConfig.Policy.HTTP_AND_HTTPS.name());
conf.set(NfsConfigKeys.NFS_HTTP_ADDRESS_KEY, "localhost:0");
conf.set(NfsConfigKeys.NFS_HTTPS_ADDRESS_KEY, "localhost:0");
// Use emphral port in case tests are running in parallel
conf.setInt(NfsConfigKeys.DFS_NFS_SERVER_PORT_KEY, 0);
conf.setInt(NfsConfigKeys.DFS_NFS_MOUNTD_PORT_KEY, 0);
File base = new File(BASEDIR);
FileUtil.fullyDelete(base);
base.mkdirs();

View File

@ -273,6 +273,8 @@ Release 2.7.0 - UNRELEASED
(Maysam Yabandeh via wang)
HDFS-7424. Add web UI for NFS gateway (brandonli)
HDFS-7449. Add metrics to NFS gateway (brandonli)
IMPROVEMENTS