YARN-5638. Introduce a collector timestamp to uniquely identify collectors creation order in collector discovery. Contributed by Li Lu.
This commit is contained in:
parent
e72e6260da
commit
78b7e070d8
@ -24,6 +24,7 @@
|
|||||||
|
|
||||||
import org.apache.hadoop.yarn.api.records.ApplicationId;
|
import org.apache.hadoop.yarn.api.records.ApplicationId;
|
||||||
import org.apache.hadoop.yarn.api.records.NodeLabel;
|
import org.apache.hadoop.yarn.api.records.NodeLabel;
|
||||||
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.api.records.MasterKey;
|
import org.apache.hadoop.yarn.server.api.records.MasterKey;
|
||||||
import org.apache.hadoop.yarn.server.api.records.NodeStatus;
|
import org.apache.hadoop.yarn.server.api.records.NodeStatus;
|
||||||
import org.apache.hadoop.yarn.util.Records;
|
import org.apache.hadoop.yarn.util.Records;
|
||||||
@ -47,7 +48,7 @@ public static NodeHeartbeatRequest newInstance(NodeStatus nodeStatus,
|
|||||||
public static NodeHeartbeatRequest newInstance(NodeStatus nodeStatus,
|
public static NodeHeartbeatRequest newInstance(NodeStatus nodeStatus,
|
||||||
MasterKey lastKnownContainerTokenMasterKey,
|
MasterKey lastKnownContainerTokenMasterKey,
|
||||||
MasterKey lastKnownNMTokenMasterKey, Set<NodeLabel> nodeLabels,
|
MasterKey lastKnownNMTokenMasterKey, Set<NodeLabel> nodeLabels,
|
||||||
Map<ApplicationId, String> registeredCollectors) {
|
Map<ApplicationId, AppCollectorData> registeringCollectors) {
|
||||||
NodeHeartbeatRequest nodeHeartbeatRequest =
|
NodeHeartbeatRequest nodeHeartbeatRequest =
|
||||||
Records.newRecord(NodeHeartbeatRequest.class);
|
Records.newRecord(NodeHeartbeatRequest.class);
|
||||||
nodeHeartbeatRequest.setNodeStatus(nodeStatus);
|
nodeHeartbeatRequest.setNodeStatus(nodeStatus);
|
||||||
@ -56,7 +57,7 @@ public static NodeHeartbeatRequest newInstance(NodeStatus nodeStatus,
|
|||||||
nodeHeartbeatRequest
|
nodeHeartbeatRequest
|
||||||
.setLastKnownNMTokenMasterKey(lastKnownNMTokenMasterKey);
|
.setLastKnownNMTokenMasterKey(lastKnownNMTokenMasterKey);
|
||||||
nodeHeartbeatRequest.setNodeLabels(nodeLabels);
|
nodeHeartbeatRequest.setNodeLabels(nodeLabels);
|
||||||
nodeHeartbeatRequest.setRegisteredCollectors(registeredCollectors);
|
nodeHeartbeatRequest.setRegisteringCollectors(registeringCollectors);
|
||||||
return nodeHeartbeatRequest;
|
return nodeHeartbeatRequest;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -79,7 +80,9 @@ public abstract void setLogAggregationReportsForApps(
|
|||||||
List<LogAggregationReport> logAggregationReportsForApps);
|
List<LogAggregationReport> logAggregationReportsForApps);
|
||||||
|
|
||||||
// This tells RM registered collectors' address info on this node
|
// This tells RM registered collectors' address info on this node
|
||||||
public abstract Map<ApplicationId, String> getRegisteredCollectors();
|
public abstract Map<ApplicationId, AppCollectorData>
|
||||||
public abstract void setRegisteredCollectors(Map<ApplicationId,
|
getRegisteringCollectors();
|
||||||
String> appCollectorsMap);
|
|
||||||
|
public abstract void setRegisteringCollectors(Map<ApplicationId,
|
||||||
|
AppCollectorData> appCollectorsMap);
|
||||||
}
|
}
|
||||||
|
@ -28,6 +28,7 @@
|
|||||||
import org.apache.hadoop.yarn.api.records.Container;
|
import org.apache.hadoop.yarn.api.records.Container;
|
||||||
import org.apache.hadoop.yarn.api.records.ContainerId;
|
import org.apache.hadoop.yarn.api.records.ContainerId;
|
||||||
import org.apache.hadoop.yarn.api.records.Resource;
|
import org.apache.hadoop.yarn.api.records.Resource;
|
||||||
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.api.records.ContainerQueuingLimit;
|
import org.apache.hadoop.yarn.server.api.records.ContainerQueuingLimit;
|
||||||
import org.apache.hadoop.yarn.server.api.records.MasterKey;
|
import org.apache.hadoop.yarn.server.api.records.MasterKey;
|
||||||
import org.apache.hadoop.yarn.server.api.records.NodeAction;
|
import org.apache.hadoop.yarn.server.api.records.NodeAction;
|
||||||
@ -47,10 +48,9 @@ public abstract class NodeHeartbeatResponse {
|
|||||||
public abstract List<ApplicationId> getApplicationsToCleanup();
|
public abstract List<ApplicationId> getApplicationsToCleanup();
|
||||||
|
|
||||||
// This tells NM the collectors' address info of related apps
|
// This tells NM the collectors' address info of related apps
|
||||||
public abstract Map<ApplicationId, String> getAppCollectorsMap();
|
public abstract Map<ApplicationId, AppCollectorData> getAppCollectors();
|
||||||
|
public abstract void setAppCollectors(
|
||||||
public abstract void setAppCollectorsMap(
|
Map<ApplicationId, AppCollectorData> appCollectorsMap);
|
||||||
Map<ApplicationId, String> appCollectorsMap);
|
|
||||||
|
|
||||||
public abstract void setResponseId(int responseId);
|
public abstract void setResponseId(int responseId);
|
||||||
|
|
||||||
|
@ -22,14 +22,14 @@
|
|||||||
|
|
||||||
import org.apache.hadoop.classification.InterfaceAudience.Private;
|
import org.apache.hadoop.classification.InterfaceAudience.Private;
|
||||||
import org.apache.hadoop.yarn.api.records.ApplicationId;
|
import org.apache.hadoop.yarn.api.records.ApplicationId;
|
||||||
import org.apache.hadoop.yarn.server.api.records.AppCollectorsMap;
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.util.Records;
|
import org.apache.hadoop.yarn.util.Records;
|
||||||
|
|
||||||
@Private
|
@Private
|
||||||
public abstract class ReportNewCollectorInfoRequest {
|
public abstract class ReportNewCollectorInfoRequest {
|
||||||
|
|
||||||
public static ReportNewCollectorInfoRequest newInstance(
|
public static ReportNewCollectorInfoRequest newInstance(
|
||||||
List<AppCollectorsMap> appCollectorsList) {
|
List<AppCollectorData> appCollectorsList) {
|
||||||
ReportNewCollectorInfoRequest request =
|
ReportNewCollectorInfoRequest request =
|
||||||
Records.newRecord(ReportNewCollectorInfoRequest.class);
|
Records.newRecord(ReportNewCollectorInfoRequest.class);
|
||||||
request.setAppCollectorsList(appCollectorsList);
|
request.setAppCollectorsList(appCollectorsList);
|
||||||
@ -41,13 +41,13 @@ public static ReportNewCollectorInfoRequest newInstance(
|
|||||||
ReportNewCollectorInfoRequest request =
|
ReportNewCollectorInfoRequest request =
|
||||||
Records.newRecord(ReportNewCollectorInfoRequest.class);
|
Records.newRecord(ReportNewCollectorInfoRequest.class);
|
||||||
request.setAppCollectorsList(
|
request.setAppCollectorsList(
|
||||||
Arrays.asList(AppCollectorsMap.newInstance(id, collectorAddr)));
|
Arrays.asList(AppCollectorData.newInstance(id, collectorAddr)));
|
||||||
return request;
|
return request;
|
||||||
}
|
}
|
||||||
|
|
||||||
public abstract List<AppCollectorsMap> getAppCollectorsList();
|
public abstract List<AppCollectorData> getAppCollectorsList();
|
||||||
|
|
||||||
public abstract void setAppCollectorsList(
|
public abstract void setAppCollectorsList(
|
||||||
List<AppCollectorsMap> appCollectorsList);
|
List<AppCollectorData> appCollectorsList);
|
||||||
|
|
||||||
}
|
}
|
||||||
|
@ -35,13 +35,14 @@
|
|||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonProtos.MasterKeyProto;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonProtos.MasterKeyProto;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonProtos.NodeStatusProto;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonProtos.NodeStatusProto;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.LogAggregationReportProto;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.LogAggregationReportProto;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.AppCollectorsMapProto;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.AppCollectorDataProto;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.NodeHeartbeatRequestProto;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.NodeHeartbeatRequestProto;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.NodeHeartbeatRequestProtoOrBuilder;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.NodeHeartbeatRequestProtoOrBuilder;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.NodeLabelsProto;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.NodeLabelsProto;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.NodeLabelsProto.Builder;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.NodeLabelsProto.Builder;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.LogAggregationReport;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.LogAggregationReport;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.NodeHeartbeatRequest;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.NodeHeartbeatRequest;
|
||||||
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.api.records.MasterKey;
|
import org.apache.hadoop.yarn.server.api.records.MasterKey;
|
||||||
import org.apache.hadoop.yarn.server.api.records.NodeStatus;
|
import org.apache.hadoop.yarn.server.api.records.NodeStatus;
|
||||||
import org.apache.hadoop.yarn.server.api.records.impl.pb.MasterKeyPBImpl;
|
import org.apache.hadoop.yarn.server.api.records.impl.pb.MasterKeyPBImpl;
|
||||||
@ -58,7 +59,7 @@ public class NodeHeartbeatRequestPBImpl extends NodeHeartbeatRequest {
|
|||||||
private Set<NodeLabel> labels = null;
|
private Set<NodeLabel> labels = null;
|
||||||
private List<LogAggregationReport> logAggregationReportsForApps = null;
|
private List<LogAggregationReport> logAggregationReportsForApps = null;
|
||||||
|
|
||||||
private Map<ApplicationId, String> registeredCollectors = null;
|
private Map<ApplicationId, AppCollectorData> registeringCollectors = null;
|
||||||
|
|
||||||
public NodeHeartbeatRequestPBImpl() {
|
public NodeHeartbeatRequestPBImpl() {
|
||||||
builder = NodeHeartbeatRequestProto.newBuilder();
|
builder = NodeHeartbeatRequestProto.newBuilder();
|
||||||
@ -114,8 +115,8 @@ private void mergeLocalToBuilder() {
|
|||||||
if (this.logAggregationReportsForApps != null) {
|
if (this.logAggregationReportsForApps != null) {
|
||||||
addLogAggregationStatusForAppsToProto();
|
addLogAggregationStatusForAppsToProto();
|
||||||
}
|
}
|
||||||
if (this.registeredCollectors != null) {
|
if (this.registeringCollectors != null) {
|
||||||
addRegisteredCollectorsToProto();
|
addRegisteringCollectorsToProto();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -158,14 +159,14 @@ private LogAggregationReportProto convertToProtoFormat(
|
|||||||
return ((LogAggregationReportPBImpl) value).getProto();
|
return ((LogAggregationReportPBImpl) value).getProto();
|
||||||
}
|
}
|
||||||
|
|
||||||
private void addRegisteredCollectorsToProto() {
|
private void addRegisteringCollectorsToProto() {
|
||||||
maybeInitBuilder();
|
maybeInitBuilder();
|
||||||
builder.clearRegisteredCollectors();
|
builder.clearRegisteringCollectors();
|
||||||
for (Map.Entry<ApplicationId, String> entry :
|
for (Map.Entry<ApplicationId, AppCollectorData> entry :
|
||||||
registeredCollectors.entrySet()) {
|
registeringCollectors.entrySet()) {
|
||||||
builder.addRegisteredCollectors(AppCollectorsMapProto.newBuilder()
|
builder.addRegisteringCollectors(AppCollectorDataProto.newBuilder()
|
||||||
.setAppId(convertToProtoFormat(entry.getKey()))
|
.setAppId(convertToProtoFormat(entry.getKey()))
|
||||||
.setAppCollectorAddr(entry.getValue()));
|
.setAppCollectorAddr(entry.getValue().getCollectorAddr()));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -251,35 +252,37 @@ public void setLastKnownNMTokenMasterKey(MasterKey masterKey) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public Map<ApplicationId, String> getRegisteredCollectors() {
|
public Map<ApplicationId, AppCollectorData> getRegisteringCollectors() {
|
||||||
if (this.registeredCollectors != null) {
|
if (this.registeringCollectors != null) {
|
||||||
return this.registeredCollectors;
|
return this.registeringCollectors;
|
||||||
}
|
}
|
||||||
initRegisteredCollectors();
|
initRegisteredCollectors();
|
||||||
return registeredCollectors;
|
return registeringCollectors;
|
||||||
}
|
}
|
||||||
|
|
||||||
private void initRegisteredCollectors() {
|
private void initRegisteredCollectors() {
|
||||||
NodeHeartbeatRequestProtoOrBuilder p = viaProto ? proto : builder;
|
NodeHeartbeatRequestProtoOrBuilder p = viaProto ? proto : builder;
|
||||||
List<AppCollectorsMapProto> list = p.getRegisteredCollectorsList();
|
List<AppCollectorDataProto> list = p.getRegisteringCollectorsList();
|
||||||
if (!list.isEmpty()) {
|
if (!list.isEmpty()) {
|
||||||
this.registeredCollectors = new HashMap<>();
|
this.registeringCollectors = new HashMap<>();
|
||||||
for (AppCollectorsMapProto c : list) {
|
for (AppCollectorDataProto c : list) {
|
||||||
ApplicationId appId = convertFromProtoFormat(c.getAppId());
|
ApplicationId appId = convertFromProtoFormat(c.getAppId());
|
||||||
this.registeredCollectors.put(appId, c.getAppCollectorAddr());
|
AppCollectorData data = AppCollectorData.newInstance(appId,
|
||||||
|
c.getAppCollectorAddr(), c.getRmIdentifier(), c.getVersion());
|
||||||
|
this.registeringCollectors.put(appId, data);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void setRegisteredCollectors(
|
public void setRegisteringCollectors(
|
||||||
Map<ApplicationId, String> registeredCollectors) {
|
Map<ApplicationId, AppCollectorData> registeredCollectors) {
|
||||||
if (registeredCollectors == null || registeredCollectors.isEmpty()) {
|
if (registeredCollectors == null || registeredCollectors.isEmpty()) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
maybeInitBuilder();
|
maybeInitBuilder();
|
||||||
this.registeredCollectors = new HashMap<ApplicationId, String>();
|
this.registeringCollectors = new HashMap<>();
|
||||||
this.registeredCollectors.putAll(registeredCollectors);
|
this.registeringCollectors.putAll(registeredCollectors);
|
||||||
}
|
}
|
||||||
|
|
||||||
private NodeStatusPBImpl convertFromProtoFormat(NodeStatusProto p) {
|
private NodeStatusPBImpl convertFromProtoFormat(NodeStatusProto p) {
|
||||||
|
@ -45,11 +45,12 @@
|
|||||||
import org.apache.hadoop.yarn.proto.YarnServiceProtos.SignalContainerRequestProto;
|
import org.apache.hadoop.yarn.proto.YarnServiceProtos.SignalContainerRequestProto;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonProtos.MasterKeyProto;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonProtos.MasterKeyProto;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonProtos.NodeActionProto;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonProtos.NodeActionProto;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.AppCollectorsMapProto;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.AppCollectorDataProto;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.NodeHeartbeatResponseProto;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.NodeHeartbeatResponseProto;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.NodeHeartbeatResponseProtoOrBuilder;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.NodeHeartbeatResponseProtoOrBuilder;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.SystemCredentialsForAppsProto;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.SystemCredentialsForAppsProto;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.NodeHeartbeatResponse;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.NodeHeartbeatResponse;
|
||||||
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.api.records.ContainerQueuingLimit;
|
import org.apache.hadoop.yarn.server.api.records.ContainerQueuingLimit;
|
||||||
import org.apache.hadoop.yarn.server.api.records.MasterKey;
|
import org.apache.hadoop.yarn.server.api.records.MasterKey;
|
||||||
import org.apache.hadoop.yarn.server.api.records.NodeAction;
|
import org.apache.hadoop.yarn.server.api.records.NodeAction;
|
||||||
@ -70,7 +71,7 @@ public class NodeHeartbeatResponsePBImpl extends NodeHeartbeatResponse {
|
|||||||
private List<ApplicationId> applicationsToCleanup = null;
|
private List<ApplicationId> applicationsToCleanup = null;
|
||||||
private Map<ApplicationId, ByteBuffer> systemCredentials = null;
|
private Map<ApplicationId, ByteBuffer> systemCredentials = null;
|
||||||
private Resource resource = null;
|
private Resource resource = null;
|
||||||
private Map<ApplicationId, String> appCollectorsMap = null;
|
private Map<ApplicationId, AppCollectorData> appCollectorsMap = null;
|
||||||
|
|
||||||
private MasterKey containerTokenMasterKey = null;
|
private MasterKey containerTokenMasterKey = null;
|
||||||
private MasterKey nmTokenMasterKey = null;
|
private MasterKey nmTokenMasterKey = null;
|
||||||
@ -146,11 +147,15 @@ private void addSystemCredentialsToProto() {
|
|||||||
|
|
||||||
private void addAppCollectorsMapToProto() {
|
private void addAppCollectorsMapToProto() {
|
||||||
maybeInitBuilder();
|
maybeInitBuilder();
|
||||||
builder.clearAppCollectorsMap();
|
builder.clearAppCollectors();
|
||||||
for (Map.Entry<ApplicationId, String> entry : appCollectorsMap.entrySet()) {
|
for (Map.Entry<ApplicationId, AppCollectorData> entry
|
||||||
builder.addAppCollectorsMap(AppCollectorsMapProto.newBuilder()
|
: appCollectorsMap.entrySet()) {
|
||||||
|
AppCollectorData data = entry.getValue();
|
||||||
|
builder.addAppCollectors(AppCollectorDataProto.newBuilder()
|
||||||
.setAppId(convertToProtoFormat(entry.getKey()))
|
.setAppId(convertToProtoFormat(entry.getKey()))
|
||||||
.setAppCollectorAddr(entry.getValue()));
|
.setAppCollectorAddr(data.getCollectorAddr())
|
||||||
|
.setRmIdentifier(data.getRMIdentifier())
|
||||||
|
.setVersion(data.getVersion()));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -568,7 +573,7 @@ public Map<ApplicationId, ByteBuffer> getSystemCredentialsForApps() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public Map<ApplicationId, String> getAppCollectorsMap() {
|
public Map<ApplicationId, AppCollectorData> getAppCollectors() {
|
||||||
if (this.appCollectorsMap != null) {
|
if (this.appCollectorsMap != null) {
|
||||||
return this.appCollectorsMap;
|
return this.appCollectorsMap;
|
||||||
}
|
}
|
||||||
@ -589,12 +594,14 @@ private void initSystemCredentials() {
|
|||||||
|
|
||||||
private void initAppCollectorsMap() {
|
private void initAppCollectorsMap() {
|
||||||
NodeHeartbeatResponseProtoOrBuilder p = viaProto ? proto : builder;
|
NodeHeartbeatResponseProtoOrBuilder p = viaProto ? proto : builder;
|
||||||
List<AppCollectorsMapProto> list = p.getAppCollectorsMapList();
|
List<AppCollectorDataProto> list = p.getAppCollectorsList();
|
||||||
if (!list.isEmpty()) {
|
if (!list.isEmpty()) {
|
||||||
this.appCollectorsMap = new HashMap<>();
|
this.appCollectorsMap = new HashMap<>();
|
||||||
for (AppCollectorsMapProto c : list) {
|
for (AppCollectorDataProto c : list) {
|
||||||
ApplicationId appId = convertFromProtoFormat(c.getAppId());
|
ApplicationId appId = convertFromProtoFormat(c.getAppId());
|
||||||
this.appCollectorsMap.put(appId, c.getAppCollectorAddr());
|
AppCollectorData data = AppCollectorData.newInstance(appId,
|
||||||
|
c.getAppCollectorAddr(), c.getRmIdentifier(), c.getVersion());
|
||||||
|
this.appCollectorsMap.put(appId, data);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -611,14 +618,14 @@ public void setSystemCredentialsForApps(
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void setAppCollectorsMap(
|
public void setAppCollectors(
|
||||||
Map<ApplicationId, String> appCollectorsMap) {
|
Map<ApplicationId, AppCollectorData> appCollectors) {
|
||||||
if (appCollectorsMap == null || appCollectorsMap.isEmpty()) {
|
if (appCollectors == null || appCollectors.isEmpty()) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
maybeInitBuilder();
|
maybeInitBuilder();
|
||||||
this.appCollectorsMap = new HashMap<ApplicationId, String>();
|
this.appCollectorsMap = new HashMap<>();
|
||||||
this.appCollectorsMap.putAll(appCollectorsMap);
|
this.appCollectorsMap.putAll(appCollectors);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -20,12 +20,12 @@
|
|||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.AppCollectorsMapProto;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.AppCollectorDataProto;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.ReportNewCollectorInfoRequestProto;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.ReportNewCollectorInfoRequestProto;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.ReportNewCollectorInfoRequestProtoOrBuilder;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.ReportNewCollectorInfoRequestProtoOrBuilder;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.ReportNewCollectorInfoRequest;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.ReportNewCollectorInfoRequest;
|
||||||
import org.apache.hadoop.yarn.server.api.records.AppCollectorsMap;
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.api.records.impl.pb.AppCollectorsMapPBImpl;
|
import org.apache.hadoop.yarn.server.api.records.impl.pb.AppCollectorDataPBImpl;
|
||||||
|
|
||||||
public class ReportNewCollectorInfoRequestPBImpl extends
|
public class ReportNewCollectorInfoRequestPBImpl extends
|
||||||
ReportNewCollectorInfoRequest {
|
ReportNewCollectorInfoRequest {
|
||||||
@ -36,7 +36,7 @@ public class ReportNewCollectorInfoRequestPBImpl extends
|
|||||||
private ReportNewCollectorInfoRequestProto.Builder builder = null;
|
private ReportNewCollectorInfoRequestProto.Builder builder = null;
|
||||||
private boolean viaProto = false;
|
private boolean viaProto = false;
|
||||||
|
|
||||||
private List<AppCollectorsMap> collectorsList = null;
|
private List<AppCollectorData> collectorsList = null;
|
||||||
|
|
||||||
public ReportNewCollectorInfoRequestPBImpl() {
|
public ReportNewCollectorInfoRequestPBImpl() {
|
||||||
builder = ReportNewCollectorInfoRequestProto.newBuilder();
|
builder = ReportNewCollectorInfoRequestProto.newBuilder();
|
||||||
@ -96,9 +96,9 @@ private void maybeInitBuilder() {
|
|||||||
private void addLocalCollectorsToProto() {
|
private void addLocalCollectorsToProto() {
|
||||||
maybeInitBuilder();
|
maybeInitBuilder();
|
||||||
builder.clearAppCollectors();
|
builder.clearAppCollectors();
|
||||||
List<AppCollectorsMapProto> protoList =
|
List<AppCollectorDataProto> protoList =
|
||||||
new ArrayList<AppCollectorsMapProto>();
|
new ArrayList<AppCollectorDataProto>();
|
||||||
for (AppCollectorsMap m : this.collectorsList) {
|
for (AppCollectorData m : this.collectorsList) {
|
||||||
protoList.add(convertToProtoFormat(m));
|
protoList.add(convertToProtoFormat(m));
|
||||||
}
|
}
|
||||||
builder.addAllAppCollectors(protoList);
|
builder.addAllAppCollectors(protoList);
|
||||||
@ -106,16 +106,16 @@ private void addLocalCollectorsToProto() {
|
|||||||
|
|
||||||
private void initLocalCollectorsList() {
|
private void initLocalCollectorsList() {
|
||||||
ReportNewCollectorInfoRequestProtoOrBuilder p = viaProto ? proto : builder;
|
ReportNewCollectorInfoRequestProtoOrBuilder p = viaProto ? proto : builder;
|
||||||
List<AppCollectorsMapProto> list =
|
List<AppCollectorDataProto> list =
|
||||||
p.getAppCollectorsList();
|
p.getAppCollectorsList();
|
||||||
this.collectorsList = new ArrayList<AppCollectorsMap>();
|
this.collectorsList = new ArrayList<AppCollectorData>();
|
||||||
for (AppCollectorsMapProto m : list) {
|
for (AppCollectorDataProto m : list) {
|
||||||
this.collectorsList.add(convertFromProtoFormat(m));
|
this.collectorsList.add(convertFromProtoFormat(m));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public List<AppCollectorsMap> getAppCollectorsList() {
|
public List<AppCollectorData> getAppCollectorsList() {
|
||||||
if (this.collectorsList == null) {
|
if (this.collectorsList == null) {
|
||||||
initLocalCollectorsList();
|
initLocalCollectorsList();
|
||||||
}
|
}
|
||||||
@ -123,7 +123,7 @@ public List<AppCollectorsMap> getAppCollectorsList() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void setAppCollectorsList(List<AppCollectorsMap> appCollectorsList) {
|
public void setAppCollectorsList(List<AppCollectorData> appCollectorsList) {
|
||||||
maybeInitBuilder();
|
maybeInitBuilder();
|
||||||
if (appCollectorsList == null) {
|
if (appCollectorsList == null) {
|
||||||
builder.clearAppCollectors();
|
builder.clearAppCollectors();
|
||||||
@ -131,14 +131,14 @@ public void setAppCollectorsList(List<AppCollectorsMap> appCollectorsList) {
|
|||||||
this.collectorsList = appCollectorsList;
|
this.collectorsList = appCollectorsList;
|
||||||
}
|
}
|
||||||
|
|
||||||
private AppCollectorsMapPBImpl convertFromProtoFormat(
|
private AppCollectorDataPBImpl convertFromProtoFormat(
|
||||||
AppCollectorsMapProto p) {
|
AppCollectorDataProto p) {
|
||||||
return new AppCollectorsMapPBImpl(p);
|
return new AppCollectorDataPBImpl(p);
|
||||||
}
|
}
|
||||||
|
|
||||||
private AppCollectorsMapProto convertToProtoFormat(
|
private AppCollectorDataProto convertToProtoFormat(
|
||||||
AppCollectorsMap m) {
|
AppCollectorData m) {
|
||||||
return ((AppCollectorsMapPBImpl) m).getProto();
|
return ((AppCollectorDataPBImpl) m).getProto();
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
@ -0,0 +1,104 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.yarn.server.api.records;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience.Private;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
import org.apache.hadoop.yarn.api.records.ApplicationId;
|
||||||
|
import org.apache.hadoop.yarn.util.Records;
|
||||||
|
|
||||||
|
|
||||||
|
@Private
|
||||||
|
@InterfaceStability.Unstable
|
||||||
|
public abstract class AppCollectorData {
|
||||||
|
|
||||||
|
protected static final long DEFAULT_TIMESTAMP_VALUE = -1;
|
||||||
|
|
||||||
|
public static AppCollectorData newInstance(
|
||||||
|
ApplicationId id, String collectorAddr, long rmIdentifier, long version) {
|
||||||
|
AppCollectorData appCollectorData =
|
||||||
|
Records.newRecord(AppCollectorData.class);
|
||||||
|
appCollectorData.setApplicationId(id);
|
||||||
|
appCollectorData.setCollectorAddr(collectorAddr);
|
||||||
|
appCollectorData.setRMIdentifier(rmIdentifier);
|
||||||
|
appCollectorData.setVersion(version);
|
||||||
|
return appCollectorData;
|
||||||
|
}
|
||||||
|
|
||||||
|
public static AppCollectorData newInstance(ApplicationId id,
|
||||||
|
String collectorAddr) {
|
||||||
|
return newInstance(id, collectorAddr, DEFAULT_TIMESTAMP_VALUE,
|
||||||
|
DEFAULT_TIMESTAMP_VALUE);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns if a collector data item happens before another one. Null data
|
||||||
|
* items happens before any other non-null items. Non-null data items A
|
||||||
|
* happens before another non-null item B when A's rmIdentifier is less than
|
||||||
|
* B's rmIdentifier. Or A's version is less than B's if they have the same
|
||||||
|
* rmIdentifier.
|
||||||
|
*
|
||||||
|
* @param dataA first collector data item.
|
||||||
|
* @param dataB second collector data item.
|
||||||
|
* @return true if dataA happens before dataB.
|
||||||
|
*/
|
||||||
|
public static boolean happensBefore(AppCollectorData dataA,
|
||||||
|
AppCollectorData dataB) {
|
||||||
|
if (dataA == null && dataB == null) {
|
||||||
|
return false;
|
||||||
|
} else if (dataA == null || dataB == null) {
|
||||||
|
return dataA == null;
|
||||||
|
}
|
||||||
|
|
||||||
|
return
|
||||||
|
(dataA.getRMIdentifier() < dataB.getRMIdentifier())
|
||||||
|
|| ((dataA.getRMIdentifier() == dataB.getRMIdentifier())
|
||||||
|
&& (dataA.getVersion() < dataB.getVersion()));
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns if the collector data has been stamped by the RM with a RM cluster
|
||||||
|
* timestamp and a version number.
|
||||||
|
*
|
||||||
|
* @return true if RM has already assigned a timestamp for this collector.
|
||||||
|
* Otherwise, it means the RM has not recognized the existence of this
|
||||||
|
* collector.
|
||||||
|
*/
|
||||||
|
public boolean isStamped() {
|
||||||
|
return (getRMIdentifier() != DEFAULT_TIMESTAMP_VALUE)
|
||||||
|
|| (getVersion() != DEFAULT_TIMESTAMP_VALUE);
|
||||||
|
}
|
||||||
|
|
||||||
|
public abstract ApplicationId getApplicationId();
|
||||||
|
|
||||||
|
public abstract void setApplicationId(ApplicationId id);
|
||||||
|
|
||||||
|
public abstract String getCollectorAddr();
|
||||||
|
|
||||||
|
public abstract void setCollectorAddr(String addr);
|
||||||
|
|
||||||
|
public abstract long getRMIdentifier();
|
||||||
|
|
||||||
|
public abstract void setRMIdentifier(long rmId);
|
||||||
|
|
||||||
|
public abstract long getVersion();
|
||||||
|
|
||||||
|
public abstract void setVersion(long version);
|
||||||
|
|
||||||
|
}
|
@ -1,46 +0,0 @@
|
|||||||
/**
|
|
||||||
* Licensed to the Apache Software Foundation (ASF) under one
|
|
||||||
* or more contributor license agreements. See the NOTICE file
|
|
||||||
* distributed with this work for additional information
|
|
||||||
* regarding copyright ownership. The ASF licenses this file
|
|
||||||
* to you under the Apache License, Version 2.0 (the
|
|
||||||
* "License"); you may not use this file except in compliance
|
|
||||||
* with the License. You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package org.apache.hadoop.yarn.server.api.records;
|
|
||||||
|
|
||||||
import org.apache.hadoop.classification.InterfaceAudience.Private;
|
|
||||||
import org.apache.hadoop.yarn.api.records.ApplicationId;
|
|
||||||
import org.apache.hadoop.yarn.util.Records;
|
|
||||||
|
|
||||||
|
|
||||||
@Private
|
|
||||||
public abstract class AppCollectorsMap {
|
|
||||||
|
|
||||||
public static AppCollectorsMap newInstance(
|
|
||||||
ApplicationId id, String collectorAddr) {
|
|
||||||
AppCollectorsMap appCollectorsMap =
|
|
||||||
Records.newRecord(AppCollectorsMap.class);
|
|
||||||
appCollectorsMap.setApplicationId(id);
|
|
||||||
appCollectorsMap.setCollectorAddr(collectorAddr);
|
|
||||||
return appCollectorsMap;
|
|
||||||
}
|
|
||||||
|
|
||||||
public abstract ApplicationId getApplicationId();
|
|
||||||
|
|
||||||
public abstract void setApplicationId(ApplicationId id);
|
|
||||||
|
|
||||||
public abstract String getCollectorAddr();
|
|
||||||
|
|
||||||
public abstract void setCollectorAddr(String addr);
|
|
||||||
|
|
||||||
}
|
|
@ -21,37 +21,39 @@
|
|||||||
import org.apache.hadoop.classification.InterfaceStability.Unstable;
|
import org.apache.hadoop.classification.InterfaceStability.Unstable;
|
||||||
import org.apache.hadoop.yarn.api.records.ApplicationId;
|
import org.apache.hadoop.yarn.api.records.ApplicationId;
|
||||||
import org.apache.hadoop.yarn.api.records.impl.pb.ApplicationIdPBImpl;
|
import org.apache.hadoop.yarn.api.records.impl.pb.ApplicationIdPBImpl;
|
||||||
import org.apache.hadoop.yarn.server.api.records.AppCollectorsMap;
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
|
|
||||||
import org.apache.hadoop.yarn.proto.YarnProtos.ApplicationIdProto;
|
import org.apache.hadoop.yarn.proto.YarnProtos.ApplicationIdProto;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.AppCollectorsMapProto;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.AppCollectorDataProto;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.AppCollectorsMapProtoOrBuilder;
|
import org.apache.hadoop.yarn.proto.YarnServerCommonServiceProtos.AppCollectorDataProtoOrBuilder;
|
||||||
|
|
||||||
import com.google.protobuf.TextFormat;
|
import com.google.protobuf.TextFormat;
|
||||||
|
|
||||||
@Private
|
@Private
|
||||||
@Unstable
|
@Unstable
|
||||||
public class AppCollectorsMapPBImpl extends AppCollectorsMap {
|
public class AppCollectorDataPBImpl extends AppCollectorData {
|
||||||
|
|
||||||
private AppCollectorsMapProto proto =
|
private AppCollectorDataProto proto =
|
||||||
AppCollectorsMapProto.getDefaultInstance();
|
AppCollectorDataProto.getDefaultInstance();
|
||||||
|
|
||||||
private AppCollectorsMapProto.Builder builder = null;
|
private AppCollectorDataProto.Builder builder = null;
|
||||||
private boolean viaProto = false;
|
private boolean viaProto = false;
|
||||||
|
|
||||||
private ApplicationId appId = null;
|
private ApplicationId appId = null;
|
||||||
private String collectorAddr = null;
|
private String collectorAddr = null;
|
||||||
|
private Long rmIdentifier = null;
|
||||||
|
private Long version = null;
|
||||||
|
|
||||||
public AppCollectorsMapPBImpl() {
|
public AppCollectorDataPBImpl() {
|
||||||
builder = AppCollectorsMapProto.newBuilder();
|
builder = AppCollectorDataProto.newBuilder();
|
||||||
}
|
}
|
||||||
|
|
||||||
public AppCollectorsMapPBImpl(AppCollectorsMapProto proto) {
|
public AppCollectorDataPBImpl(AppCollectorDataProto proto) {
|
||||||
this.proto = proto;
|
this.proto = proto;
|
||||||
viaProto = true;
|
viaProto = true;
|
||||||
}
|
}
|
||||||
|
|
||||||
public AppCollectorsMapProto getProto() {
|
public AppCollectorDataProto getProto() {
|
||||||
mergeLocalToProto();
|
mergeLocalToProto();
|
||||||
proto = viaProto ? proto : builder.build();
|
proto = viaProto ? proto : builder.build();
|
||||||
viaProto = true;
|
viaProto = true;
|
||||||
@ -81,7 +83,7 @@ public String toString() {
|
|||||||
|
|
||||||
@Override
|
@Override
|
||||||
public ApplicationId getApplicationId() {
|
public ApplicationId getApplicationId() {
|
||||||
AppCollectorsMapProtoOrBuilder p = viaProto ? proto : builder;
|
AppCollectorDataProtoOrBuilder p = viaProto ? proto : builder;
|
||||||
if (this.appId == null && p.hasAppId()) {
|
if (this.appId == null && p.hasAppId()) {
|
||||||
this.appId = convertFromProtoFormat(p.getAppId());
|
this.appId = convertFromProtoFormat(p.getAppId());
|
||||||
}
|
}
|
||||||
@ -90,7 +92,7 @@ public ApplicationId getApplicationId() {
|
|||||||
|
|
||||||
@Override
|
@Override
|
||||||
public String getCollectorAddr() {
|
public String getCollectorAddr() {
|
||||||
AppCollectorsMapProtoOrBuilder p = viaProto ? proto : builder;
|
AppCollectorDataProtoOrBuilder p = viaProto ? proto : builder;
|
||||||
if (this.collectorAddr == null
|
if (this.collectorAddr == null
|
||||||
&& p.hasAppCollectorAddr()) {
|
&& p.hasAppCollectorAddr()) {
|
||||||
this.collectorAddr = p.getAppCollectorAddr();
|
this.collectorAddr = p.getAppCollectorAddr();
|
||||||
@ -116,6 +118,46 @@ public void setCollectorAddr(String collectorAddr) {
|
|||||||
this.collectorAddr = collectorAddr;
|
this.collectorAddr = collectorAddr;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public long getRMIdentifier() {
|
||||||
|
AppCollectorDataProtoOrBuilder p = viaProto ? proto : builder;
|
||||||
|
if (this.rmIdentifier == null && p.hasRmIdentifier()) {
|
||||||
|
this.rmIdentifier = p.getRmIdentifier();
|
||||||
|
}
|
||||||
|
if (this.rmIdentifier != null) {
|
||||||
|
return this.rmIdentifier;
|
||||||
|
} else {
|
||||||
|
return AppCollectorData.DEFAULT_TIMESTAMP_VALUE;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void setRMIdentifier(long rmId) {
|
||||||
|
maybeInitBuilder();
|
||||||
|
this.rmIdentifier = rmId;
|
||||||
|
builder.setRmIdentifier(rmId);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public long getVersion() {
|
||||||
|
AppCollectorDataProtoOrBuilder p = viaProto ? proto : builder;
|
||||||
|
if (this.version == null && p.hasRmIdentifier()) {
|
||||||
|
this.version = p.getRmIdentifier();
|
||||||
|
}
|
||||||
|
if (this.version != null) {
|
||||||
|
return this.version;
|
||||||
|
} else {
|
||||||
|
return AppCollectorData.DEFAULT_TIMESTAMP_VALUE;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void setVersion(long version) {
|
||||||
|
maybeInitBuilder();
|
||||||
|
this.version = version;
|
||||||
|
builder.setVersion(version);
|
||||||
|
}
|
||||||
|
|
||||||
private ApplicationIdPBImpl convertFromProtoFormat(ApplicationIdProto p) {
|
private ApplicationIdPBImpl convertFromProtoFormat(ApplicationIdProto p) {
|
||||||
return new ApplicationIdPBImpl(p);
|
return new ApplicationIdPBImpl(p);
|
||||||
}
|
}
|
||||||
@ -126,7 +168,7 @@ private ApplicationIdProto convertToProtoFormat(ApplicationId t) {
|
|||||||
|
|
||||||
private void maybeInitBuilder() {
|
private void maybeInitBuilder() {
|
||||||
if (viaProto || builder == null) {
|
if (viaProto || builder == null) {
|
||||||
builder = AppCollectorsMapProto.newBuilder(proto);
|
builder = AppCollectorDataProto.newBuilder(proto);
|
||||||
}
|
}
|
||||||
viaProto = false;
|
viaProto = false;
|
||||||
}
|
}
|
||||||
@ -147,6 +189,12 @@ private void mergeLocalToBuilder() {
|
|||||||
if (this.collectorAddr != null) {
|
if (this.collectorAddr != null) {
|
||||||
builder.setAppCollectorAddr(this.collectorAddr);
|
builder.setAppCollectorAddr(this.collectorAddr);
|
||||||
}
|
}
|
||||||
|
if (this.rmIdentifier != null) {
|
||||||
|
builder.setRmIdentifier(this.rmIdentifier);
|
||||||
|
}
|
||||||
|
if (this.version != null) {
|
||||||
|
builder.setVersion(this.version);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
@ -0,0 +1,19 @@
|
|||||||
|
/*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
/** Server records PB implementations. */
|
||||||
|
package org.apache.hadoop.yarn.server.api.records.impl.pb;
|
@ -90,7 +90,7 @@ message NodeHeartbeatRequestProto {
|
|||||||
optional MasterKeyProto last_known_nm_token_master_key = 3;
|
optional MasterKeyProto last_known_nm_token_master_key = 3;
|
||||||
optional NodeLabelsProto nodeLabels = 4;
|
optional NodeLabelsProto nodeLabels = 4;
|
||||||
repeated LogAggregationReportProto log_aggregation_reports_for_apps = 5;
|
repeated LogAggregationReportProto log_aggregation_reports_for_apps = 5;
|
||||||
repeated AppCollectorsMapProto registered_collectors = 6;
|
repeated AppCollectorDataProto registering_collectors = 6;
|
||||||
}
|
}
|
||||||
|
|
||||||
message LogAggregationReportProto {
|
message LogAggregationReportProto {
|
||||||
@ -116,7 +116,7 @@ message NodeHeartbeatResponseProto {
|
|||||||
repeated SignalContainerRequestProto containers_to_signal = 13;
|
repeated SignalContainerRequestProto containers_to_signal = 13;
|
||||||
optional ResourceProto resource = 14;
|
optional ResourceProto resource = 14;
|
||||||
optional ContainerQueuingLimitProto container_queuing_limit = 15;
|
optional ContainerQueuingLimitProto container_queuing_limit = 15;
|
||||||
repeated AppCollectorsMapProto app_collectors_map = 16;
|
repeated AppCollectorDataProto app_collectors = 16;
|
||||||
// to be used in place of containers_to_decrease
|
// to be used in place of containers_to_decrease
|
||||||
repeated ContainerProto containers_to_update = 17;
|
repeated ContainerProto containers_to_update = 17;
|
||||||
}
|
}
|
||||||
@ -134,16 +134,18 @@ message SystemCredentialsForAppsProto {
|
|||||||
////////////////////////////////////////////////////////////////////////
|
////////////////////////////////////////////////////////////////////////
|
||||||
////// From collector_nodemanager_protocol ////////////////////////////
|
////// From collector_nodemanager_protocol ////////////////////////////
|
||||||
////////////////////////////////////////////////////////////////////////
|
////////////////////////////////////////////////////////////////////////
|
||||||
message AppCollectorsMapProto {
|
message AppCollectorDataProto {
|
||||||
optional ApplicationIdProto appId = 1;
|
optional ApplicationIdProto app_id = 1;
|
||||||
optional string appCollectorAddr = 2;
|
optional string app_collector_addr = 2;
|
||||||
|
optional int64 rm_identifier = 3 [default = -1];
|
||||||
|
optional int64 version = 4 [default = -1];
|
||||||
}
|
}
|
||||||
|
|
||||||
//////////////////////////////////////////////////////
|
//////////////////////////////////////////////////////
|
||||||
/////// collector_nodemanager_protocol //////////////
|
/////// collector_nodemanager_protocol //////////////
|
||||||
//////////////////////////////////////////////////////
|
//////////////////////////////////////////////////////
|
||||||
message ReportNewCollectorInfoRequestProto {
|
message ReportNewCollectorInfoRequestProto {
|
||||||
repeated AppCollectorsMapProto app_collectors = 1;
|
repeated AppCollectorDataProto app_collectors = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message ReportNewCollectorInfoResponseProto {
|
message ReportNewCollectorInfoResponseProto {
|
||||||
|
@ -77,7 +77,7 @@
|
|||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.GetTimelineCollectorContextResponse;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.GetTimelineCollectorContextResponse;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.ReportNewCollectorInfoRequest;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.ReportNewCollectorInfoRequest;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.ReportNewCollectorInfoResponse;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.ReportNewCollectorInfoResponse;
|
||||||
import org.apache.hadoop.yarn.server.api.records.AppCollectorsMap;
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.util.Records;
|
import org.apache.hadoop.yarn.util.Records;
|
||||||
import org.junit.Assert;
|
import org.junit.Assert;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
@ -429,10 +429,10 @@ public class DummyNMCollectorService
|
|||||||
public ReportNewCollectorInfoResponse reportNewCollectorInfo(
|
public ReportNewCollectorInfoResponse reportNewCollectorInfo(
|
||||||
ReportNewCollectorInfoRequest request)
|
ReportNewCollectorInfoRequest request)
|
||||||
throws YarnException, IOException {
|
throws YarnException, IOException {
|
||||||
List<AppCollectorsMap> appCollectors = request.getAppCollectorsList();
|
List<AppCollectorData> appCollectors = request.getAppCollectorsList();
|
||||||
if (appCollectors.size() == 1) {
|
if (appCollectors.size() == 1) {
|
||||||
// check default appID and collectorAddr
|
// check default appID and collectorAddr
|
||||||
AppCollectorsMap appCollector = appCollectors.get(0);
|
AppCollectorData appCollector = appCollectors.get(0);
|
||||||
Assert.assertEquals(appCollector.getApplicationId(),
|
Assert.assertEquals(appCollector.getApplicationId(),
|
||||||
DEFAULT_APP_ID);
|
DEFAULT_APP_ID);
|
||||||
Assert.assertEquals(appCollector.getCollectorAddr(),
|
Assert.assertEquals(appCollector.getCollectorAddr(),
|
||||||
|
@ -48,6 +48,7 @@
|
|||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.impl.pb.RegisterNodeManagerRequestPBImpl;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.impl.pb.RegisterNodeManagerRequestPBImpl;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.impl.pb.RegisterNodeManagerResponsePBImpl;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.impl.pb.RegisterNodeManagerResponsePBImpl;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.impl.pb.UnRegisterNodeManagerRequestPBImpl;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.impl.pb.UnRegisterNodeManagerRequestPBImpl;
|
||||||
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.api.records.MasterKey;
|
import org.apache.hadoop.yarn.server.api.records.MasterKey;
|
||||||
import org.apache.hadoop.yarn.server.api.records.NodeAction;
|
import org.apache.hadoop.yarn.server.api.records.NodeAction;
|
||||||
import org.apache.hadoop.yarn.server.api.records.NodeHealthStatus;
|
import org.apache.hadoop.yarn.server.api.records.NodeHealthStatus;
|
||||||
@ -109,14 +110,14 @@ public void testNodeHeartbeatRequestPBImpl() {
|
|||||||
original.setLastKnownNMTokenMasterKey(getMasterKey());
|
original.setLastKnownNMTokenMasterKey(getMasterKey());
|
||||||
original.setNodeStatus(getNodeStatus());
|
original.setNodeStatus(getNodeStatus());
|
||||||
original.setNodeLabels(getValidNodeLabels());
|
original.setNodeLabels(getValidNodeLabels());
|
||||||
Map<ApplicationId, String> collectors = getCollectors();
|
Map<ApplicationId, AppCollectorData> collectors = getCollectors();
|
||||||
original.setRegisteredCollectors(collectors);
|
original.setRegisteringCollectors(collectors);
|
||||||
NodeHeartbeatRequestPBImpl copy = new NodeHeartbeatRequestPBImpl(
|
NodeHeartbeatRequestPBImpl copy = new NodeHeartbeatRequestPBImpl(
|
||||||
original.getProto());
|
original.getProto());
|
||||||
assertEquals(1, copy.getLastKnownContainerTokenMasterKey().getKeyId());
|
assertEquals(1, copy.getLastKnownContainerTokenMasterKey().getKeyId());
|
||||||
assertEquals(1, copy.getLastKnownNMTokenMasterKey().getKeyId());
|
assertEquals(1, copy.getLastKnownNMTokenMasterKey().getKeyId());
|
||||||
assertEquals("localhost", copy.getNodeStatus().getNodeId().getHost());
|
assertEquals("localhost", copy.getNodeStatus().getNodeId().getHost());
|
||||||
assertEquals(collectors, copy.getRegisteredCollectors());
|
assertEquals(collectors, copy.getRegisteringCollectors());
|
||||||
// check labels are coming with valid values
|
// check labels are coming with valid values
|
||||||
Assert.assertTrue(original.getNodeLabels()
|
Assert.assertTrue(original.getNodeLabels()
|
||||||
.containsAll(copy.getNodeLabels()));
|
.containsAll(copy.getNodeLabels()));
|
||||||
@ -153,8 +154,8 @@ public void testNodeHeartbeatResponsePBImpl() {
|
|||||||
original.setNextHeartBeatInterval(1000);
|
original.setNextHeartBeatInterval(1000);
|
||||||
original.setNodeAction(NodeAction.NORMAL);
|
original.setNodeAction(NodeAction.NORMAL);
|
||||||
original.setResponseId(100);
|
original.setResponseId(100);
|
||||||
Map<ApplicationId, String> collectors = getCollectors();
|
Map<ApplicationId, AppCollectorData> collectors = getCollectors();
|
||||||
original.setAppCollectorsMap(collectors);
|
original.setAppCollectors(collectors);
|
||||||
|
|
||||||
NodeHeartbeatResponsePBImpl copy = new NodeHeartbeatResponsePBImpl(
|
NodeHeartbeatResponsePBImpl copy = new NodeHeartbeatResponsePBImpl(
|
||||||
original.getProto());
|
original.getProto());
|
||||||
@ -164,7 +165,7 @@ public void testNodeHeartbeatResponsePBImpl() {
|
|||||||
assertEquals(1, copy.getContainerTokenMasterKey().getKeyId());
|
assertEquals(1, copy.getContainerTokenMasterKey().getKeyId());
|
||||||
assertEquals(1, copy.getNMTokenMasterKey().getKeyId());
|
assertEquals(1, copy.getNMTokenMasterKey().getKeyId());
|
||||||
assertEquals("testDiagnosticMessage", copy.getDiagnosticsMessage());
|
assertEquals("testDiagnosticMessage", copy.getDiagnosticsMessage());
|
||||||
assertEquals(collectors, copy.getAppCollectorsMap());
|
assertEquals(collectors, copy.getAppCollectors());
|
||||||
assertEquals(false, copy.getAreNodeLabelsAcceptedByRM());
|
assertEquals(false, copy.getAreNodeLabelsAcceptedByRM());
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -347,12 +348,13 @@ private HashSet<NodeLabel> getValidNodeLabels() {
|
|||||||
return nodeLabels;
|
return nodeLabels;
|
||||||
}
|
}
|
||||||
|
|
||||||
private Map<ApplicationId, String> getCollectors() {
|
private Map<ApplicationId, AppCollectorData> getCollectors() {
|
||||||
ApplicationId appID = ApplicationId.newInstance(1L, 1);
|
ApplicationId appID = ApplicationId.newInstance(1L, 1);
|
||||||
String collectorAddr = "localhost:0";
|
String collectorAddr = "localhost:0";
|
||||||
Map<ApplicationId, String> collectorMap =
|
AppCollectorData data = AppCollectorData.newInstance(appID, collectorAddr);
|
||||||
new HashMap<ApplicationId, String>();
|
Map<ApplicationId, AppCollectorData> collectorMap =
|
||||||
collectorMap.put(appID, collectorAddr);
|
new HashMap<>();
|
||||||
|
collectorMap.put(appID, data);
|
||||||
return collectorMap;
|
return collectorMap;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -28,6 +28,7 @@
|
|||||||
import org.apache.hadoop.yarn.api.records.ContainerId;
|
import org.apache.hadoop.yarn.api.records.ContainerId;
|
||||||
import org.apache.hadoop.yarn.api.records.NodeId;
|
import org.apache.hadoop.yarn.api.records.NodeId;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.LogAggregationReport;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.LogAggregationReport;
|
||||||
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.api.records.NodeHealthStatus;
|
import org.apache.hadoop.yarn.server.api.records.NodeHealthStatus;
|
||||||
import org.apache.hadoop.yarn.server.nodemanager.containermanager.ContainerManager;
|
import org.apache.hadoop.yarn.server.nodemanager.containermanager.ContainerManager;
|
||||||
import org.apache.hadoop.yarn.server.nodemanager.containermanager.application.Application;
|
import org.apache.hadoop.yarn.server.nodemanager.containermanager.application.Application;
|
||||||
@ -65,11 +66,18 @@ public interface Context {
|
|||||||
Map<ApplicationId, Credentials> getSystemCredentialsForApps();
|
Map<ApplicationId, Credentials> getSystemCredentialsForApps();
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Get the registered collectors that located on this NM.
|
* Get the list of collectors that are registering with the RM from this node.
|
||||||
* @return registered collectors, or null if the timeline service v.2 is not
|
* @return registering collectors, or null if the timeline service v.2 is not
|
||||||
* enabled
|
* enabled
|
||||||
*/
|
*/
|
||||||
Map<ApplicationId, String> getRegisteredCollectors();
|
Map<ApplicationId, AppCollectorData> getRegisteringCollectors();
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get the list of collectors registered with the RM and known by this node.
|
||||||
|
* @return known collectors, or null if the timeline service v.2 is not
|
||||||
|
* enabled.
|
||||||
|
*/
|
||||||
|
Map<ApplicationId, AppCollectorData> getKnownCollectors();
|
||||||
|
|
||||||
ConcurrentMap<ContainerId, Container> getContainers();
|
ConcurrentMap<ContainerId, Container> getContainers();
|
||||||
|
|
||||||
|
@ -57,6 +57,7 @@
|
|||||||
import org.apache.hadoop.yarn.exceptions.YarnRuntimeException;
|
import org.apache.hadoop.yarn.exceptions.YarnRuntimeException;
|
||||||
import org.apache.hadoop.yarn.factory.providers.RecordFactoryProvider;
|
import org.apache.hadoop.yarn.factory.providers.RecordFactoryProvider;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.LogAggregationReport;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.LogAggregationReport;
|
||||||
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.api.records.NodeHealthStatus;
|
import org.apache.hadoop.yarn.server.api.records.NodeHealthStatus;
|
||||||
import org.apache.hadoop.yarn.server.nodemanager.containermanager.ContainerManager;
|
import org.apache.hadoop.yarn.server.nodemanager.containermanager.ContainerManager;
|
||||||
import org.apache.hadoop.yarn.server.nodemanager.collectormanager.NMCollectorService;
|
import org.apache.hadoop.yarn.server.nodemanager.collectormanager.NMCollectorService;
|
||||||
@ -492,7 +493,9 @@ public static class NMContext implements Context {
|
|||||||
protected final ConcurrentMap<ContainerId, Container> containers =
|
protected final ConcurrentMap<ContainerId, Container> containers =
|
||||||
new ConcurrentSkipListMap<ContainerId, Container>();
|
new ConcurrentSkipListMap<ContainerId, Container>();
|
||||||
|
|
||||||
private Map<ApplicationId, String> registeredCollectors;
|
private Map<ApplicationId, AppCollectorData> registeringCollectors;
|
||||||
|
|
||||||
|
private Map<ApplicationId, AppCollectorData> knownCollectors;
|
||||||
|
|
||||||
protected final ConcurrentMap<ContainerId,
|
protected final ConcurrentMap<ContainerId,
|
||||||
org.apache.hadoop.yarn.api.records.Container> increasedContainers =
|
org.apache.hadoop.yarn.api.records.Container> increasedContainers =
|
||||||
@ -526,7 +529,8 @@ public NMContext(NMContainerTokenSecretManager containerTokenSecretManager,
|
|||||||
NMStateStoreService stateStore, boolean isDistSchedulingEnabled,
|
NMStateStoreService stateStore, boolean isDistSchedulingEnabled,
|
||||||
Configuration conf) {
|
Configuration conf) {
|
||||||
if (YarnConfiguration.timelineServiceV2Enabled(conf)) {
|
if (YarnConfiguration.timelineServiceV2Enabled(conf)) {
|
||||||
this.registeredCollectors = new ConcurrentHashMap<>();
|
this.registeringCollectors = new ConcurrentHashMap<>();
|
||||||
|
this.knownCollectors = new ConcurrentHashMap<>();
|
||||||
}
|
}
|
||||||
this.containerTokenSecretManager = containerTokenSecretManager;
|
this.containerTokenSecretManager = containerTokenSecretManager;
|
||||||
this.nmTokenSecretManager = nmTokenSecretManager;
|
this.nmTokenSecretManager = nmTokenSecretManager;
|
||||||
@ -681,18 +685,13 @@ public OpportunisticContainerAllocator getContainerAllocator() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public Map<ApplicationId, String> getRegisteredCollectors() {
|
public Map<ApplicationId, AppCollectorData> getRegisteringCollectors() {
|
||||||
return this.registeredCollectors;
|
return this.registeringCollectors;
|
||||||
}
|
}
|
||||||
|
|
||||||
public void addRegisteredCollectors(
|
@Override
|
||||||
Map<ApplicationId, String> newRegisteredCollectors) {
|
public Map<ApplicationId, AppCollectorData> getKnownCollectors() {
|
||||||
if (registeredCollectors != null) {
|
return this.knownCollectors;
|
||||||
this.registeredCollectors.putAll(newRegisteredCollectors);
|
|
||||||
} else {
|
|
||||||
LOG.warn("collectors are added when the registered collectors are " +
|
|
||||||
"initialized");
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -72,6 +72,7 @@
|
|||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.RegisterNodeManagerResponse;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.RegisterNodeManagerResponse;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.UnRegisterNodeManagerRequest;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.UnRegisterNodeManagerRequest;
|
||||||
|
|
||||||
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.api.records.ContainerQueuingLimit;
|
import org.apache.hadoop.yarn.server.api.records.ContainerQueuingLimit;
|
||||||
import org.apache.hadoop.yarn.server.api.records.OpportunisticContainersStatus;
|
import org.apache.hadoop.yarn.server.api.records.OpportunisticContainersStatus;
|
||||||
import org.apache.hadoop.yarn.server.api.records.MasterKey;
|
import org.apache.hadoop.yarn.server.api.records.MasterKey;
|
||||||
@ -760,7 +761,6 @@ private static Map<ApplicationId, Credentials> parseCredentials(
|
|||||||
}
|
}
|
||||||
|
|
||||||
protected void startStatusUpdater() {
|
protected void startStatusUpdater() {
|
||||||
|
|
||||||
statusUpdaterRunnable = new StatusUpdaterRunnable();
|
statusUpdaterRunnable = new StatusUpdaterRunnable();
|
||||||
statusUpdater =
|
statusUpdater =
|
||||||
new Thread(statusUpdaterRunnable, "Node Status Updater");
|
new Thread(statusUpdaterRunnable, "Node Status Updater");
|
||||||
@ -1043,7 +1043,7 @@ public void run() {
|
|||||||
.getNMTokenSecretManager().getCurrentKey(),
|
.getNMTokenSecretManager().getCurrentKey(),
|
||||||
nodeLabelsForHeartbeat,
|
nodeLabelsForHeartbeat,
|
||||||
NodeStatusUpdaterImpl.this.context
|
NodeStatusUpdaterImpl.this.context
|
||||||
.getRegisteredCollectors());
|
.getRegisteringCollectors());
|
||||||
|
|
||||||
if (logAggregationEnabled) {
|
if (logAggregationEnabled) {
|
||||||
// pull log aggregation status for application running in this NM
|
// pull log aggregation status for application running in this NM
|
||||||
@ -1134,7 +1134,7 @@ public void run() {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (YarnConfiguration.timelineServiceV2Enabled(context.getConf())) {
|
if (YarnConfiguration.timelineServiceV2Enabled(context.getConf())) {
|
||||||
updateTimelineClientsAddress(response);
|
updateTimelineCollectorData(response);
|
||||||
}
|
}
|
||||||
|
|
||||||
} catch (ConnectException e) {
|
} catch (ConnectException e) {
|
||||||
@ -1164,40 +1164,48 @@ public void run() {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
private void updateTimelineClientsAddress(
|
private void updateTimelineCollectorData(
|
||||||
NodeHeartbeatResponse response) {
|
NodeHeartbeatResponse response) {
|
||||||
Map<ApplicationId, String> knownCollectorsMap =
|
Map<ApplicationId, AppCollectorData> incomingCollectorsMap =
|
||||||
response.getAppCollectorsMap();
|
response.getAppCollectors();
|
||||||
if (knownCollectorsMap == null) {
|
if (incomingCollectorsMap == null) {
|
||||||
if (LOG.isDebugEnabled()) {
|
if (LOG.isDebugEnabled()) {
|
||||||
LOG.debug("No collectors to update RM");
|
LOG.debug("No collectors to update RM");
|
||||||
}
|
}
|
||||||
} else {
|
return;
|
||||||
Set<Map.Entry<ApplicationId, String>> rmKnownCollectors =
|
}
|
||||||
knownCollectorsMap.entrySet();
|
Map<ApplicationId, AppCollectorData> knownCollectors =
|
||||||
for (Map.Entry<ApplicationId, String> entry : rmKnownCollectors) {
|
context.getKnownCollectors();
|
||||||
ApplicationId appId = entry.getKey();
|
for (Map.Entry<ApplicationId, AppCollectorData> entry
|
||||||
String collectorAddr = entry.getValue();
|
: incomingCollectorsMap.entrySet()) {
|
||||||
|
ApplicationId appId = entry.getKey();
|
||||||
|
AppCollectorData collectorData = entry.getValue();
|
||||||
|
|
||||||
// Only handle applications running on local node.
|
// Only handle applications running on local node.
|
||||||
// Not include apps with timeline collectors running in local
|
Application application = context.getApplications().get(appId);
|
||||||
Application application = context.getApplications().get(appId);
|
if (application != null) {
|
||||||
// TODO this logic could be problematic if the collector address
|
// Update collector data if the newly received data happens after
|
||||||
// gets updated due to NM restart or collector service failure
|
// the known data (updates the known data).
|
||||||
if (application != null &&
|
AppCollectorData existingData = knownCollectors.get(appId);
|
||||||
!context.getRegisteredCollectors().containsKey(appId)) {
|
if (AppCollectorData.happensBefore(existingData, collectorData)) {
|
||||||
if (LOG.isDebugEnabled()) {
|
if (LOG.isDebugEnabled()) {
|
||||||
LOG.debug("Sync a new collector address: " + collectorAddr +
|
LOG.debug("Sync a new collector address: "
|
||||||
" for application: " + appId + " from RM.");
|
+ collectorData.getCollectorAddr()
|
||||||
|
+ " for application: " + appId + " from RM.");
|
||||||
}
|
}
|
||||||
|
// Update information for clients.
|
||||||
NMTimelinePublisher nmTimelinePublisher =
|
NMTimelinePublisher nmTimelinePublisher =
|
||||||
context.getNMTimelinePublisher();
|
context.getNMTimelinePublisher();
|
||||||
if (nmTimelinePublisher != null) {
|
if (nmTimelinePublisher != null) {
|
||||||
nmTimelinePublisher.setTimelineServiceAddress(
|
nmTimelinePublisher.setTimelineServiceAddress(
|
||||||
application.getAppId(), collectorAddr);
|
application.getAppId(), collectorData.getCollectorAddr());
|
||||||
}
|
}
|
||||||
|
// Update information for the node manager itself.
|
||||||
|
knownCollectors.put(appId, collectorData);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
// Remove the registering collector data
|
||||||
|
context.getRegisteringCollectors().remove(entry.getKey());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -37,9 +37,8 @@
|
|||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.GetTimelineCollectorContextResponse;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.GetTimelineCollectorContextResponse;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.ReportNewCollectorInfoRequest;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.ReportNewCollectorInfoRequest;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.ReportNewCollectorInfoResponse;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.ReportNewCollectorInfoResponse;
|
||||||
import org.apache.hadoop.yarn.server.api.records.AppCollectorsMap;
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.nodemanager.Context;
|
import org.apache.hadoop.yarn.server.nodemanager.Context;
|
||||||
import org.apache.hadoop.yarn.server.nodemanager.NodeManager;
|
|
||||||
import org.apache.hadoop.yarn.server.nodemanager.containermanager.application.Application;
|
import org.apache.hadoop.yarn.server.nodemanager.containermanager.application.Application;
|
||||||
import org.apache.hadoop.yarn.server.nodemanager.timelineservice.NMTimelinePublisher;
|
import org.apache.hadoop.yarn.server.nodemanager.timelineservice.NMTimelinePublisher;
|
||||||
|
|
||||||
@ -108,23 +107,31 @@ public void serviceStop() throws Exception {
|
|||||||
@Override
|
@Override
|
||||||
public ReportNewCollectorInfoResponse reportNewCollectorInfo(
|
public ReportNewCollectorInfoResponse reportNewCollectorInfo(
|
||||||
ReportNewCollectorInfoRequest request) throws YarnException, IOException {
|
ReportNewCollectorInfoRequest request) throws YarnException, IOException {
|
||||||
List<AppCollectorsMap> newCollectorsList = request.getAppCollectorsList();
|
List<AppCollectorData> newCollectorsList = request.getAppCollectorsList();
|
||||||
if (newCollectorsList != null && !newCollectorsList.isEmpty()) {
|
if (newCollectorsList != null && !newCollectorsList.isEmpty()) {
|
||||||
Map<ApplicationId, String> newCollectorsMap =
|
Map<ApplicationId, AppCollectorData> newCollectorsMap =
|
||||||
new HashMap<ApplicationId, String>();
|
new HashMap<>();
|
||||||
for (AppCollectorsMap collector : newCollectorsList) {
|
for (AppCollectorData collector : newCollectorsList) {
|
||||||
ApplicationId appId = collector.getApplicationId();
|
ApplicationId appId = collector.getApplicationId();
|
||||||
String collectorAddr = collector.getCollectorAddr();
|
newCollectorsMap.put(appId, collector);
|
||||||
newCollectorsMap.put(appId, collectorAddr);
|
|
||||||
// set registered collector address to TimelineClient.
|
// set registered collector address to TimelineClient.
|
||||||
|
// TODO: Do we need to do this after we received confirmation from
|
||||||
|
// the RM?
|
||||||
NMTimelinePublisher nmTimelinePublisher =
|
NMTimelinePublisher nmTimelinePublisher =
|
||||||
context.getNMTimelinePublisher();
|
context.getNMTimelinePublisher();
|
||||||
if (nmTimelinePublisher != null) {
|
if (nmTimelinePublisher != null) {
|
||||||
nmTimelinePublisher.setTimelineServiceAddress(appId, collectorAddr);
|
nmTimelinePublisher.setTimelineServiceAddress(appId,
|
||||||
|
collector.getCollectorAddr());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
((NodeManager.NMContext)context).addRegisteredCollectors(
|
Map<ApplicationId, AppCollectorData> registeringCollectors
|
||||||
newCollectorsMap);
|
= context.getRegisteringCollectors();
|
||||||
|
if (registeringCollectors != null) {
|
||||||
|
registeringCollectors.putAll(newCollectorsMap);
|
||||||
|
} else {
|
||||||
|
LOG.warn("collectors are added when the registered collectors are " +
|
||||||
|
"initialized");
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return ReportNewCollectorInfoResponse.newInstance();
|
return ReportNewCollectorInfoResponse.newInstance();
|
||||||
|
@ -45,6 +45,7 @@
|
|||||||
import org.apache.hadoop.yarn.proto.YarnProtos;
|
import org.apache.hadoop.yarn.proto.YarnProtos;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerNodemanagerRecoveryProtos.ContainerManagerApplicationProto;
|
import org.apache.hadoop.yarn.proto.YarnServerNodemanagerRecoveryProtos.ContainerManagerApplicationProto;
|
||||||
import org.apache.hadoop.yarn.proto.YarnServerNodemanagerRecoveryProtos.FlowContextProto;
|
import org.apache.hadoop.yarn.proto.YarnServerNodemanagerRecoveryProtos.FlowContextProto;
|
||||||
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.nodemanager.Context;
|
import org.apache.hadoop.yarn.server.nodemanager.Context;
|
||||||
import org.apache.hadoop.yarn.server.nodemanager.containermanager.AuxServicesEvent;
|
import org.apache.hadoop.yarn.server.nodemanager.containermanager.AuxServicesEvent;
|
||||||
import org.apache.hadoop.yarn.server.nodemanager.containermanager.AuxServicesEventType;
|
import org.apache.hadoop.yarn.server.nodemanager.containermanager.AuxServicesEventType;
|
||||||
@ -558,6 +559,29 @@ public ApplicationState transition(ApplicationImpl app,
|
|||||||
@SuppressWarnings("unchecked")
|
@SuppressWarnings("unchecked")
|
||||||
static class AppCompletelyDoneTransition implements
|
static class AppCompletelyDoneTransition implements
|
||||||
SingleArcTransition<ApplicationImpl, ApplicationEvent> {
|
SingleArcTransition<ApplicationImpl, ApplicationEvent> {
|
||||||
|
|
||||||
|
private void updateCollectorStatus(ApplicationImpl app) {
|
||||||
|
// Remove collectors info for finished apps.
|
||||||
|
// TODO check we remove related collectors info in failure cases
|
||||||
|
// (YARN-3038)
|
||||||
|
Map<ApplicationId, AppCollectorData> registeringCollectors
|
||||||
|
= app.context.getRegisteringCollectors();
|
||||||
|
if (registeringCollectors != null) {
|
||||||
|
registeringCollectors.remove(app.getAppId());
|
||||||
|
}
|
||||||
|
Map<ApplicationId, AppCollectorData> knownCollectors =
|
||||||
|
app.context.getKnownCollectors();
|
||||||
|
if (knownCollectors != null) {
|
||||||
|
knownCollectors.remove(app.getAppId());
|
||||||
|
}
|
||||||
|
// stop timelineClient when application get finished.
|
||||||
|
NMTimelinePublisher nmTimelinePublisher =
|
||||||
|
app.context.getNMTimelinePublisher();
|
||||||
|
if (nmTimelinePublisher != null) {
|
||||||
|
nmTimelinePublisher.stopTimelineClient(app.getAppId());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void transition(ApplicationImpl app, ApplicationEvent event) {
|
public void transition(ApplicationImpl app, ApplicationEvent event) {
|
||||||
|
|
||||||
@ -566,20 +590,7 @@ public void transition(ApplicationImpl app, ApplicationEvent event) {
|
|||||||
new LogHandlerAppFinishedEvent(app.appId));
|
new LogHandlerAppFinishedEvent(app.appId));
|
||||||
|
|
||||||
app.context.getNMTokenSecretManager().appFinished(app.getAppId());
|
app.context.getNMTokenSecretManager().appFinished(app.getAppId());
|
||||||
// Remove collectors info for finished apps.
|
updateCollectorStatus(app);
|
||||||
// TODO check we remove related collectors info in failure cases
|
|
||||||
// (YARN-3038)
|
|
||||||
Map<ApplicationId, String> registeredCollectors =
|
|
||||||
app.context.getRegisteredCollectors();
|
|
||||||
if (registeredCollectors != null) {
|
|
||||||
registeredCollectors.remove(app.getAppId());
|
|
||||||
}
|
|
||||||
// stop timelineClient when application get finished.
|
|
||||||
NMTimelinePublisher nmTimelinePublisher =
|
|
||||||
app.context.getNMTimelinePublisher();
|
|
||||||
if (nmTimelinePublisher != null) {
|
|
||||||
nmTimelinePublisher.stopTimelineClient(app.getAppId());
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -59,6 +59,7 @@
|
|||||||
import org.apache.hadoop.yarn.exceptions.YarnException;
|
import org.apache.hadoop.yarn.exceptions.YarnException;
|
||||||
import org.apache.hadoop.yarn.security.AMRMTokenIdentifier;
|
import org.apache.hadoop.yarn.security.AMRMTokenIdentifier;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.LogAggregationReport;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.LogAggregationReport;
|
||||||
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.api.records.NodeHealthStatus;
|
import org.apache.hadoop.yarn.server.api.records.NodeHealthStatus;
|
||||||
import org.apache.hadoop.yarn.server.nodemanager.ContainerExecutor;
|
import org.apache.hadoop.yarn.server.nodemanager.ContainerExecutor;
|
||||||
import org.apache.hadoop.yarn.server.nodemanager.Context;
|
import org.apache.hadoop.yarn.server.nodemanager.Context;
|
||||||
@ -658,8 +659,11 @@ public Map<ApplicationId, Credentials> getSystemCredentialsForApps() {
|
|||||||
return null;
|
return null;
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
public Map<ApplicationId, AppCollectorData> getRegisteringCollectors() {
|
||||||
public Map<ApplicationId, String> getRegisteredCollectors() {
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override public Map<ApplicationId, AppCollectorData> getKnownCollectors() {
|
||||||
return null;
|
return null;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -61,6 +61,7 @@
|
|||||||
import org.apache.hadoop.yarn.security.AMRMTokenIdentifier;
|
import org.apache.hadoop.yarn.security.AMRMTokenIdentifier;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.RMAuditLogger.AuditConstants;
|
import org.apache.hadoop.yarn.server.resourcemanager.RMAuditLogger.AuditConstants;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMApp;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMApp;
|
||||||
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppImpl;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.AMLivelinessMonitor;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.AMLivelinessMonitor;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttempt;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttempt;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptImpl;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptImpl;
|
||||||
@ -254,7 +255,7 @@ public FinishApplicationMasterResponse finishApplicationMaster(
|
|||||||
|
|
||||||
// Remove collector address when app get finished.
|
// Remove collector address when app get finished.
|
||||||
if (YarnConfiguration.timelineServiceV2Enabled(getConfig())) {
|
if (YarnConfiguration.timelineServiceV2Enabled(getConfig())) {
|
||||||
rmApp.removeCollectorAddr();
|
((RMAppImpl) rmApp).removeCollectorData();
|
||||||
}
|
}
|
||||||
// checking whether the app exits in RMStateStore at first not to throw
|
// checking whether the app exits in RMStateStore at first not to throw
|
||||||
// ApplicationDoesNotExistInCacheException before and after
|
// ApplicationDoesNotExistInCacheException before and after
|
||||||
|
@ -54,6 +54,7 @@
|
|||||||
import org.apache.hadoop.yarn.exceptions.YarnException;
|
import org.apache.hadoop.yarn.exceptions.YarnException;
|
||||||
import org.apache.hadoop.yarn.factories.RecordFactory;
|
import org.apache.hadoop.yarn.factories.RecordFactory;
|
||||||
import org.apache.hadoop.yarn.factory.providers.RecordFactoryProvider;
|
import org.apache.hadoop.yarn.factory.providers.RecordFactoryProvider;
|
||||||
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMApp;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMApp;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttempt;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttempt;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt
|
||||||
@ -293,9 +294,10 @@ public void allocate(ApplicationAttemptId appAttemptId,
|
|||||||
// add collector address for this application
|
// add collector address for this application
|
||||||
if (YarnConfiguration.timelineServiceV2Enabled(
|
if (YarnConfiguration.timelineServiceV2Enabled(
|
||||||
getRmContext().getYarnConfiguration())) {
|
getRmContext().getYarnConfiguration())) {
|
||||||
response.setCollectorAddr(
|
AppCollectorData data = app.getCollectorData();
|
||||||
getRmContext().getRMApps().get(appAttemptId.getApplicationId())
|
if (data != null) {
|
||||||
.getCollectorAddr());
|
response.setCollectorAddr(data.getCollectorAddr());
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// add preemption to the allocateResponse message (if any)
|
// add preemption to the allocateResponse message (if any)
|
||||||
|
@ -27,6 +27,7 @@
|
|||||||
import java.util.Map;
|
import java.util.Map;
|
||||||
import java.util.Set;
|
import java.util.Set;
|
||||||
import java.util.concurrent.ConcurrentMap;
|
import java.util.concurrent.ConcurrentMap;
|
||||||
|
import java.util.concurrent.atomic.AtomicLong;
|
||||||
import java.util.concurrent.locks.ReentrantReadWriteLock;
|
import java.util.concurrent.locks.ReentrantReadWriteLock;
|
||||||
import java.util.concurrent.locks.ReentrantReadWriteLock.ReadLock;
|
import java.util.concurrent.locks.ReentrantReadWriteLock.ReadLock;
|
||||||
import java.util.concurrent.locks.ReentrantReadWriteLock.WriteLock;
|
import java.util.concurrent.locks.ReentrantReadWriteLock.WriteLock;
|
||||||
@ -63,12 +64,14 @@
|
|||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.RegisterNodeManagerResponse;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.RegisterNodeManagerResponse;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.UnRegisterNodeManagerRequest;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.UnRegisterNodeManagerRequest;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.UnRegisterNodeManagerResponse;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.UnRegisterNodeManagerResponse;
|
||||||
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.api.records.MasterKey;
|
import org.apache.hadoop.yarn.server.api.records.MasterKey;
|
||||||
import org.apache.hadoop.yarn.server.api.records.NodeAction;
|
import org.apache.hadoop.yarn.server.api.records.NodeAction;
|
||||||
import org.apache.hadoop.yarn.server.api.records.NodeStatus;
|
import org.apache.hadoop.yarn.server.api.records.NodeStatus;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.nodelabels.NodeLabelsUtils;
|
import org.apache.hadoop.yarn.server.resourcemanager.nodelabels.NodeLabelsUtils;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.resource.DynamicResourceConfiguration;
|
import org.apache.hadoop.yarn.server.resourcemanager.resource.DynamicResourceConfiguration;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMApp;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMApp;
|
||||||
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppImpl;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttempt;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttempt;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.event.RMAppAttemptContainerFinishedEvent;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.event.RMAppAttemptContainerFinishedEvent;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNode;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNode;
|
||||||
@ -118,6 +121,8 @@ public class ResourceTrackerService extends AbstractService implements
|
|||||||
private boolean isDelegatedCentralizedNodeLabelsConf;
|
private boolean isDelegatedCentralizedNodeLabelsConf;
|
||||||
private DynamicResourceConfiguration drConf;
|
private DynamicResourceConfiguration drConf;
|
||||||
|
|
||||||
|
private final AtomicLong timelineCollectorVersion = new AtomicLong(0);
|
||||||
|
|
||||||
public ResourceTrackerService(RMContext rmContext,
|
public ResourceTrackerService(RMContext rmContext,
|
||||||
NodesListManager nodesListManager,
|
NodesListManager nodesListManager,
|
||||||
NMLivelinessMonitor nmLivelinessMonitor,
|
NMLivelinessMonitor nmLivelinessMonitor,
|
||||||
@ -525,9 +530,6 @@ public NodeHeartbeatResponse nodeHeartbeat(NodeHeartbeatRequest request)
|
|||||||
YarnConfiguration.timelineServiceV2Enabled(getConfig());
|
YarnConfiguration.timelineServiceV2Enabled(getConfig());
|
||||||
if (timelineV2Enabled) {
|
if (timelineV2Enabled) {
|
||||||
// Check & update collectors info from request.
|
// Check & update collectors info from request.
|
||||||
// TODO make sure it won't have race condition issue for AM failed over
|
|
||||||
// case that the older registration could possible override the newer
|
|
||||||
// one.
|
|
||||||
updateAppCollectorsMap(request);
|
updateAppCollectorsMap(request);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -613,14 +615,14 @@ public NodeHeartbeatResponse nodeHeartbeat(NodeHeartbeatRequest request)
|
|||||||
|
|
||||||
private void setAppCollectorsMapToResponse(
|
private void setAppCollectorsMapToResponse(
|
||||||
List<ApplicationId> runningApps, NodeHeartbeatResponse response) {
|
List<ApplicationId> runningApps, NodeHeartbeatResponse response) {
|
||||||
Map<ApplicationId, String> liveAppCollectorsMap = new
|
Map<ApplicationId, AppCollectorData> liveAppCollectorsMap = new
|
||||||
HashMap<ApplicationId, String>();
|
HashMap<>();
|
||||||
Map<ApplicationId, RMApp> rmApps = rmContext.getRMApps();
|
Map<ApplicationId, RMApp> rmApps = rmContext.getRMApps();
|
||||||
// Set collectors for all running apps on this node.
|
// Set collectors for all running apps on this node.
|
||||||
for (ApplicationId appId : runningApps) {
|
for (ApplicationId appId : runningApps) {
|
||||||
String appCollectorAddr = rmApps.get(appId).getCollectorAddr();
|
AppCollectorData appCollectorData = rmApps.get(appId).getCollectorData();
|
||||||
if (appCollectorAddr != null) {
|
if (appCollectorData != null) {
|
||||||
liveAppCollectorsMap.put(appId, appCollectorAddr);
|
liveAppCollectorsMap.put(appId, appCollectorData);
|
||||||
} else {
|
} else {
|
||||||
if (LOG.isDebugEnabled()) {
|
if (LOG.isDebugEnabled()) {
|
||||||
LOG.debug("Collector for applicaton: " + appId +
|
LOG.debug("Collector for applicaton: " + appId +
|
||||||
@ -628,29 +630,43 @@ private void setAppCollectorsMapToResponse(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
response.setAppCollectorsMap(liveAppCollectorsMap);
|
response.setAppCollectors(liveAppCollectorsMap);
|
||||||
}
|
}
|
||||||
|
|
||||||
private void updateAppCollectorsMap(NodeHeartbeatRequest request) {
|
private void updateAppCollectorsMap(NodeHeartbeatRequest request) {
|
||||||
Map<ApplicationId, String> registeredCollectorsMap =
|
Map<ApplicationId, AppCollectorData> registeringCollectorsMap =
|
||||||
request.getRegisteredCollectors();
|
request.getRegisteringCollectors();
|
||||||
if (registeredCollectorsMap != null
|
if (registeringCollectorsMap != null
|
||||||
&& !registeredCollectorsMap.isEmpty()) {
|
&& !registeringCollectorsMap.isEmpty()) {
|
||||||
Map<ApplicationId, RMApp> rmApps = rmContext.getRMApps();
|
Map<ApplicationId, RMApp> rmApps = rmContext.getRMApps();
|
||||||
for (Map.Entry<ApplicationId, String> entry:
|
for (Map.Entry<ApplicationId, AppCollectorData> entry:
|
||||||
registeredCollectorsMap.entrySet()) {
|
registeringCollectorsMap.entrySet()) {
|
||||||
ApplicationId appId = entry.getKey();
|
ApplicationId appId = entry.getKey();
|
||||||
String collectorAddr = entry.getValue();
|
AppCollectorData collectorData = entry.getValue();
|
||||||
if (collectorAddr != null && !collectorAddr.isEmpty()) {
|
if (collectorData != null) {
|
||||||
|
if (!collectorData.isStamped()) {
|
||||||
|
// Stamp the collector if we have not done so
|
||||||
|
collectorData.setRMIdentifier(
|
||||||
|
ResourceManager.getClusterTimeStamp());
|
||||||
|
collectorData.setVersion(
|
||||||
|
timelineCollectorVersion.getAndIncrement());
|
||||||
|
}
|
||||||
RMApp rmApp = rmApps.get(appId);
|
RMApp rmApp = rmApps.get(appId);
|
||||||
if (rmApp == null) {
|
if (rmApp == null) {
|
||||||
LOG.warn("Cannot update collector info because application ID: " +
|
LOG.warn("Cannot update collector info because application ID: " +
|
||||||
appId + " is not found in RMContext!");
|
appId + " is not found in RMContext!");
|
||||||
} else {
|
} else {
|
||||||
String previousCollectorAddr = rmApp.getCollectorAddr();
|
AppCollectorData previousCollectorData = rmApp.getCollectorData();
|
||||||
if (previousCollectorAddr == null
|
if (AppCollectorData.happensBefore(previousCollectorData,
|
||||||
|| !previousCollectorAddr.equals(collectorAddr)) {
|
collectorData)) {
|
||||||
rmApp.setCollectorAddr(collectorAddr);
|
// Sending collector update event.
|
||||||
|
// Note: RM has to store the newly received collector data
|
||||||
|
// synchronously. Otherwise, the RM may send out stale collector
|
||||||
|
// data before this update is done, and the RM then crashes, the
|
||||||
|
// newly updated collector data will get lost.
|
||||||
|
LOG.info("Update collector information for application " + appId
|
||||||
|
+ " with new address: " + collectorData.getCollectorAddr());
|
||||||
|
((RMAppImpl) rmApp).setCollectorData(collectorData);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -23,6 +23,8 @@
|
|||||||
import java.util.Map;
|
import java.util.Map;
|
||||||
import java.util.Set;
|
import java.util.Set;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
import org.apache.hadoop.ipc.CallerContext;
|
import org.apache.hadoop.ipc.CallerContext;
|
||||||
import org.apache.hadoop.yarn.api.protocolrecords.FinishApplicationMasterRequest;
|
import org.apache.hadoop.yarn.api.protocolrecords.FinishApplicationMasterRequest;
|
||||||
import org.apache.hadoop.yarn.api.records.ApplicationAttemptId;
|
import org.apache.hadoop.yarn.api.records.ApplicationAttemptId;
|
||||||
@ -39,6 +41,7 @@
|
|||||||
import org.apache.hadoop.yarn.api.records.YarnApplicationState;
|
import org.apache.hadoop.yarn.api.records.YarnApplicationState;
|
||||||
import org.apache.hadoop.yarn.event.EventHandler;
|
import org.apache.hadoop.yarn.event.EventHandler;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.LogAggregationReport;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.LogAggregationReport;
|
||||||
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttempt;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttempt;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNode;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNode;
|
||||||
|
|
||||||
@ -180,27 +183,16 @@ ApplicationReport createAndGetApplicationReport(String clientUserName,
|
|||||||
String getTrackingUrl();
|
String getTrackingUrl();
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* The collector address for the application. It should be used only if the
|
* The timeline collector information for the application. It should be used
|
||||||
* timeline service v.2 is enabled.
|
* only if the timeline service v.2 is enabled.
|
||||||
*
|
*
|
||||||
* @return the address for the application's collector, or null if the
|
* @return the data for the application's collector, including collector
|
||||||
* timeline service v.2 is not enabled.
|
* address, collector ID. Return null if the timeline service v.2 is not
|
||||||
|
* enabled.
|
||||||
*/
|
*/
|
||||||
String getCollectorAddr();
|
@InterfaceAudience.Private
|
||||||
|
@InterfaceStability.Unstable
|
||||||
/**
|
AppCollectorData getCollectorData();
|
||||||
* Set collector address for the application. It should be used only if the
|
|
||||||
* timeline service v.2 is enabled.
|
|
||||||
*
|
|
||||||
* @param collectorAddr the address of collector
|
|
||||||
*/
|
|
||||||
void setCollectorAddr(String collectorAddr);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Remove collector address when application is finished or killed. It should
|
|
||||||
* be used only if the timeline service v.2 is enabled.
|
|
||||||
*/
|
|
||||||
void removeCollectorAddr();
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* The original tracking url for the application master.
|
* The original tracking url for the application master.
|
||||||
|
@ -30,9 +30,6 @@ public enum RMAppEventType {
|
|||||||
// Source: Scheduler
|
// Source: Scheduler
|
||||||
APP_ACCEPTED,
|
APP_ACCEPTED,
|
||||||
|
|
||||||
// TODO add source later
|
|
||||||
COLLECTOR_UPDATE,
|
|
||||||
|
|
||||||
// Source: RMAppAttempt
|
// Source: RMAppAttempt
|
||||||
ATTEMPT_REGISTERED,
|
ATTEMPT_REGISTERED,
|
||||||
ATTEMPT_UNREGISTERED,
|
ATTEMPT_UNREGISTERED,
|
||||||
|
@ -72,6 +72,7 @@
|
|||||||
import org.apache.hadoop.yarn.security.AMRMTokenIdentifier;
|
import org.apache.hadoop.yarn.security.AMRMTokenIdentifier;
|
||||||
import org.apache.hadoop.yarn.security.client.ClientToAMTokenIdentifier;
|
import org.apache.hadoop.yarn.security.client.ClientToAMTokenIdentifier;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.LogAggregationReport;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.LogAggregationReport;
|
||||||
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService;
|
import org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.RMAppManagerEvent;
|
import org.apache.hadoop.yarn.server.resourcemanager.RMAppManagerEvent;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.RMAppManagerEventType;
|
import org.apache.hadoop.yarn.server.resourcemanager.RMAppManagerEventType;
|
||||||
@ -165,7 +166,7 @@ public class RMAppImpl implements RMApp, Recoverable {
|
|||||||
private long storedFinishTime = 0;
|
private long storedFinishTime = 0;
|
||||||
private int firstAttemptIdInStateStore = 1;
|
private int firstAttemptIdInStateStore = 1;
|
||||||
private int nextAttemptId = 1;
|
private int nextAttemptId = 1;
|
||||||
private volatile String collectorAddr;
|
private AppCollectorData collectorData;
|
||||||
// This field isn't protected by readlock now.
|
// This field isn't protected by readlock now.
|
||||||
private volatile RMAppAttempt currentAttempt;
|
private volatile RMAppAttempt currentAttempt;
|
||||||
private String queue;
|
private String queue;
|
||||||
@ -611,18 +612,16 @@ public void setQueue(String queue) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public String getCollectorAddr() {
|
public AppCollectorData getCollectorData() {
|
||||||
return this.collectorAddr;
|
return this.collectorData;
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
public void setCollectorData(AppCollectorData incomingData) {
|
||||||
public void setCollectorAddr(String collectorAddress) {
|
this.collectorData = incomingData;
|
||||||
this.collectorAddr = collectorAddress;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
public void removeCollectorData() {
|
||||||
public void removeCollectorAddr() {
|
this.collectorData = null;
|
||||||
this.collectorAddr = null;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -63,6 +63,7 @@
|
|||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.RegisterNodeManagerRequest;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.RegisterNodeManagerRequest;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.RegisterNodeManagerResponse;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.RegisterNodeManagerResponse;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.UnRegisterNodeManagerRequest;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.UnRegisterNodeManagerRequest;
|
||||||
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.api.records.NodeAction;
|
import org.apache.hadoop.yarn.server.api.records.NodeAction;
|
||||||
import org.apache.hadoop.yarn.server.api.records.NodeHealthStatus;
|
import org.apache.hadoop.yarn.server.api.records.NodeHealthStatus;
|
||||||
import org.apache.hadoop.yarn.server.api.records.NodeStatus;
|
import org.apache.hadoop.yarn.server.api.records.NodeStatus;
|
||||||
@ -70,6 +71,7 @@
|
|||||||
import org.apache.hadoop.yarn.server.resourcemanager.nodelabels.NullRMNodeLabelsManager;
|
import org.apache.hadoop.yarn.server.resourcemanager.nodelabels.NullRMNodeLabelsManager;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.nodelabels.RMNodeLabelsManager;
|
import org.apache.hadoop.yarn.server.resourcemanager.nodelabels.RMNodeLabelsManager;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMApp;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMApp;
|
||||||
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppImpl;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppState;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppState;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptImpl;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptImpl;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNode;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNode;
|
||||||
@ -1011,13 +1013,23 @@ public void testNodeHeartbeatForAppCollectorsMap() throws Exception {
|
|||||||
RMNodeImpl node2 =
|
RMNodeImpl node2 =
|
||||||
(RMNodeImpl) rm.getRMContext().getRMNodes().get(nm2.getNodeId());
|
(RMNodeImpl) rm.getRMContext().getRMNodes().get(nm2.getNodeId());
|
||||||
|
|
||||||
RMApp app1 = rm.submitApp(1024);
|
RMAppImpl app1 = (RMAppImpl) rm.submitApp(1024);
|
||||||
String collectorAddr1 = "1.2.3.4:5";
|
String collectorAddr1 = "1.2.3.4:5";
|
||||||
app1.setCollectorAddr(collectorAddr1);
|
app1.setCollectorData(AppCollectorData.newInstance(
|
||||||
|
app1.getApplicationId(), collectorAddr1));
|
||||||
|
|
||||||
String collectorAddr2 = "5.4.3.2:1";
|
String collectorAddr2 = "5.4.3.2:1";
|
||||||
RMApp app2 = rm.submitApp(1024);
|
RMAppImpl app2 = (RMAppImpl) rm.submitApp(1024);
|
||||||
app2.setCollectorAddr(collectorAddr2);
|
app2.setCollectorData(AppCollectorData.newInstance(
|
||||||
|
app2.getApplicationId(), collectorAddr2));
|
||||||
|
|
||||||
|
String collectorAddr3 = "5.4.3.2:2";
|
||||||
|
app2.setCollectorData(AppCollectorData.newInstance(
|
||||||
|
app2.getApplicationId(), collectorAddr3, 0, 1));
|
||||||
|
|
||||||
|
String collectorAddr4 = "5.4.3.2:3";
|
||||||
|
app2.setCollectorData(AppCollectorData.newInstance(
|
||||||
|
app2.getApplicationId(), collectorAddr4, 1, 0));
|
||||||
|
|
||||||
// Create a running container for app1 running on nm1
|
// Create a running container for app1 running on nm1
|
||||||
ContainerId runningContainerId1 = BuilderUtils.newContainerId(
|
ContainerId runningContainerId1 = BuilderUtils.newContainerId(
|
||||||
@ -1055,14 +1067,18 @@ public void testNodeHeartbeatForAppCollectorsMap() throws Exception {
|
|||||||
Assert.assertEquals(app2.getApplicationId(), node2.getRunningApps().get(0));
|
Assert.assertEquals(app2.getApplicationId(), node2.getRunningApps().get(0));
|
||||||
|
|
||||||
nodeHeartbeat1 = nm1.nodeHeartbeat(true);
|
nodeHeartbeat1 = nm1.nodeHeartbeat(true);
|
||||||
Map<ApplicationId, String> map1 = nodeHeartbeat1.getAppCollectorsMap();
|
Map<ApplicationId, AppCollectorData> map1
|
||||||
|
= nodeHeartbeat1.getAppCollectors();
|
||||||
Assert.assertEquals(1, map1.size());
|
Assert.assertEquals(1, map1.size());
|
||||||
Assert.assertEquals(collectorAddr1, map1.get(app1.getApplicationId()));
|
Assert.assertEquals(collectorAddr1,
|
||||||
|
map1.get(app1.getApplicationId()).getCollectorAddr());
|
||||||
|
|
||||||
nodeHeartbeat2 = nm2.nodeHeartbeat(true);
|
nodeHeartbeat2 = nm2.nodeHeartbeat(true);
|
||||||
Map<ApplicationId, String> map2 = nodeHeartbeat2.getAppCollectorsMap();
|
Map<ApplicationId, AppCollectorData> map2
|
||||||
|
= nodeHeartbeat2.getAppCollectors();
|
||||||
Assert.assertEquals(1, map2.size());
|
Assert.assertEquals(1, map2.size());
|
||||||
Assert.assertEquals(collectorAddr2, map2.get(app2.getApplicationId()));
|
Assert.assertEquals(collectorAddr4,
|
||||||
|
map2.get(app2.getApplicationId()).getCollectorAddr());
|
||||||
}
|
}
|
||||||
|
|
||||||
private void checkRebootedNMCount(MockRM rm2, int count)
|
private void checkRebootedNMCount(MockRM rm2, int count)
|
||||||
|
@ -43,6 +43,7 @@
|
|||||||
import org.apache.hadoop.yarn.api.records.YarnApplicationState;
|
import org.apache.hadoop.yarn.api.records.YarnApplicationState;
|
||||||
import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.LogAggregationReport;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.LogAggregationReport;
|
||||||
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMApp;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMApp;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppEvent;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppEvent;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppMetrics;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppMetrics;
|
||||||
@ -97,15 +98,7 @@ public StringBuilder getDiagnostics() {
|
|||||||
throw new UnsupportedOperationException("Not supported yet.");
|
throw new UnsupportedOperationException("Not supported yet.");
|
||||||
}
|
}
|
||||||
@Override
|
@Override
|
||||||
public String getCollectorAddr() {
|
public AppCollectorData getCollectorData() {
|
||||||
throw new UnsupportedOperationException("Not supported yet.");
|
|
||||||
}
|
|
||||||
@Override
|
|
||||||
public void setCollectorAddr(String collectorAddr) {
|
|
||||||
throw new UnsupportedOperationException("Not supported yet.");
|
|
||||||
}
|
|
||||||
@Override
|
|
||||||
public void removeCollectorAddr() {
|
|
||||||
throw new UnsupportedOperationException("Not supported yet.");
|
throw new UnsupportedOperationException("Not supported yet.");
|
||||||
}
|
}
|
||||||
@Override
|
@Override
|
||||||
|
@ -43,6 +43,7 @@
|
|||||||
import org.apache.hadoop.yarn.api.records.impl.pb.ApplicationSubmissionContextPBImpl;
|
import org.apache.hadoop.yarn.api.records.impl.pb.ApplicationSubmissionContextPBImpl;
|
||||||
import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
||||||
import org.apache.hadoop.yarn.server.api.protocolrecords.LogAggregationReport;
|
import org.apache.hadoop.yarn.server.api.protocolrecords.LogAggregationReport;
|
||||||
|
import org.apache.hadoop.yarn.server.api.records.AppCollectorData;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttempt;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttempt;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNode;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNode;
|
||||||
|
|
||||||
@ -305,17 +306,8 @@ public CallerContext getCallerContext() {
|
|||||||
throw new UnsupportedOperationException("Not supported yet.");
|
throw new UnsupportedOperationException("Not supported yet.");
|
||||||
}
|
}
|
||||||
|
|
||||||
public String getCollectorAddr() {
|
|
||||||
throw new UnsupportedOperationException("Not supported yet.");
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void removeCollectorAddr() {
|
public AppCollectorData getCollectorData() {
|
||||||
throw new UnsupportedOperationException("Not supported yet.");
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public void setCollectorAddr(String collectorAddr) {
|
|
||||||
throw new UnsupportedOperationException("Not supported yet.");
|
throw new UnsupportedOperationException("Not supported yet.");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user