YARN-5575. Many classes use bare yarn. properties instead of the defined constants. Contributed by Daniel Templeton.
This commit is contained in:
parent
287effff93
commit
d3bb69a667
@ -33,6 +33,7 @@
|
||||
import org.apache.hadoop.mapreduce.MRConfig;
|
||||
import org.apache.hadoop.mapreduce.v2.MiniMRYarnCluster;
|
||||
import org.apache.hadoop.net.StandardSocketFactory;
|
||||
import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Test;
|
||||
|
||||
@ -87,9 +88,9 @@ public void testSocketFactory() throws IOException {
|
||||
jconf.set("hadoop.rpc.socket.factory.class.default",
|
||||
"org.apache.hadoop.ipc.DummySocketFactory");
|
||||
jconf.set(MRConfig.FRAMEWORK_NAME, MRConfig.YARN_FRAMEWORK_NAME);
|
||||
String rmAddress = jconf.get("yarn.resourcemanager.address");
|
||||
String rmAddress = jconf.get(YarnConfiguration.RM_ADDRESS);
|
||||
String[] split = rmAddress.split(":");
|
||||
jconf.set("yarn.resourcemanager.address", split[0] + ':'
|
||||
jconf.set(YarnConfiguration.RM_ADDRESS, split[0] + ':'
|
||||
+ (Integer.parseInt(split[1]) + 10));
|
||||
client = new JobClient(jconf);
|
||||
|
||||
|
@ -163,7 +163,7 @@ public static void setup() throws IOException {
|
||||
Configuration conf = new Configuration();
|
||||
conf.set("fs.defaultFS", remoteFs.getUri().toString()); // use HDFS
|
||||
conf.set(MRJobConfig.MR_AM_STAGING_DIR, "/apps_staging_dir");
|
||||
conf.setInt("yarn.cluster.max-application-priority", 10);
|
||||
conf.setInt(YarnConfiguration.MAX_CLUSTER_LEVEL_APPLICATION_PRIORITY, 10);
|
||||
mrCluster.init(conf);
|
||||
mrCluster.start();
|
||||
}
|
||||
@ -418,8 +418,7 @@ public void testJobWithChangePriority() throws Exception {
|
||||
// set master address to local to test that local mode applied if framework
|
||||
// equals local
|
||||
sleepConf.set(MRConfig.MASTER_ADDRESS, "local");
|
||||
sleepConf
|
||||
.setInt("yarn.app.mapreduce.am.scheduler.heartbeat.interval-ms", 5);
|
||||
sleepConf.setInt(MRJobConfig.MR_AM_TO_RM_HEARTBEAT_INTERVAL_MS, 5);
|
||||
|
||||
SleepJob sleepJob = new SleepJob();
|
||||
sleepJob.setConf(sleepConf);
|
||||
|
@ -11,6 +11,7 @@
|
||||
import org.apache.hadoop.mapreduce.MRJobConfig;
|
||||
import org.apache.hadoop.mapreduce.server.jobtracker.JTConfig;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import static org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacitySchedulerConfiguration.PREFIX;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
@ -45,8 +46,8 @@ public static void initCluster(Class<?> caller) throws IOException {
|
||||
Configuration conf = new Configuration();
|
||||
// conf.set("mapred.queue.names", "default,q1,q2");
|
||||
conf.set("mapred.queue.names", "default");
|
||||
conf.set("yarn.scheduler.capacity.root.queues", "default");
|
||||
conf.set("yarn.scheduler.capacity.root.default.capacity", "100.0");
|
||||
conf.set(PREFIX + "root.queues", "default");
|
||||
conf.set(PREFIX + "root.default.capacity", "100.0");
|
||||
|
||||
|
||||
conf.setBoolean(GRIDMIX_USE_QUEUE_IN_TRACE, false);
|
||||
|
@ -220,7 +220,8 @@ private void setupInternal(int numNodeManager, float timelineVersion)
|
||||
throw new RuntimeException("Could not find 'yarn-site.xml' dummy file in classpath");
|
||||
}
|
||||
Configuration yarnClusterConfig = yarnCluster.getConfig();
|
||||
yarnClusterConfig.set("yarn.application.classpath", new File(url.getPath()).getParent());
|
||||
yarnClusterConfig.set(YarnConfiguration.YARN_APPLICATION_CLASSPATH,
|
||||
new File(url.getPath()).getParent());
|
||||
//write the document to a buffer (not directly to the file, as that
|
||||
//can cause the file being written to get read -which will then fail.
|
||||
ByteArrayOutputStream bytesOut = new ByteArrayOutputStream();
|
||||
@ -234,8 +235,7 @@ private void setupInternal(int numNodeManager, float timelineVersion)
|
||||
FileContext fsContext = FileContext.getLocalFSFileContext();
|
||||
fsContext
|
||||
.delete(
|
||||
new Path(conf
|
||||
.get("yarn.timeline-service.leveldb-timeline-store.path")),
|
||||
new Path(conf.get(YarnConfiguration.TIMELINE_SERVICE_LEVELDB_PATH)),
|
||||
true);
|
||||
try {
|
||||
Thread.sleep(2000);
|
||||
@ -263,8 +263,7 @@ public void tearDown() throws IOException {
|
||||
FileContext fsContext = FileContext.getLocalFSFileContext();
|
||||
fsContext
|
||||
.delete(
|
||||
new Path(conf
|
||||
.get("yarn.timeline-service.leveldb-timeline-store.path")),
|
||||
new Path(conf.get(YarnConfiguration.TIMELINE_SERVICE_LEVELDB_PATH)),
|
||||
true);
|
||||
}
|
||||
|
||||
|
@ -28,6 +28,7 @@
|
||||
import org.apache.hadoop.yarn.server.nodemanager.NodeManager;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.RMContext;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.nodelabels.RMNodeLabelsManager;
|
||||
import static org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacitySchedulerConfiguration.PREFIX;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
@ -57,13 +58,12 @@ private void initializeNodeLabels() throws IOException {
|
||||
labelsMgr.addToCluserNodeLabelsWithDefaultExclusivity(labels);
|
||||
|
||||
// Setup queue access to node labels
|
||||
distShellTest.conf.set("yarn.scheduler.capacity.root.accessible-node-labels", "x");
|
||||
distShellTest.conf.set("yarn.scheduler.capacity.root.accessible-node-labels.x.capacity",
|
||||
"100");
|
||||
distShellTest.conf.set("yarn.scheduler.capacity.root.default.accessible-node-labels", "x");
|
||||
distShellTest.conf.set(
|
||||
"yarn.scheduler.capacity.root.default.accessible-node-labels.x.capacity",
|
||||
distShellTest.conf.set(PREFIX + "root.accessible-node-labels", "x");
|
||||
distShellTest.conf.set(PREFIX + "root.accessible-node-labels.x.capacity",
|
||||
"100");
|
||||
distShellTest.conf.set(PREFIX + "root.default.accessible-node-labels", "x");
|
||||
distShellTest.conf.set(PREFIX
|
||||
+ "root.default.accessible-node-labels.x.capacity", "100");
|
||||
|
||||
rmContext.getScheduler().reinitialize(distShellTest.conf, rmContext);
|
||||
|
||||
|
@ -80,6 +80,7 @@
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacitySchedulerConfiguration;
|
||||
import org.apache.hadoop.yarn.util.Records;
|
||||
import org.apache.hadoop.yarn.util.Times;
|
||||
import static org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacitySchedulerConfiguration.PREFIX;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
@ -1704,8 +1705,7 @@ public void testGetQueueInfoPreemptionDisabled() throws Exception {
|
||||
"org.apache.hadoop.yarn.server.resourcemanager.monitor.capacity."
|
||||
+ "ProportionalCapacityPreemptionPolicy");
|
||||
conf.setBoolean(YarnConfiguration.RM_SCHEDULER_ENABLE_MONITORS, true);
|
||||
conf.setBoolean(
|
||||
"yarn.scheduler.capacity.root.a.a1.disable_preemption", true);
|
||||
conf.setBoolean(PREFIX + "root.a.a1.disable_preemption", true);
|
||||
MiniYARNCluster cluster =
|
||||
new MiniYARNCluster("testReservationAPIs", 2, 1, 1);
|
||||
|
||||
|
@ -23,6 +23,7 @@
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.http.FilterContainer;
|
||||
import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
||||
import static org.apache.hadoop.yarn.server.timeline.security.TimelineAuthenticationFilterInitializer.PREFIX;
|
||||
import org.junit.Test;
|
||||
import org.mockito.Mockito;
|
||||
|
||||
@ -35,36 +36,30 @@ public void testProxyUserConfiguration() {
|
||||
for (int i = 0; i < 3; ++i) {
|
||||
Configuration conf = new YarnConfiguration();
|
||||
switch (i) {
|
||||
case 0:
|
||||
// hadoop.proxyuser prefix
|
||||
conf.set("hadoop.proxyuser.foo.hosts", "*");
|
||||
conf.set("hadoop.proxyuser.foo.users", "*");
|
||||
conf.set("hadoop.proxyuser.foo.groups", "*");
|
||||
break;
|
||||
case 1:
|
||||
// yarn.timeline-service.http-authentication.proxyuser prefix
|
||||
conf.set("yarn.timeline-service.http-authentication.proxyuser.foo.hosts",
|
||||
"*");
|
||||
conf.set("yarn.timeline-service.http-authentication.proxyuser.foo.users",
|
||||
"*");
|
||||
conf.set("yarn.timeline-service.http-authentication.proxyuser.foo.groups",
|
||||
"*");
|
||||
break;
|
||||
case 2:
|
||||
// hadoop.proxyuser prefix has been overwritten by
|
||||
// yarn.timeline-service.http-authentication.proxyuser prefix
|
||||
conf.set("hadoop.proxyuser.foo.hosts", "bar");
|
||||
conf.set("hadoop.proxyuser.foo.users", "bar");
|
||||
conf.set("hadoop.proxyuser.foo.groups", "bar");
|
||||
conf.set("yarn.timeline-service.http-authentication.proxyuser.foo.hosts",
|
||||
"*");
|
||||
conf.set("yarn.timeline-service.http-authentication.proxyuser.foo.users",
|
||||
"*");
|
||||
conf.set("yarn.timeline-service.http-authentication.proxyuser.foo.groups",
|
||||
"*");
|
||||
break;
|
||||
default:
|
||||
break;
|
||||
case 0:
|
||||
// hadoop.proxyuser prefix
|
||||
conf.set("hadoop.proxyuser.foo.hosts", "*");
|
||||
conf.set("hadoop.proxyuser.foo.users", "*");
|
||||
conf.set("hadoop.proxyuser.foo.groups", "*");
|
||||
break;
|
||||
case 1:
|
||||
// yarn.timeline-service.http-authentication.proxyuser prefix
|
||||
conf.set(PREFIX + "proxyuser.foo.hosts", "*");
|
||||
conf.set(PREFIX + "proxyuser.foo.users", "*");
|
||||
conf.set(PREFIX + "proxyuser.foo.groups", "*");
|
||||
break;
|
||||
case 2:
|
||||
// hadoop.proxyuser prefix has been overwritten by
|
||||
// yarn.timeline-service.http-authentication.proxyuser prefix
|
||||
conf.set("hadoop.proxyuser.foo.hosts", "bar");
|
||||
conf.set("hadoop.proxyuser.foo.users", "bar");
|
||||
conf.set("hadoop.proxyuser.foo.groups", "bar");
|
||||
conf.set(PREFIX + "proxyuser.foo.hosts", "*");
|
||||
conf.set(PREFIX + "proxyuser.foo.users", "*");
|
||||
conf.set(PREFIX + "proxyuser.foo.groups", "*");
|
||||
break;
|
||||
default:
|
||||
break;
|
||||
}
|
||||
|
||||
TimelineAuthenticationFilterInitializer initializer =
|
||||
|
@ -549,8 +549,8 @@ private static Configuration createCapacitySchedulerConfiguration() {
|
||||
csConf.setBoolean(YarnConfiguration.RM_RESERVATION_SYSTEM_ENABLE, true);
|
||||
csConf.setBoolean(YarnConfiguration.YARN_ACL_ENABLE, true);
|
||||
csConf.setBoolean(YarnConfiguration.YARN_RESERVATION_ACL_ENABLE, true);
|
||||
csConf.set("yarn.resourcemanager.scheduler.class", CapacityScheduler
|
||||
.class.getName());
|
||||
csConf.set(YarnConfiguration.RM_SCHEDULER,
|
||||
CapacityScheduler.class.getName());
|
||||
|
||||
return csConf;
|
||||
}
|
||||
@ -602,8 +602,7 @@ private static Configuration createFairSchedulerConfiguration() throws
|
||||
fsConf.setBoolean(YarnConfiguration.RM_RESERVATION_SYSTEM_ENABLE, true);
|
||||
fsConf.setBoolean(YarnConfiguration.YARN_ACL_ENABLE, true);
|
||||
fsConf.setBoolean(YarnConfiguration.YARN_RESERVATION_ACL_ENABLE, true);
|
||||
fsConf.set("yarn.resourcemanager.scheduler.class", FairScheduler.class
|
||||
.getName());
|
||||
fsConf.set(YarnConfiguration.RM_SCHEDULER, FairScheduler.class.getName());
|
||||
|
||||
return fsConf;
|
||||
}
|
||||
|
@ -80,6 +80,7 @@
|
||||
import org.apache.hadoop.yarn.util.Records;
|
||||
import org.apache.hadoop.yarn.util.resource.ResourceCalculator;
|
||||
import org.apache.hadoop.yarn.util.resource.Resources;
|
||||
import static org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacitySchedulerConfiguration.PREFIX;
|
||||
import org.junit.After;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
@ -291,12 +292,11 @@ public void testQueueSubmitWithNoPermission() throws IOException {
|
||||
YarnConfiguration conf = new YarnConfiguration();
|
||||
conf.set(YarnConfiguration.RM_SCHEDULER,
|
||||
CapacityScheduler.class.getCanonicalName());
|
||||
conf.set("yarn.scheduler.capacity.root.acl_submit_applications", " ");
|
||||
conf.set("yarn.scheduler.capacity.root.acl_administer_queue", " ");
|
||||
conf.set(PREFIX + "root.acl_submit_applications", " ");
|
||||
conf.set(PREFIX + "root.acl_administer_queue", " ");
|
||||
|
||||
conf.set("yarn.scheduler.capacity.root.default.acl_submit_applications",
|
||||
" ");
|
||||
conf.set("yarn.scheduler.capacity.root.default.acl_administer_queue", " ");
|
||||
conf.set(PREFIX + "root.default.acl_submit_applications", " ");
|
||||
conf.set(PREFIX + "root.default.acl_administer_queue", " ");
|
||||
conf.set(YarnConfiguration.YARN_ACL_ENABLE, "true");
|
||||
MockRM mockRM = new MockRM(conf);
|
||||
ClientRMService rmService = mockRM.getClientRMService();
|
||||
|
@ -69,6 +69,7 @@
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptState;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNode;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacitySchedulerConfiguration;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.security.NMTokenSecretManagerInRM;
|
||||
import org.apache.log4j.Level;
|
||||
import org.apache.log4j.LogManager;
|
||||
@ -136,7 +137,7 @@ public void testAppWithNoContainers() throws Exception {
|
||||
public void testAppOnMultiNode() throws Exception {
|
||||
Logger rootLogger = LogManager.getRootLogger();
|
||||
rootLogger.setLevel(Level.DEBUG);
|
||||
conf.set("yarn.scheduler.capacity.node-locality-delay", "-1");
|
||||
conf.set(CapacitySchedulerConfiguration.NODE_LOCALITY_DELAY, "-1");
|
||||
MockRM rm = new MockRM(conf);
|
||||
rm.start();
|
||||
MockNM nm1 = rm.registerNode("h1:1234", 5120);
|
||||
|
@ -68,6 +68,9 @@
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNodeImpl;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacitySchedulerConfiguration;
|
||||
import static org.apache.hadoop.yarn.conf.YarnConfiguration.RM_PROXY_USER_PREFIX;
|
||||
import static org.apache.hadoop.yarn.server.resourcemanager.resource.DynamicResourceConfiguration.NODES;
|
||||
import static org.apache.hadoop.yarn.server.resourcemanager.resource.DynamicResourceConfiguration.PREFIX;
|
||||
import org.junit.After;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
@ -165,7 +168,8 @@ public void testAdminRefreshQueuesWithFileSystemBasedConfigurationProvider()
|
||||
|
||||
CapacitySchedulerConfiguration csConf =
|
||||
new CapacitySchedulerConfiguration();
|
||||
csConf.set("yarn.scheduler.capacity.maximum-applications", "5000");
|
||||
csConf.set(CapacitySchedulerConfiguration.MAXIMUM_SYSTEM_APPLICATIONS,
|
||||
"5000");
|
||||
uploadConfiguration(csConf, "capacity-scheduler.xml");
|
||||
|
||||
rm.adminService.refreshQueues(RefreshQueuesRequest.newInstance());
|
||||
@ -226,9 +230,9 @@ public void testRefreshNodesResourceWithFileSystemBasedConfigurationProvider()
|
||||
|
||||
DynamicResourceConfiguration drConf =
|
||||
new DynamicResourceConfiguration();
|
||||
drConf.set("yarn.resource.dynamic.nodes", "h1:1234");
|
||||
drConf.set("yarn.resource.dynamic.h1:1234.vcores", "4");
|
||||
drConf.set("yarn.resource.dynamic.h1:1234.memory", "4096");
|
||||
drConf.set(PREFIX + NODES, "h1:1234");
|
||||
drConf.set(PREFIX + "h1:1234.vcores", "4");
|
||||
drConf.set(PREFIX + "h1:1234.memory", "4096");
|
||||
uploadConfiguration(drConf, "dynamic-resources.xml");
|
||||
|
||||
rm.adminService.refreshNodesResources(
|
||||
@ -266,9 +270,9 @@ public void testRefreshNodesResourceWithResourceReturnInRegistration()
|
||||
|
||||
DynamicResourceConfiguration drConf =
|
||||
new DynamicResourceConfiguration();
|
||||
drConf.set("yarn.resource.dynamic.nodes", "h1:1234");
|
||||
drConf.set("yarn.resource.dynamic.h1:1234.vcores", "4");
|
||||
drConf.set("yarn.resource.dynamic.h1:1234.memory", "4096");
|
||||
drConf.set(PREFIX + NODES, "h1:1234");
|
||||
drConf.set(PREFIX + "h1:1234.vcores", "4");
|
||||
drConf.set(PREFIX + "h1:1234.memory", "4096");
|
||||
uploadConfiguration(drConf, "dynamic-resources.xml");
|
||||
|
||||
rm.adminService.refreshNodesResources(
|
||||
@ -316,9 +320,9 @@ public void testRefreshNodesResourceWithResourceReturnInHeartbeat()
|
||||
|
||||
DynamicResourceConfiguration drConf =
|
||||
new DynamicResourceConfiguration();
|
||||
drConf.set("yarn.resource.dynamic.nodes", "h1:1234");
|
||||
drConf.set("yarn.resource.dynamic.h1:1234.vcores", "4");
|
||||
drConf.set("yarn.resource.dynamic.h1:1234.memory", "4096");
|
||||
drConf.set(PREFIX + NODES, "h1:1234");
|
||||
drConf.set(PREFIX + "h1:1234.vcores", "4");
|
||||
drConf.set(PREFIX + "h1:1234.memory", "4096");
|
||||
uploadConfiguration(drConf, "dynamic-resources.xml");
|
||||
|
||||
rm.adminService.refreshNodesResources(
|
||||
@ -364,9 +368,9 @@ public void testResourcePersistentForNMRegistrationWithNewResource()
|
||||
|
||||
DynamicResourceConfiguration drConf =
|
||||
new DynamicResourceConfiguration();
|
||||
drConf.set("yarn.resource.dynamic.nodes", "h1:1234");
|
||||
drConf.set("yarn.resource.dynamic.h1:1234.vcores", "4");
|
||||
drConf.set("yarn.resource.dynamic.h1:1234.memory", "4096");
|
||||
drConf.set(PREFIX + NODES, "h1:1234");
|
||||
drConf.set(PREFIX + "h1:1234.vcores", "4");
|
||||
drConf.set(PREFIX + "h1:1234.memory", "4096");
|
||||
uploadConfiguration(drConf, "dynamic-resources.xml");
|
||||
|
||||
rm.adminService.refreshNodesResources(
|
||||
@ -611,8 +615,8 @@ private void verifyServiceACLsRefresh(ServiceAuthorizationManager manager,
|
||||
.get("hadoop.proxyuser.test.hosts").contains("test_hosts"));
|
||||
|
||||
Configuration yarnConf = new Configuration(false);
|
||||
yarnConf.set("yarn.resourcemanager.proxyuser.test.groups", "test_groups_1");
|
||||
yarnConf.set("yarn.resourcemanager.proxyuser.test.hosts", "test_hosts_1");
|
||||
yarnConf.set(RM_PROXY_USER_PREFIX + "test.groups", "test_groups_1");
|
||||
yarnConf.set(RM_PROXY_USER_PREFIX + "test.hosts", "test_hosts_1");
|
||||
uploadConfiguration(yarnConf, "yarn-site.xml");
|
||||
|
||||
// RM specific configs will overwrite the common ones
|
||||
@ -809,7 +813,8 @@ public void testRMHAWithFileSystemBasedConfiguration() throws IOException,
|
||||
|
||||
CapacitySchedulerConfiguration csConf =
|
||||
new CapacitySchedulerConfiguration();
|
||||
csConf.set("yarn.scheduler.capacity.maximum-applications", "5000");
|
||||
csConf.set(CapacitySchedulerConfiguration.MAXIMUM_SYSTEM_APPLICATIONS,
|
||||
"5000");
|
||||
uploadConfiguration(csConf, "capacity-scheduler.xml");
|
||||
|
||||
rm1.adminService.refreshQueues(RefreshQueuesRequest.newInstance());
|
||||
@ -897,7 +902,8 @@ public void testRMInitialsWithFileSystemBasedConfigurationProvider()
|
||||
|
||||
CapacitySchedulerConfiguration csConf =
|
||||
new CapacitySchedulerConfiguration();
|
||||
csConf.set("yarn.scheduler.capacity.maximum-applications", "5000");
|
||||
csConf.set(CapacitySchedulerConfiguration.MAXIMUM_SYSTEM_APPLICATIONS,
|
||||
"5000");
|
||||
uploadConfiguration(csConf, "capacity-scheduler.xml");
|
||||
|
||||
String aclsString = "alice,bob users,wheel";
|
||||
|
@ -26,6 +26,7 @@
|
||||
import org.apache.hadoop.security.authorize.AuthorizationException;
|
||||
import org.apache.hadoop.security.authorize.ProxyUsers;
|
||||
import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
||||
import static org.apache.hadoop.yarn.conf.YarnConfiguration.RM_PROXY_USER_PREFIX;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Test;
|
||||
import org.junit.runner.RunWith;
|
||||
@ -50,30 +51,30 @@ public static Collection<Object[]> headers() {
|
||||
public TestRMProxyUsersConf(int round) {
|
||||
conf = new YarnConfiguration();
|
||||
switch (round) {
|
||||
case 0:
|
||||
// hadoop.proxyuser prefix
|
||||
conf.set("hadoop.proxyuser.foo.hosts", ipAddress);
|
||||
conf.set("hadoop.proxyuser.foo.users", "bar");
|
||||
conf.set("hadoop.proxyuser.foo.groups", "bar_group");
|
||||
break;
|
||||
case 1:
|
||||
// yarn.resourcemanager.proxyuser prefix
|
||||
conf.set("yarn.resourcemanager.proxyuser.foo.hosts", ipAddress);
|
||||
conf.set("yarn.resourcemanager.proxyuser.foo.users", "bar");
|
||||
conf.set("yarn.resourcemanager.proxyuser.foo.groups", "bar_group");
|
||||
break;
|
||||
case 2:
|
||||
// hadoop.proxyuser prefix has been overwritten by
|
||||
// yarn.resourcemanager.proxyuser prefix
|
||||
conf.set("hadoop.proxyuser.foo.hosts", "xyz");
|
||||
conf.set("hadoop.proxyuser.foo.users", "xyz");
|
||||
conf.set("hadoop.proxyuser.foo.groups", "xyz");
|
||||
conf.set("yarn.resourcemanager.proxyuser.foo.hosts", ipAddress);
|
||||
conf.set("yarn.resourcemanager.proxyuser.foo.users", "bar");
|
||||
conf.set("yarn.resourcemanager.proxyuser.foo.groups", "bar_group");
|
||||
break;
|
||||
default:
|
||||
break;
|
||||
case 0:
|
||||
// hadoop.proxyuser prefix
|
||||
conf.set("hadoop.proxyuser.foo.hosts", ipAddress);
|
||||
conf.set("hadoop.proxyuser.foo.users", "bar");
|
||||
conf.set("hadoop.proxyuser.foo.groups", "bar_group");
|
||||
break;
|
||||
case 1:
|
||||
// yarn.resourcemanager.proxyuser prefix
|
||||
conf.set(RM_PROXY_USER_PREFIX + "foo.hosts", ipAddress);
|
||||
conf.set(RM_PROXY_USER_PREFIX + "foo.users", "bar");
|
||||
conf.set(RM_PROXY_USER_PREFIX + "foo.groups", "bar_group");
|
||||
break;
|
||||
case 2:
|
||||
// hadoop.proxyuser prefix has been overwritten by
|
||||
// yarn.resourcemanager.proxyuser prefix
|
||||
conf.set("hadoop.proxyuser.foo.hosts", "xyz");
|
||||
conf.set("hadoop.proxyuser.foo.users", "xyz");
|
||||
conf.set("hadoop.proxyuser.foo.groups", "xyz");
|
||||
conf.set(RM_PROXY_USER_PREFIX + "foo.hosts", ipAddress);
|
||||
conf.set(RM_PROXY_USER_PREFIX + "foo.users", "bar");
|
||||
conf.set(RM_PROXY_USER_PREFIX + "foo.groups", "bar_group");
|
||||
break;
|
||||
default:
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -94,6 +94,7 @@
|
||||
import java.util.Map;
|
||||
import java.util.Set;
|
||||
import java.util.concurrent.TimeUnit;
|
||||
import static org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacitySchedulerConfiguration.PREFIX;
|
||||
|
||||
import static org.junit.Assert.*;
|
||||
import static org.mockito.Mockito.mock;
|
||||
@ -644,7 +645,7 @@ public void testCapacitySchedulerRecovery() throws Exception {
|
||||
rm1.clearQueueMetrics(app1_2);
|
||||
rm1.clearQueueMetrics(app2);
|
||||
|
||||
csConf.set("yarn.scheduler.capacity.root.Default.QueueB.state", "STOPPED");
|
||||
csConf.set(PREFIX + "root.Default.QueueB.state", "STOPPED");
|
||||
|
||||
// Re-start RM
|
||||
rm2 = new MockRM(csConf, memStore);
|
||||
|
@ -389,10 +389,9 @@ private void testRMWritingMassiveHistory(boolean isFS) throws Exception {
|
||||
YarnConfiguration conf = new YarnConfiguration();
|
||||
if (isFS) {
|
||||
conf.setBoolean(FairSchedulerConfiguration.ASSIGN_MULTIPLE, true);
|
||||
conf.set("yarn.resourcemanager.scheduler.class",
|
||||
FairScheduler.class.getName());
|
||||
conf.set(YarnConfiguration.RM_SCHEDULER, FairScheduler.class.getName());
|
||||
} else {
|
||||
conf.set("yarn.resourcemanager.scheduler.class",
|
||||
conf.set(YarnConfiguration.RM_SCHEDULER,
|
||||
CapacityScheduler.class.getName());
|
||||
}
|
||||
// don't process history events
|
||||
|
@ -157,8 +157,7 @@ public void setup() {
|
||||
conf.setBoolean(YarnConfiguration.RM_SCHEDULER_ENABLE_MONITORS, true);
|
||||
// FairScheduler doesn't support this test,
|
||||
// Set CapacityScheduler as the scheduler for this test.
|
||||
conf.set("yarn.resourcemanager.scheduler.class",
|
||||
CapacityScheduler.class.getName());
|
||||
conf.set(YarnConfiguration.RM_SCHEDULER, CapacityScheduler.class.getName());
|
||||
|
||||
mClock = mock(Clock.class);
|
||||
mCS = mock(CapacityScheduler.class);
|
||||
|
@ -60,6 +60,7 @@
|
||||
import org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator;
|
||||
import org.apache.hadoop.yarn.util.resource.ResourceCalculator;
|
||||
import org.apache.hadoop.yarn.util.resource.Resources;
|
||||
import static org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacitySchedulerConfiguration.PREFIX;
|
||||
import org.junit.After;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
@ -339,11 +340,8 @@ public void testLimitsComputation() throws Exception {
|
||||
);
|
||||
|
||||
// Change the per-queue max AM resources percentage.
|
||||
csConf.setFloat(
|
||||
"yarn.scheduler.capacity." +
|
||||
queue.getQueuePath() +
|
||||
".maximum-am-resource-percent",
|
||||
0.5f);
|
||||
csConf.setFloat(PREFIX + queue.getQueuePath()
|
||||
+ ".maximum-am-resource-percent", 0.5f);
|
||||
// Re-create queues to get new configs.
|
||||
queues = new HashMap<String, CSQueue>();
|
||||
root =
|
||||
@ -364,10 +362,8 @@ public void testLimitsComputation() throws Exception {
|
||||
Resource.newInstance(400*GB, 1));
|
||||
|
||||
// Change the per-queue max applications.
|
||||
csConf.setInt(
|
||||
"yarn.scheduler.capacity." +
|
||||
queue.getQueuePath() +
|
||||
".maximum-applications", 9999);
|
||||
csConf.setInt(PREFIX + queue.getQueuePath() + ".maximum-applications",
|
||||
9999);
|
||||
// Re-create queues to get new configs.
|
||||
queues = new HashMap<String, CSQueue>();
|
||||
root =
|
||||
|
@ -66,7 +66,8 @@ protected Configuration createConfiguration() {
|
||||
csConf.setAcls(CapacitySchedulerConfiguration.ROOT, aclsOnRootQueue);
|
||||
|
||||
csConf.setBoolean(YarnConfiguration.YARN_ACL_ENABLE, true);
|
||||
csConf.set("yarn.resourcemanager.scheduler.class", CapacityScheduler.class.getName());
|
||||
csConf.set(YarnConfiguration.RM_SCHEDULER,
|
||||
CapacityScheduler.class.getName());
|
||||
|
||||
return csConf;
|
||||
}
|
||||
|
@ -154,9 +154,9 @@ private void setUpInternal(ResourceCalculator rC) throws Exception {
|
||||
|
||||
csConf =
|
||||
new CapacitySchedulerConfiguration();
|
||||
csConf.setBoolean("yarn.scheduler.capacity.user-metrics.enable", true);
|
||||
csConf.setBoolean(
|
||||
"yarn.scheduler.capacity.reservations-continue-look-all-nodes", false);
|
||||
csConf.setBoolean(CapacitySchedulerConfiguration.ENABLE_USER_METRICS, true);
|
||||
csConf.setBoolean(CapacitySchedulerConfiguration.RESERVE_CONT_LOOK_ALL_NODES,
|
||||
false);
|
||||
final String newRoot = "root" + System.currentTimeMillis();
|
||||
setupQueueConfiguration(csConf, newRoot);
|
||||
YarnConfiguration conf = new YarnConfiguration();
|
||||
|
@ -112,7 +112,7 @@ private void setup(CapacitySchedulerConfiguration csConf) throws Exception {
|
||||
private void setup(CapacitySchedulerConfiguration csConf,
|
||||
boolean addUserLimits) throws Exception {
|
||||
|
||||
csConf.setBoolean("yarn.scheduler.capacity.user-metrics.enable", true);
|
||||
csConf.setBoolean(CapacitySchedulerConfiguration.ENABLE_USER_METRICS, true);
|
||||
final String newRoot = "root" + System.currentTimeMillis();
|
||||
// final String newRoot = "root";
|
||||
|
||||
@ -502,8 +502,8 @@ public void testReservationNoContinueLook() throws Exception {
|
||||
queues = new HashMap<String, CSQueue>();
|
||||
// test that the deadlock occurs when turned off
|
||||
CapacitySchedulerConfiguration csConf = new CapacitySchedulerConfiguration();
|
||||
csConf.setBoolean(
|
||||
"yarn.scheduler.capacity.reservations-continue-look-all-nodes", false);
|
||||
csConf.setBoolean(CapacitySchedulerConfiguration.RESERVE_CONT_LOOK_ALL_NODES,
|
||||
false);
|
||||
setup(csConf);
|
||||
|
||||
// Manipulate queue 'a'
|
||||
|
@ -45,7 +45,7 @@ public void setUp() throws IOException {
|
||||
Configuration conf = new YarnConfiguration();
|
||||
conf.setClass(YarnConfiguration.RM_SCHEDULER, FairScheduler.class,
|
||||
ResourceScheduler.class);
|
||||
conf.set("yarn.scheduler.fair.event-log-enabled", "true");
|
||||
conf.set(FairSchedulerConfiguration.EVENT_LOG_ENABLED, "true");
|
||||
|
||||
// All tests assume only one assignment per node update
|
||||
conf.set(FairSchedulerConfiguration.ASSIGN_MULTIPLE, "false");
|
||||
|
@ -55,7 +55,7 @@ protected Configuration createConfiguration() throws IOException {
|
||||
fsConf.set(FairSchedulerConfiguration.ALLOCATION_FILE, ALLOC_FILE);
|
||||
|
||||
fsConf.setBoolean(YarnConfiguration.YARN_ACL_ENABLE, true);
|
||||
fsConf.set("yarn.resourcemanager.scheduler.class", FairScheduler.class.getName());
|
||||
fsConf.set(YarnConfiguration.RM_SCHEDULER, FairScheduler.class.getName());
|
||||
|
||||
return fsConf;
|
||||
}
|
||||
|
@ -189,7 +189,7 @@ protected void configureServlets() {
|
||||
private class CapTestServletModule extends TestServletModule {
|
||||
@Override
|
||||
public void configureScheduler() {
|
||||
conf.set("yarn.resourcemanager.scheduler.class",
|
||||
conf.set(YarnConfiguration.RM_SCHEDULER,
|
||||
CapacityScheduler.class.getName());
|
||||
}
|
||||
}
|
||||
@ -215,8 +215,7 @@ public void configureScheduler() {
|
||||
} catch(IOException e) {
|
||||
}
|
||||
conf.set(FairSchedulerConfiguration.ALLOCATION_FILE, FS_ALLOC_FILE);
|
||||
conf.set("yarn.resourcemanager.scheduler.class",
|
||||
FairScheduler.class.getName());
|
||||
conf.set(YarnConfiguration.RM_SCHEDULER, FairScheduler.class.getName());
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -58,7 +58,6 @@
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.ResourceScheduler;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.webapp.dao.ApplicationSubmissionContextInfo;
|
||||
import org.apache.hadoop.yarn.util.ConverterUtils;
|
||||
import org.codehaus.jettison.json.JSONObject;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Assert;
|
||||
@ -66,6 +65,7 @@
|
||||
import org.junit.Test;
|
||||
|
||||
import com.sun.jersey.api.client.ClientResponse.Status;
|
||||
import static org.apache.hadoop.yarn.conf.YarnConfiguration.RM_PROXY_USER_PREFIX;
|
||||
import org.junit.runner.RunWith;
|
||||
import org.junit.runners.Parameterized;
|
||||
|
||||
@ -159,8 +159,8 @@ private static void setupAndStartRM() throws Exception {
|
||||
rmconf.set(YarnConfiguration.NM_WEBAPP_SPNEGO_KEYTAB_FILE_KEY,
|
||||
httpSpnegoKeytabFile.getAbsolutePath());
|
||||
rmconf.setBoolean("mockrm.webapp.enabled", true);
|
||||
rmconf.set("yarn.resourcemanager.proxyuser.client.hosts", "*");
|
||||
rmconf.set("yarn.resourcemanager.proxyuser.client.groups", "*");
|
||||
rmconf.set(RM_PROXY_USER_PREFIX + "client.hosts", "*");
|
||||
rmconf.set(RM_PROXY_USER_PREFIX + "client.groups", "*");
|
||||
UserGroupInformation.setConfiguration(rmconf);
|
||||
rm = new MockRM(rmconf);
|
||||
rm.start();
|
||||
|
@ -104,8 +104,8 @@ private static void setupAndStartRM() throws Exception {
|
||||
rmconf.setBoolean(YarnConfiguration.YARN_ACL_ENABLE, true);
|
||||
rmconf.set(CommonConfigurationKeysPublic.HADOOP_SECURITY_AUTHENTICATION,
|
||||
"kerberos");
|
||||
rmconf.set("yarn.resourcemanager.principal", spnegoPrincipal);
|
||||
rmconf.set("yarn.resourcemanager.keytab",
|
||||
rmconf.set(YarnConfiguration.RM_PRINCIPAL, spnegoPrincipal);
|
||||
rmconf.set(YarnConfiguration.RM_KEYTAB,
|
||||
spnegoKeytabFile.getAbsolutePath());
|
||||
rmconf.setBoolean("mockrm.webapp.enabled", true);
|
||||
UserGroupInformation.setConfiguration(rmconf);
|
||||
|
@ -170,7 +170,7 @@ protected void configureServlets() {
|
||||
private static class CapTestServletModule extends TestServletModule {
|
||||
@Override
|
||||
public void configureScheduler() {
|
||||
conf.set("yarn.resourcemanager.scheduler.class",
|
||||
conf.set(YarnConfiguration.RM_SCHEDULER,
|
||||
CapacityScheduler.class.getName());
|
||||
}
|
||||
}
|
||||
@ -196,8 +196,7 @@ public void configureScheduler() {
|
||||
} catch (IOException e) {
|
||||
}
|
||||
conf.set(FairSchedulerConfiguration.ALLOCATION_FILE, FS_ALLOC_FILE);
|
||||
conf.set("yarn.resourcemanager.scheduler.class",
|
||||
FairScheduler.class.getName());
|
||||
conf.set(YarnConfiguration.RM_SCHEDULER, FairScheduler.class.getName());
|
||||
}
|
||||
}
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user