HDFS-6090. Use MiniDFSCluster.Builder instead of deprecated constructors. Contributed by Akira AJISAKA.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1578669 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
92317e3459
commit
0b3ee72af8
@ -409,6 +409,9 @@ Release 2.4.0 - UNRELEASED
|
||||
HDFS-6106. Reduce default for
|
||||
dfs.namenode.path.based.cache.refresh.interval.ms (cmccabe)
|
||||
|
||||
HDFS-6090. Use MiniDFSCluster.Builder instead of deprecated constructors.
|
||||
(Akira AJISAKA via jing9)
|
||||
|
||||
OPTIMIZATIONS
|
||||
|
||||
HDFS-5790. LeaseManager.findPath is very slow when many leases need recovery
|
||||
|
@ -523,7 +523,8 @@ static void runTestCache(int port) throws Exception {
|
||||
Configuration conf = new Configuration();
|
||||
MiniDFSCluster cluster = null;
|
||||
try {
|
||||
cluster = new MiniDFSCluster(port, conf, 2, true, true, null, null);
|
||||
cluster = new MiniDFSCluster.Builder(conf).nameNodePort(port)
|
||||
.numDataNodes(2).build();
|
||||
URI uri = cluster.getFileSystem().getUri();
|
||||
LOG.info("uri=" + uri);
|
||||
|
||||
|
@ -146,7 +146,8 @@ protected void setUp() throws Exception {
|
||||
fileSystem = FileSystem.getLocal(new JobConf());
|
||||
}
|
||||
else {
|
||||
dfsCluster = new MiniDFSCluster(new JobConf(), dataNodes, true, null);
|
||||
dfsCluster = new MiniDFSCluster.Builder(new JobConf())
|
||||
.numDataNodes(dataNodes).build();
|
||||
fileSystem = dfsCluster.getFileSystem();
|
||||
}
|
||||
if (localMR) {
|
||||
|
@ -51,7 +51,7 @@ public void testJobShell() throws Exception {
|
||||
try {
|
||||
Configuration conf = new Configuration();
|
||||
//start the mini mr and dfs cluster.
|
||||
dfs = new MiniDFSCluster(conf, 2 , true, null);
|
||||
dfs = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||
fs = dfs.getFileSystem();
|
||||
FSDataOutputStream stream = fs.create(testFile);
|
||||
stream.write("teststring".getBytes());
|
||||
|
@ -125,7 +125,7 @@ public void testWithDFS() throws IOException {
|
||||
|
||||
JobConf conf = new JobConf();
|
||||
conf.set(JTConfig.JT_SYSTEM_DIR, "/tmp/custom/mapred/system");
|
||||
dfs = new MiniDFSCluster(conf, 4, true, null);
|
||||
dfs = new MiniDFSCluster.Builder(conf).numDataNodes(4).build();
|
||||
fileSys = dfs.getFileSystem();
|
||||
mr = new MiniMRCluster(taskTrackers, fileSys.getUri().toString(), 1, null, null, conf);
|
||||
|
||||
|
@ -141,7 +141,8 @@ public void testLazyOutput() throws Exception {
|
||||
Configuration conf = new Configuration();
|
||||
|
||||
// Start the mini-MR and mini-DFS clusters
|
||||
dfs = new MiniDFSCluster(conf, NUM_HADOOP_SLAVES, true, null);
|
||||
dfs = new MiniDFSCluster.Builder(conf).numDataNodes(NUM_HADOOP_SLAVES)
|
||||
.build();
|
||||
fileSys = dfs.getFileSystem();
|
||||
mr = new MiniMRCluster(NUM_HADOOP_SLAVES, fileSys.getUri().toString(), 1);
|
||||
|
||||
|
@ -44,11 +44,10 @@ public class TestMRCJCFileInputFormat extends TestCase {
|
||||
MiniDFSCluster dfs = null;
|
||||
|
||||
private MiniDFSCluster newDFSCluster(JobConf conf) throws Exception {
|
||||
return new MiniDFSCluster(conf, 4, true,
|
||||
new String[]{"/rack0", "/rack0",
|
||||
"/rack1", "/rack1"},
|
||||
new String[]{"host0", "host1",
|
||||
"host2", "host3"});
|
||||
return new MiniDFSCluster.Builder(conf).numDataNodes(4)
|
||||
.racks(new String[]{"/rack0", "/rack0", "/rack1", "/rack1"})
|
||||
.hosts(new String[]{"host0", "host1", "host2", "host3"})
|
||||
.build();
|
||||
}
|
||||
|
||||
public void testLocality() throws Exception {
|
||||
@ -162,7 +161,7 @@ public void testMultiLevelInput() throws Exception {
|
||||
JobConf job = new JobConf(conf);
|
||||
|
||||
job.setBoolean("dfs.replication.considerLoad", false);
|
||||
dfs = new MiniDFSCluster(job, 1, true, rack1, hosts1);
|
||||
dfs = new MiniDFSCluster.Builder(job).racks(rack1).hosts(hosts1).build();
|
||||
dfs.waitActive();
|
||||
|
||||
String namenode = (dfs.getFileSystem()).getUri().getHost() + ":" +
|
||||
|
@ -76,7 +76,8 @@ public void testMerge() throws Exception {
|
||||
try {
|
||||
Configuration conf = new Configuration();
|
||||
// Start the mini-MR and mini-DFS clusters
|
||||
dfsCluster = new MiniDFSCluster(conf, NUM_HADOOP_DATA_NODES, true, null);
|
||||
dfsCluster = new MiniDFSCluster.Builder(conf)
|
||||
.numDataNodes(NUM_HADOOP_DATA_NODES).build();
|
||||
fileSystem = dfsCluster.getFileSystem();
|
||||
mrCluster = MiniMRClientClusterFactory.create(this.getClass(),
|
||||
NUM_HADOOP_DATA_NODES, conf);
|
||||
|
@ -324,7 +324,7 @@ public void reduce(WritableComparable key, Iterator<Writable> values,
|
||||
@BeforeClass
|
||||
public static void setup() throws IOException {
|
||||
// create configuration, dfs, file system and mapred cluster
|
||||
dfs = new MiniDFSCluster(conf, 1, true, null);
|
||||
dfs = new MiniDFSCluster.Builder(conf).build();
|
||||
fileSys = dfs.getFileSystem();
|
||||
|
||||
if (!(new File(MiniMRYarnCluster.APPJAR)).exists()) {
|
||||
|
@ -169,7 +169,7 @@ public void testClassPath() throws IOException {
|
||||
final int jobTrackerPort = 60050;
|
||||
|
||||
Configuration conf = new Configuration();
|
||||
dfs = new MiniDFSCluster(conf, 1, true, null);
|
||||
dfs = new MiniDFSCluster.Builder(conf).build();
|
||||
fileSys = dfs.getFileSystem();
|
||||
namenode = fileSys.getUri().toString();
|
||||
mr = new MiniMRCluster(taskTrackers, namenode, 3);
|
||||
@ -201,7 +201,7 @@ public void testExternalWritable()
|
||||
final int taskTrackers = 4;
|
||||
|
||||
Configuration conf = new Configuration();
|
||||
dfs = new MiniDFSCluster(conf, 1, true, null);
|
||||
dfs = new MiniDFSCluster.Builder(conf).build();
|
||||
fileSys = dfs.getFileSystem();
|
||||
namenode = fileSys.getUri().toString();
|
||||
mr = new MiniMRCluster(taskTrackers, namenode, 3);
|
||||
|
@ -38,7 +38,7 @@ public void testWithDFS() throws IOException {
|
||||
FileSystem fileSys = null;
|
||||
try {
|
||||
JobConf conf = new JobConf();
|
||||
dfs = new MiniDFSCluster(conf, 1, true, null);
|
||||
dfs = new MiniDFSCluster.Builder(conf).build();
|
||||
fileSys = dfs.getFileSystem();
|
||||
mr = new MiniMRCluster(2, fileSys.getUri().toString(), 4);
|
||||
MRCaching.setupCache("/cachedir", fileSys);
|
||||
|
@ -75,7 +75,7 @@ public RunningJob run() throws IOException {
|
||||
|
||||
@Before
|
||||
public void setUp() throws Exception {
|
||||
dfs = new MiniDFSCluster(conf, 4, true, null);
|
||||
dfs = new MiniDFSCluster.Builder(conf).numDataNodes(4).build();
|
||||
|
||||
fs = DFS_UGI.doAs(new PrivilegedExceptionAction<FileSystem>() {
|
||||
public FileSystem run() throws IOException {
|
||||
|
@ -92,8 +92,8 @@ private void testCachingAtLevel(int level) throws Exception {
|
||||
String rack2 = getRack(1, level);
|
||||
Configuration conf = new Configuration();
|
||||
// Run a datanode on host1 under /a/b/c/..../d1/e1/f1
|
||||
dfs = new MiniDFSCluster(conf, 1, true, new String[] {rack1},
|
||||
new String[] {"host1.com"});
|
||||
dfs = new MiniDFSCluster.Builder(conf).racks(new String[] {rack1})
|
||||
.hosts(new String[] {"host1.com"}).build();
|
||||
dfs.waitActive();
|
||||
fileSys = dfs.getFileSystem();
|
||||
if (!fileSys.mkdirs(inDir)) {
|
||||
|
@ -57,7 +57,7 @@ public static Test suite() {
|
||||
TestSetup setup = new TestSetup(mySuite) {
|
||||
protected void setUp() throws Exception {
|
||||
Configuration conf = new Configuration();
|
||||
dfsCluster = new MiniDFSCluster(conf, 2, true, null);
|
||||
dfsCluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||
mrCluster = new MiniMRCluster(2,
|
||||
dfsCluster.getFileSystem().getUri().toString(), 1);
|
||||
}
|
||||
|
@ -106,7 +106,7 @@ public void testJobWithDFS() throws IOException {
|
||||
final int taskTrackers = 4;
|
||||
final int jobTrackerPort = 60050;
|
||||
Configuration conf = new Configuration();
|
||||
dfs = new MiniDFSCluster(conf, 1, true, null);
|
||||
dfs = new MiniDFSCluster.Builder(conf).build();
|
||||
fileSys = dfs.getFileSystem();
|
||||
namenode = fileSys.getUri().toString();
|
||||
mr = new MiniMRCluster(taskTrackers, namenode, 2);
|
||||
|
@ -62,7 +62,7 @@ public static Test suite() {
|
||||
TestSetup setup = new TestSetup(new TestSuite(TestDatamerge.class)) {
|
||||
protected void setUp() throws Exception {
|
||||
Configuration conf = new Configuration();
|
||||
cluster = new MiniDFSCluster(conf, 2, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||
}
|
||||
protected void tearDown() throws Exception {
|
||||
if (cluster != null) {
|
||||
|
@ -39,9 +39,10 @@ public void testSplitting() throws Exception {
|
||||
JobConf conf = new JobConf();
|
||||
MiniDFSCluster dfs = null;
|
||||
try {
|
||||
dfs = new MiniDFSCluster(conf, 4, true, new String[] { "/rack0",
|
||||
"/rack0", "/rack1", "/rack1" }, new String[] { "host0", "host1",
|
||||
"host2", "host3" });
|
||||
dfs = new MiniDFSCluster.Builder(conf).numDataNodes(4)
|
||||
.racks(new String[] { "/rack0", "/rack0", "/rack1", "/rack1" })
|
||||
.hosts(new String[] { "host0", "host1", "host2", "host3" })
|
||||
.build();
|
||||
FileSystem fs = dfs.getFileSystem();
|
||||
|
||||
Path path = getPath("/foo/bar", fs);
|
||||
|
@ -79,7 +79,7 @@ public void testPipes() throws IOException {
|
||||
try {
|
||||
final int numSlaves = 2;
|
||||
Configuration conf = new Configuration();
|
||||
dfs = new MiniDFSCluster(conf, numSlaves, true, null);
|
||||
dfs = new MiniDFSCluster.Builder(conf).numDataNodes(numSlaves).build();
|
||||
mr = new MiniMRCluster(numSlaves, dfs.getFileSystem().getUri().toString(), 1);
|
||||
writeInputFile(dfs.getFileSystem(), inputPath);
|
||||
runProgram(mr, dfs, wordCountSimple,
|
||||
|
@ -151,8 +151,8 @@ private void sleepForever() {
|
||||
public void start() throws IOException, FileNotFoundException,
|
||||
URISyntaxException {
|
||||
if (!noDFS) {
|
||||
dfs = new MiniDFSCluster(nnPort, conf, numDataNodes, true, true,
|
||||
dfsOpts, null, null);
|
||||
dfs = new MiniDFSCluster.Builder(conf).nameNodePort(nnPort)
|
||||
.numDataNodes(numDataNodes).startupOption(dfsOpts).build();
|
||||
LOG.info("Started MiniDFSCluster -- namenode on port "
|
||||
+ dfs.getNameNodePort());
|
||||
}
|
||||
|
@ -131,7 +131,8 @@ public void testLazyOutput() throws Exception {
|
||||
Configuration conf = new Configuration();
|
||||
|
||||
// Start the mini-MR and mini-DFS clusters
|
||||
dfs = new MiniDFSCluster(conf, NUM_HADOOP_SLAVES, true, null);
|
||||
dfs = new MiniDFSCluster.Builder(conf).numDataNodes(NUM_HADOOP_SLAVES)
|
||||
.build();
|
||||
fileSys = dfs.getFileSystem();
|
||||
mr = new MiniMRCluster(NUM_HADOOP_SLAVES, fileSys.getUri().toString(), 1);
|
||||
|
||||
|
@ -313,7 +313,8 @@ public void testSplitPlacement() throws Exception {
|
||||
*/
|
||||
Configuration conf = new Configuration();
|
||||
conf.setBoolean("dfs.replication.considerLoad", false);
|
||||
dfs = new MiniDFSCluster(conf, 1, true, rack1, hosts1);
|
||||
dfs = new MiniDFSCluster.Builder(conf).racks(rack1).hosts(hosts1)
|
||||
.build();
|
||||
dfs.waitActive();
|
||||
|
||||
fileSys = dfs.getFileSystem();
|
||||
@ -855,7 +856,8 @@ public void testSplitPlacementForCompressedFiles() throws Exception {
|
||||
*/
|
||||
Configuration conf = new Configuration();
|
||||
conf.setBoolean("dfs.replication.considerLoad", false);
|
||||
dfs = new MiniDFSCluster(conf, 1, true, rack1, hosts1);
|
||||
dfs = new MiniDFSCluster.Builder(conf).racks(rack1).hosts(hosts1)
|
||||
.build();
|
||||
dfs.waitActive();
|
||||
|
||||
fileSys = dfs.getFileSystem();
|
||||
@ -1197,7 +1199,8 @@ public void testMissingBlocks() throws Exception {
|
||||
Configuration conf = new Configuration();
|
||||
conf.set("fs.hdfs.impl", MissingBlockFileSystem.class.getName());
|
||||
conf.setBoolean("dfs.replication.considerLoad", false);
|
||||
dfs = new MiniDFSCluster(conf, 1, true, rack1, hosts1);
|
||||
dfs = new MiniDFSCluster.Builder(conf).racks(rack1).hosts(hosts1)
|
||||
.build();
|
||||
dfs.waitActive();
|
||||
|
||||
namenode = (dfs.getFileSystem()).getUri().getHost() + ":" +
|
||||
|
@ -37,9 +37,10 @@ public void testSplitting() throws Exception {
|
||||
Job job = Job.getInstance();
|
||||
MiniDFSCluster dfs = null;
|
||||
try {
|
||||
dfs = new MiniDFSCluster(job.getConfiguration(), 4, true, new String[] { "/rack0",
|
||||
"/rack0", "/rack1", "/rack1" }, new String[] { "host0", "host1",
|
||||
"host2", "host3" });
|
||||
dfs = new MiniDFSCluster.Builder(job.getConfiguration()).numDataNodes(4)
|
||||
.racks(new String[] { "/rack0", "/rack0", "/rack1", "/rack1" })
|
||||
.hosts(new String[] { "host0", "host1", "host2", "host3" })
|
||||
.build();
|
||||
FileSystem fs = dfs.getFileSystem();
|
||||
|
||||
Path path = getPath("/foo/bar", fs);
|
||||
|
@ -45,7 +45,7 @@ public static Test suite() {
|
||||
TestSetup setup = new TestSetup(new TestSuite(TestJoinDatamerge.class)) {
|
||||
protected void setUp() throws Exception {
|
||||
Configuration conf = new Configuration();
|
||||
cluster = new MiniDFSCluster(conf, 2, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||
}
|
||||
protected void tearDown() throws Exception {
|
||||
if (cluster != null) {
|
||||
|
@ -50,7 +50,7 @@ public static Test suite() {
|
||||
TestSetup setup = new TestSetup(new TestSuite(TestJoinProperties.class)) {
|
||||
protected void setUp() throws Exception {
|
||||
Configuration conf = new Configuration();
|
||||
cluster = new MiniDFSCluster(conf, 2, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||
base = cluster.getFileSystem().makeQualified(new Path("/nested"));
|
||||
src = generateSources(conf);
|
||||
}
|
||||
|
@ -58,7 +58,8 @@ public class TestMRCredentials {
|
||||
public static void setUp() throws Exception {
|
||||
System.setProperty("hadoop.log.dir", "logs");
|
||||
Configuration conf = new Configuration();
|
||||
dfsCluster = new MiniDFSCluster(conf, numSlaves, true, null);
|
||||
dfsCluster = new MiniDFSCluster.Builder(conf).numDataNodes(numSlaves)
|
||||
.build();
|
||||
jConf = new JobConf(conf);
|
||||
FileSystem.setDefaultUri(conf, dfsCluster.getFileSystem().getUri().toString());
|
||||
mrCluster = MiniMRClientClusterFactory.create(TestMRCredentials.class, 1, jConf);
|
||||
|
@ -92,7 +92,7 @@ private void startCluster(Configuration conf) throws Exception {
|
||||
YarnConfiguration.DEFAULT_YARN_CROSS_PLATFORM_APPLICATION_CLASSPATH))
|
||||
+ File.pathSeparator + classpathDir;
|
||||
conf.set(YarnConfiguration.YARN_APPLICATION_CLASSPATH, cp);
|
||||
dfsCluster = new MiniDFSCluster(conf, 1, true, null);
|
||||
dfsCluster = new MiniDFSCluster.Builder(conf).build();
|
||||
FileSystem fileSystem = dfsCluster.getFileSystem();
|
||||
fileSystem.mkdirs(new Path("/tmp"));
|
||||
fileSystem.mkdirs(new Path("/user"));
|
||||
|
@ -70,7 +70,8 @@ protected void setUp() throws Exception {
|
||||
UserGroupInformation.createUserForTesting("u1", userGroups);
|
||||
UserGroupInformation.createUserForTesting("u2", new String[]{"gg"});
|
||||
|
||||
dfsCluster = new MiniDFSCluster(conf, dataNodes, true, null);
|
||||
dfsCluster = new MiniDFSCluster.Builder(conf).numDataNodes(dataNodes)
|
||||
.build();
|
||||
FileSystem fileSystem = dfsCluster.getFileSystem();
|
||||
fileSystem.mkdirs(new Path("/tmp"));
|
||||
fileSystem.mkdirs(new Path("/user"));
|
||||
|
@ -57,7 +57,8 @@ protected void setUp() throws Exception {
|
||||
conf.set("dfs.permissions", "true");
|
||||
conf.set("hadoop.security.authentication", "simple");
|
||||
|
||||
dfsCluster = new MiniDFSCluster(conf, dataNodes, true, null);
|
||||
dfsCluster = new MiniDFSCluster.Builder(conf).numDataNodes(dataNodes)
|
||||
.build();
|
||||
FileSystem fileSystem = dfsCluster.getFileSystem();
|
||||
fileSystem.mkdirs(new Path("/tmp"));
|
||||
fileSystem.mkdirs(new Path("/user"));
|
||||
|
@ -43,7 +43,7 @@ public static Test suite() {
|
||||
TestSetup setup = new TestSetup(new TestSuite(TestDataJoin.class)) {
|
||||
protected void setUp() throws Exception {
|
||||
Configuration conf = new Configuration();
|
||||
cluster = new MiniDFSCluster(conf, 2, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||
}
|
||||
protected void tearDown() throws Exception {
|
||||
if (cluster != null) {
|
||||
|
@ -49,7 +49,7 @@ public class TestGlobbedCopyListing {
|
||||
|
||||
@BeforeClass
|
||||
public static void setup() throws Exception {
|
||||
cluster = new MiniDFSCluster(new Configuration(), 1, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(new Configuration()).build();
|
||||
createSourceData();
|
||||
}
|
||||
|
||||
|
@ -280,7 +280,7 @@ public void testCopyFromDfsToDfs() throws Exception {
|
||||
MiniDFSCluster cluster = null;
|
||||
try {
|
||||
Configuration conf = new Configuration();
|
||||
cluster = new MiniDFSCluster(conf, 2, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||
final FileSystem hdfs = cluster.getFileSystem();
|
||||
namenode = FileSystem.getDefaultUri(conf).toString();
|
||||
if (namenode.startsWith("hdfs://")) {
|
||||
@ -310,7 +310,7 @@ public void testEmptyDir() throws Exception {
|
||||
MiniDFSCluster cluster = null;
|
||||
try {
|
||||
Configuration conf = new Configuration();
|
||||
cluster = new MiniDFSCluster(conf, 2, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||
final FileSystem hdfs = cluster.getFileSystem();
|
||||
namenode = FileSystem.getDefaultUri(conf).toString();
|
||||
if (namenode.startsWith("hdfs://")) {
|
||||
@ -340,7 +340,7 @@ public void testCopyFromLocalToDfs() throws Exception {
|
||||
MiniDFSCluster cluster = null;
|
||||
try {
|
||||
Configuration conf = new Configuration();
|
||||
cluster = new MiniDFSCluster(conf, 1, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(conf).build();
|
||||
final FileSystem hdfs = cluster.getFileSystem();
|
||||
final String namenode = hdfs.getUri().toString();
|
||||
if (namenode.startsWith("hdfs://")) {
|
||||
@ -369,7 +369,7 @@ public void testCopyFromDfsToLocal() throws Exception {
|
||||
try {
|
||||
Configuration conf = new Configuration();
|
||||
final FileSystem localfs = FileSystem.get(LOCAL_FS, conf);
|
||||
cluster = new MiniDFSCluster(conf, 1, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(conf).build();
|
||||
final FileSystem hdfs = cluster.getFileSystem();
|
||||
final String namenode = FileSystem.getDefaultUri(conf).toString();
|
||||
if (namenode.startsWith("hdfs://")) {
|
||||
@ -396,7 +396,7 @@ public void testCopyDfsToDfsUpdateOverwrite() throws Exception {
|
||||
MiniDFSCluster cluster = null;
|
||||
try {
|
||||
Configuration conf = new Configuration();
|
||||
cluster = new MiniDFSCluster(conf, 2, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||
final FileSystem hdfs = cluster.getFileSystem();
|
||||
final String namenode = hdfs.getUri().toString();
|
||||
if (namenode.startsWith("hdfs://")) {
|
||||
@ -456,7 +456,7 @@ public void testCopyDfsToDfsUpdateWithSkipCRC() throws Exception {
|
||||
MiniDFSCluster cluster = null;
|
||||
try {
|
||||
Configuration conf = new Configuration();
|
||||
cluster = new MiniDFSCluster(conf, 2, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||
final FileSystem hdfs = cluster.getFileSystem();
|
||||
final String namenode = hdfs.getUri().toString();
|
||||
|
||||
@ -614,7 +614,7 @@ public void testBasedir() throws Exception {
|
||||
MiniDFSCluster cluster = null;
|
||||
try {
|
||||
Configuration conf = new Configuration();
|
||||
cluster = new MiniDFSCluster(conf, 2, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||
final FileSystem hdfs = cluster.getFileSystem();
|
||||
namenode = FileSystem.getDefaultUri(conf).toString();
|
||||
if (namenode.startsWith("hdfs://")) {
|
||||
@ -639,7 +639,7 @@ public void testPreserveOption() throws Exception {
|
||||
Configuration conf = new Configuration();
|
||||
MiniDFSCluster cluster = null;
|
||||
try {
|
||||
cluster = new MiniDFSCluster(conf, 2, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||
String nnUri = FileSystem.getDefaultUri(conf).toString();
|
||||
FileSystem fs = FileSystem.get(URI.create(nnUri), conf);
|
||||
|
||||
@ -791,7 +791,7 @@ public void testLimits() throws Exception {
|
||||
Configuration conf = new Configuration();
|
||||
MiniDFSCluster cluster = null;
|
||||
try {
|
||||
cluster = new MiniDFSCluster(conf, 2, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||
final String nnUri = FileSystem.getDefaultUri(conf).toString();
|
||||
final FileSystem fs = FileSystem.get(URI.create(nnUri), conf);
|
||||
final DistCpV1 distcp = new DistCpV1(conf);
|
||||
@ -899,7 +899,7 @@ public void testHftpAccessControl() throws Exception {
|
||||
|
||||
//start cluster by DFS_UGI
|
||||
final Configuration dfsConf = new Configuration();
|
||||
cluster = new MiniDFSCluster(dfsConf, 2, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(dfsConf).numDataNodes(2).build();
|
||||
cluster.waitActive();
|
||||
|
||||
final String httpAdd = dfsConf.get("dfs.http.address");
|
||||
@ -955,7 +955,7 @@ public void testDelete() throws Exception {
|
||||
conf.setInt("fs.trash.interval", 60);
|
||||
MiniDFSCluster cluster = null;
|
||||
try {
|
||||
cluster = new MiniDFSCluster(conf, 2, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||
final URI nnURI = FileSystem.getDefaultUri(conf);
|
||||
final String nnUri = nnURI.toString();
|
||||
final FileSystem fs = FileSystem.get(URI.create(nnUri), conf);
|
||||
@ -1027,7 +1027,7 @@ public void testDeleteLocal() throws Exception {
|
||||
try {
|
||||
Configuration conf = new Configuration();
|
||||
final FileSystem localfs = FileSystem.get(LOCAL_FS, conf);
|
||||
cluster = new MiniDFSCluster(conf, 1, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(conf).build();
|
||||
final FileSystem hdfs = cluster.getFileSystem();
|
||||
final String namenode = FileSystem.getDefaultUri(conf).toString();
|
||||
if (namenode.startsWith("hdfs://")) {
|
||||
@ -1060,7 +1060,7 @@ public void testGlobbing() throws Exception {
|
||||
MiniDFSCluster cluster = null;
|
||||
try {
|
||||
Configuration conf = new Configuration();
|
||||
cluster = new MiniDFSCluster(conf, 2, true, null);
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||
final FileSystem hdfs = cluster.getFileSystem();
|
||||
namenode = FileSystem.getDefaultUri(conf).toString();
|
||||
if (namenode.startsWith("hdfs://")) {
|
||||
|
@ -39,7 +39,8 @@ public class TestDumpTypedBytes {
|
||||
@Test
|
||||
public void testDumping() throws Exception {
|
||||
Configuration conf = new Configuration();
|
||||
MiniDFSCluster cluster = new MiniDFSCluster(conf, 2, true, null);
|
||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2)
|
||||
.build();
|
||||
FileSystem fs = cluster.getFileSystem();
|
||||
PrintStream psBackup = System.out;
|
||||
ByteArrayOutputStream out = new ByteArrayOutputStream();
|
||||
|
@ -54,7 +54,7 @@ public TestFileArgs() throws IOException
|
||||
{
|
||||
// Set up mini cluster
|
||||
conf = new Configuration();
|
||||
dfs = new MiniDFSCluster(conf, 1, true, null);
|
||||
dfs = new MiniDFSCluster.Builder(conf).build();
|
||||
fileSys = dfs.getFileSystem();
|
||||
namenode = fileSys.getUri().getAuthority();
|
||||
mr = new MiniMRCluster(1, namenode, 1);
|
||||
|
@ -39,7 +39,8 @@ public class TestLoadTypedBytes {
|
||||
@Test
|
||||
public void testLoading() throws Exception {
|
||||
Configuration conf = new Configuration();
|
||||
MiniDFSCluster cluster = new MiniDFSCluster(conf, 2, true, null);
|
||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2)
|
||||
.build();
|
||||
FileSystem fs = cluster.getFileSystem();
|
||||
|
||||
ByteArrayOutputStream out = new ByteArrayOutputStream();
|
||||
|
@ -65,7 +65,7 @@ public TestMultipleArchiveFiles() throws Exception {
|
||||
input = "HADOOP";
|
||||
expectedOutput = "HADOOP\t\nHADOOP\t\n";
|
||||
conf = new Configuration();
|
||||
dfs = new MiniDFSCluster(conf, 1, true, null);
|
||||
dfs = new MiniDFSCluster.Builder(conf).build();
|
||||
fileSys = dfs.getFileSystem();
|
||||
namenode = fileSys.getUri().getAuthority();
|
||||
mr = new MiniMRCluster(1, namenode, 1);
|
||||
|
@ -69,7 +69,7 @@ public void testMultipleCachefiles() throws Exception
|
||||
MiniDFSCluster dfs = null;
|
||||
try{
|
||||
Configuration conf = new Configuration();
|
||||
dfs = new MiniDFSCluster(conf, 1, true, null);
|
||||
dfs = new MiniDFSCluster.Builder(conf).build();
|
||||
FileSystem fileSys = dfs.getFileSystem();
|
||||
String namenode = fileSys.getUri().toString();
|
||||
|
||||
|
@ -61,7 +61,7 @@ public void testSymLink() throws Exception
|
||||
MiniDFSCluster dfs = null;
|
||||
try {
|
||||
Configuration conf = new Configuration();
|
||||
dfs = new MiniDFSCluster(conf, 1, true, null);
|
||||
dfs = new MiniDFSCluster.Builder(conf).build();
|
||||
FileSystem fileSys = dfs.getFileSystem();
|
||||
String namenode = fileSys.getUri().toString();
|
||||
mr = new MiniMRCluster(1, namenode, 3);
|
||||
|
Loading…
Reference in New Issue
Block a user