HDFS-5835. Add a new option for starting Namenode when rolling upgrade is in progress.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/HDFS-5535@1561943 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
a9110e1788
commit
917502ef31
@ -10,5 +10,8 @@ HDFS-5535 subtasks:
|
||||
|
||||
HDFS-5786. Support QUERY and FINALIZE actions of rolling upgrade. (szetszwo)
|
||||
|
||||
HDFS-5753. Add new NN startup options for downgrade and rollback using
|
||||
HDFS-5753. Add new Namenode startup options for downgrade and rollback using
|
||||
upgrade marker. (szetszwo)
|
||||
|
||||
HDFS-5835. Add a new option for starting Namenode when rolling upgrade is
|
||||
in progress. (szetszwo)
|
||||
|
@ -58,7 +58,25 @@ public boolean isFinalized() {
|
||||
public long getFinalizeTime() {
|
||||
return finalizeTime;
|
||||
}
|
||||
|
||||
|
||||
@Override
|
||||
public int hashCode() {
|
||||
//only use lower 32 bits
|
||||
return (int)startTime ^ (int)finalizeTime;
|
||||
}
|
||||
|
||||
@Override
|
||||
public boolean equals(Object obj) {
|
||||
if (obj == this) {
|
||||
return true;
|
||||
} else if (obj == null || !(obj instanceof RollingUpgradeInfo)) {
|
||||
return false;
|
||||
}
|
||||
final RollingUpgradeInfo that = (RollingUpgradeInfo)obj;
|
||||
return this.startTime == that.startTime
|
||||
&& this.finalizeTime == that.finalizeTime;
|
||||
}
|
||||
|
||||
@Override
|
||||
public String toString() {
|
||||
return " Start Time: " + (startTime == 0? "<NOT STARTED>": timestamp2String(startTime))
|
||||
|
@ -47,7 +47,7 @@ static public enum NodeType {
|
||||
|
||||
/** Startup options for rolling upgrade. */
|
||||
public static enum RollingUpgradeStartupOption{
|
||||
ROLLBACK, DOWNGRADE;
|
||||
ROLLBACK, DOWNGRADE, STARTED;
|
||||
|
||||
private static final RollingUpgradeStartupOption[] VALUES = values();
|
||||
|
||||
|
@ -79,6 +79,7 @@
|
||||
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.TimesOp;
|
||||
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.UpdateBlocksOp;
|
||||
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.UpdateMasterKeyOp;
|
||||
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.UpgradeMarkerOp;
|
||||
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.UpgradeMarkerOp.UpgradeMarkerException;
|
||||
import org.apache.hadoop.hdfs.server.namenode.INode.BlocksMapUpdateInfo;
|
||||
import org.apache.hadoop.hdfs.server.namenode.LeaseManager.Lease;
|
||||
@ -677,6 +678,12 @@ private long applyEditLogOp(FSEditLogOp op, FSDirectory fsDir,
|
||||
== RollingUpgradeStartupOption.DOWNGRADE) {
|
||||
//ignore upgrade marker
|
||||
break;
|
||||
} else if (startOpt.getRollingUpgradeStartupOption()
|
||||
== RollingUpgradeStartupOption.STARTED) {
|
||||
//rolling upgrade is already started, set info
|
||||
final UpgradeMarkerOp upgradeOp = (UpgradeMarkerOp)op;
|
||||
fsNamesys.setRollingUpgradeInfo(upgradeOp.getStartTime());
|
||||
break;
|
||||
}
|
||||
}
|
||||
throw new RollingUpgradeException(
|
||||
|
@ -3398,6 +3398,10 @@ static UpgradeMarkerOp getInstance(OpInstanceCache cache) {
|
||||
return (UpgradeMarkerOp) cache.get(OP_UPGRADE_MARKER);
|
||||
}
|
||||
|
||||
long getStartTime() {
|
||||
return startTime;
|
||||
}
|
||||
|
||||
void setStartTime(long startTime) {
|
||||
this.startTime = startTime;
|
||||
}
|
||||
|
@ -7110,7 +7110,7 @@ RollingUpgradeInfo startRollingUpgrade() throws IOException {
|
||||
final CheckpointSignature cs = getFSImage().rollEditLog();
|
||||
LOG.info("Successfully rolled edit log for preparing rolling upgrade."
|
||||
+ " Checkpoint signature: " + cs);
|
||||
rollingUpgradeInfo = new RollingUpgradeInfo(now());
|
||||
setRollingUpgradeInfo(now());
|
||||
getEditLog().logUpgradeMarker(rollingUpgradeInfo.getStartTime());
|
||||
} finally {
|
||||
writeUnlock();
|
||||
@ -7123,6 +7123,10 @@ RollingUpgradeInfo startRollingUpgrade() throws IOException {
|
||||
return rollingUpgradeInfo;
|
||||
}
|
||||
|
||||
void setRollingUpgradeInfo(long startTime) {
|
||||
rollingUpgradeInfo = new RollingUpgradeInfo(startTime);;
|
||||
}
|
||||
|
||||
RollingUpgradeInfo finalizeRollingUpgrade() throws IOException {
|
||||
checkSuperuserPrivilege();
|
||||
checkOperation(OperationCategory.WRITE);
|
||||
|
@ -914,6 +914,10 @@ private static void initNameNodeAddress(Configuration conf,
|
||||
}
|
||||
|
||||
private static String[] createArgs(StartupOption operation) {
|
||||
if (operation == StartupOption.ROLLINGUPGRADE) {
|
||||
return new String[]{operation.getName(),
|
||||
operation.getRollingUpgradeStartupOption().name()};
|
||||
}
|
||||
String[] args = (operation == null ||
|
||||
operation == StartupOption.FORMAT ||
|
||||
operation == StartupOption.REGULAR) ?
|
||||
|
@ -17,8 +17,18 @@
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import java.io.File;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.FileUtil;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hdfs.protocol.HdfsConstants.RollingUpgradeAction;
|
||||
import org.apache.hadoop.hdfs.protocol.RollingUpgradeInfo;
|
||||
import org.apache.hadoop.hdfs.qjournal.MiniJournalCluster;
|
||||
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption;
|
||||
import org.apache.hadoop.hdfs.tools.DFSAdmin;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Test;
|
||||
@ -28,6 +38,12 @@
|
||||
* This class tests rolling upgrade.
|
||||
*/
|
||||
public class TestRollingUpgrade {
|
||||
private static final Log LOG = LogFactory.getLog(TestRollingUpgrade.class);
|
||||
|
||||
private void runCmd(DFSAdmin dfsadmin, String... args) throws Exception {
|
||||
Assert.assertEquals(0, dfsadmin.run(args));
|
||||
}
|
||||
|
||||
/**
|
||||
* Test DFSAdmin Upgrade Command.
|
||||
*/
|
||||
@ -55,39 +71,23 @@ public void testDFSAdminRollingUpgradeCommands() throws Exception {
|
||||
Assert.assertTrue(dfsadmin.run(args) != 0);
|
||||
}
|
||||
|
||||
{
|
||||
//query rolling upgrade
|
||||
final String[] args = {"-rollingUpgrade"};
|
||||
Assert.assertEquals(0, dfsadmin.run(args));
|
||||
}
|
||||
//query rolling upgrade
|
||||
runCmd(dfsadmin, "-rollingUpgrade");
|
||||
|
||||
{
|
||||
//start rolling upgrade
|
||||
final String[] args = {"-rollingUpgrade", "start"};
|
||||
Assert.assertEquals(0, dfsadmin.run(args));
|
||||
}
|
||||
//start rolling upgrade
|
||||
runCmd(dfsadmin, "-rollingUpgrade", "start");
|
||||
|
||||
{
|
||||
//query rolling upgrade
|
||||
final String[] args = {"-rollingUpgrade", "query"};
|
||||
Assert.assertEquals(0, dfsadmin.run(args));
|
||||
}
|
||||
//query rolling upgrade
|
||||
runCmd(dfsadmin, "-rollingUpgrade", "query");
|
||||
|
||||
dfs.mkdirs(bar);
|
||||
|
||||
{
|
||||
//finalize rolling upgrade
|
||||
final String[] args = {"-rollingUpgrade", "finalize"};
|
||||
Assert.assertEquals(0, dfsadmin.run(args));
|
||||
}
|
||||
//finalize rolling upgrade
|
||||
runCmd(dfsadmin, "-rollingUpgrade", "finalize");
|
||||
|
||||
dfs.mkdirs(baz);
|
||||
|
||||
{
|
||||
//query rolling upgrade
|
||||
final String[] args = {"-rollingUpgrade"};
|
||||
Assert.assertEquals(0, dfsadmin.run(args));
|
||||
}
|
||||
runCmd(dfsadmin, "-rollingUpgrade");
|
||||
|
||||
Assert.assertTrue(dfs.exists(foo));
|
||||
Assert.assertTrue(dfs.exists(bar));
|
||||
@ -105,4 +105,94 @@ public void testDFSAdminRollingUpgradeCommands() throws Exception {
|
||||
if(cluster != null) cluster.shutdown();
|
||||
}
|
||||
}
|
||||
|
||||
private static Configuration setConf(Configuration conf, File dir,
|
||||
MiniJournalCluster mjc) {
|
||||
conf.set(DFSConfigKeys.DFS_NAMENODE_NAME_DIR_KEY, dir.getAbsolutePath());
|
||||
conf.set(DFSConfigKeys.DFS_NAMENODE_EDITS_DIR_KEY,
|
||||
mjc.getQuorumJournalURI("myjournal").toString());
|
||||
return conf;
|
||||
}
|
||||
|
||||
@Test (timeout = 30000)
|
||||
public void testRollingUpgradeWithQJM() throws Exception {
|
||||
String nnDirPrefix = MiniDFSCluster.getBaseDirectory() + "/nn/";
|
||||
final File nn1Dir = new File(nnDirPrefix + "image1");
|
||||
final File nn2Dir = new File(nnDirPrefix + "image2");
|
||||
|
||||
final Configuration conf = new HdfsConfiguration();
|
||||
final MiniJournalCluster mjc = new MiniJournalCluster.Builder(conf).build();
|
||||
setConf(conf, nn1Dir, mjc);
|
||||
|
||||
{
|
||||
// Start the cluster once to generate the dfs dirs
|
||||
final MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf)
|
||||
.numDataNodes(0)
|
||||
.manageNameDfsDirs(false)
|
||||
.checkExitOnShutdown(false)
|
||||
.build();
|
||||
// Shutdown the cluster before making a copy of the namenode dir to release
|
||||
// all file locks, otherwise, the copy will fail on some platforms.
|
||||
cluster.shutdown();
|
||||
}
|
||||
|
||||
MiniDFSCluster cluster2 = null;
|
||||
try {
|
||||
// Start a second NN pointed to the same quorum.
|
||||
// We need to copy the image dir from the first NN -- or else
|
||||
// the new NN will just be rejected because of Namespace mismatch.
|
||||
FileUtil.fullyDelete(nn2Dir);
|
||||
FileUtil.copy(nn1Dir, FileSystem.getLocal(conf).getRaw(),
|
||||
new Path(nn2Dir.getAbsolutePath()), false, conf);
|
||||
|
||||
// Start the cluster again
|
||||
final MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf)
|
||||
.numDataNodes(0)
|
||||
.format(false)
|
||||
.manageNameDfsDirs(false)
|
||||
.checkExitOnShutdown(false)
|
||||
.build();
|
||||
|
||||
final Path foo = new Path("/foo");
|
||||
final Path bar = new Path("/bar");
|
||||
|
||||
final RollingUpgradeInfo info1;
|
||||
{
|
||||
final DistributedFileSystem dfs = cluster.getFileSystem();
|
||||
dfs.mkdirs(foo);
|
||||
|
||||
//start rolling upgrade
|
||||
info1 = dfs.rollingUpgrade(RollingUpgradeAction.START);
|
||||
LOG.info("start " + info1);
|
||||
|
||||
//query rolling upgrade
|
||||
Assert.assertEquals(info1, dfs.rollingUpgrade(RollingUpgradeAction.QUERY));
|
||||
|
||||
dfs.mkdirs(bar);
|
||||
}
|
||||
|
||||
// cluster2 takes over QJM
|
||||
final Configuration conf2 = setConf(new Configuration(), nn2Dir, mjc);
|
||||
//set startup option to downgrade for ignoring upgrade marker in editlog
|
||||
StartupOption.ROLLINGUPGRADE.setRollingUpgradeStartupOption("started");
|
||||
cluster2 = new MiniDFSCluster.Builder(conf2)
|
||||
.numDataNodes(0)
|
||||
.format(false)
|
||||
.manageNameDfsDirs(false)
|
||||
.startupOption(StartupOption.ROLLINGUPGRADE)
|
||||
.build();
|
||||
final DistributedFileSystem dfs2 = cluster2.getFileSystem();
|
||||
|
||||
// Check that cluster2 sees the edits made on cluster1
|
||||
Assert.assertTrue(dfs2.exists(foo));
|
||||
Assert.assertTrue(dfs2.exists(bar));
|
||||
|
||||
//query rolling upgrade in cluster2
|
||||
Assert.assertEquals(info1, dfs2.rollingUpgrade(RollingUpgradeAction.QUERY));
|
||||
|
||||
LOG.info("finalize: " + dfs2.rollingUpgrade(RollingUpgradeAction.FINALIZE));
|
||||
} finally {
|
||||
if (cluster2 != null) cluster2.shutdown();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user