HDFS-3695. Genericize format() to non-file JournalManagers. Contributed by Todd Lipcon.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1371513 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
0733adf670
commit
9d0f8792a9
@ -112,6 +112,8 @@ Trunk (unreleased changes)
|
|||||||
HDFS-3768. Exception in TestJettyHelper is incorrect.
|
HDFS-3768. Exception in TestJettyHelper is incorrect.
|
||||||
(Eli Reisman via jghoman)
|
(Eli Reisman via jghoman)
|
||||||
|
|
||||||
|
HDFS-3695. Genericize format() to non-file JournalManagers. (todd)
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
BUG FIXES
|
BUG FIXES
|
||||||
|
@ -271,6 +271,23 @@ public void processResult(int rc, String path, Object ctx, String name) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void format(NamespaceInfo ns) {
|
||||||
|
// Currently, BKJM automatically formats itself when first accessed.
|
||||||
|
// TODO: change over to explicit formatting so that the admin can
|
||||||
|
// clear out the BK storage when reformatting a cluster.
|
||||||
|
LOG.info("Not formatting " + this + " - BKJM does not currently " +
|
||||||
|
"support reformatting. If it has not been used before, it will" +
|
||||||
|
"be formatted automatically upon first use.");
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean hasSomeData() throws IOException {
|
||||||
|
// Don't confirm format on BKJM, since format() is currently a
|
||||||
|
// no-op anyway
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Start a new log segment in a BookKeeper ledger.
|
* Start a new log segment in a BookKeeper ledger.
|
||||||
* First ensure that we have the write lock for this journal.
|
* First ensure that we have the write lock for this journal.
|
||||||
|
@ -39,6 +39,7 @@
|
|||||||
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NodeType;
|
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NodeType;
|
||||||
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption;
|
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption;
|
||||||
import org.apache.hadoop.fs.FileUtil;
|
import org.apache.hadoop.fs.FileUtil;
|
||||||
|
import org.apache.hadoop.util.ToolRunner;
|
||||||
import org.apache.hadoop.util.VersionInfo;
|
import org.apache.hadoop.util.VersionInfo;
|
||||||
|
|
||||||
|
|
||||||
@ -219,7 +220,7 @@ public String listStorageDirectories() {
|
|||||||
* One of the storage directories.
|
* One of the storage directories.
|
||||||
*/
|
*/
|
||||||
@InterfaceAudience.Private
|
@InterfaceAudience.Private
|
||||||
public static class StorageDirectory {
|
public static class StorageDirectory implements FormatConfirmable {
|
||||||
final File root; // root directory
|
final File root; // root directory
|
||||||
final boolean useLock; // flag to enable storage lock
|
final boolean useLock; // flag to enable storage lock
|
||||||
final StorageDirType dirType; // storage dir type
|
final StorageDirType dirType; // storage dir type
|
||||||
@ -575,6 +576,32 @@ public void doRecover(StorageState curState) throws IOException {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @return true if the storage directory should prompt the user prior
|
||||||
|
* to formatting (i.e if the directory appears to contain some data)
|
||||||
|
* @throws IOException if the SD cannot be accessed due to an IO error
|
||||||
|
*/
|
||||||
|
@Override
|
||||||
|
public boolean hasSomeData() throws IOException {
|
||||||
|
// Its alright for a dir not to exist, or to exist (properly accessible)
|
||||||
|
// and be completely empty.
|
||||||
|
if (!root.exists()) return false;
|
||||||
|
|
||||||
|
if (!root.isDirectory()) {
|
||||||
|
// a file where you expect a directory should not cause silent
|
||||||
|
// formatting
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (FileUtil.listFiles(root).length == 0) {
|
||||||
|
// Empty dir can format without prompt.
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Lock storage to provide exclusive access.
|
* Lock storage to provide exclusive access.
|
||||||
*
|
*
|
||||||
@ -767,6 +794,68 @@ public static void checkVersionUpgradable(int oldVersion)
|
|||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Iterate over each of the {@link FormatConfirmable} objects,
|
||||||
|
* potentially checking with the user whether it should be formatted.
|
||||||
|
*
|
||||||
|
* If running in interactive mode, will prompt the user for each
|
||||||
|
* directory to allow them to format anyway. Otherwise, returns
|
||||||
|
* false, unless 'force' is specified.
|
||||||
|
*
|
||||||
|
* @param force format regardless of whether dirs exist
|
||||||
|
* @param interactive prompt the user when a dir exists
|
||||||
|
* @return true if formatting should proceed
|
||||||
|
* @throws IOException if some storage cannot be accessed
|
||||||
|
*/
|
||||||
|
public static boolean confirmFormat(
|
||||||
|
Iterable<? extends FormatConfirmable> items,
|
||||||
|
boolean force, boolean interactive) throws IOException {
|
||||||
|
for (FormatConfirmable item : items) {
|
||||||
|
if (!item.hasSomeData())
|
||||||
|
continue;
|
||||||
|
if (force) { // Don't confirm, always format.
|
||||||
|
System.err.println(
|
||||||
|
"Data exists in " + item + ". Formatting anyway.");
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
if (!interactive) { // Don't ask - always don't format
|
||||||
|
System.err.println(
|
||||||
|
"Running in non-interactive mode, and data appears to exist in " +
|
||||||
|
item + ". Not formatting.");
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
if (!ToolRunner.confirmPrompt("Re-format filesystem in " + item + " ?")) {
|
||||||
|
System.err.println("Format aborted in " + item);
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Interface for classes which need to have the user confirm their
|
||||||
|
* formatting during NameNode -format and other similar operations.
|
||||||
|
*
|
||||||
|
* This is currently a storage directory or journal manager.
|
||||||
|
*/
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
public interface FormatConfirmable {
|
||||||
|
/**
|
||||||
|
* @return true if the storage seems to have some valid data in it,
|
||||||
|
* and the user should be required to confirm the format. Otherwise,
|
||||||
|
* false.
|
||||||
|
* @throws IOException if the storage cannot be accessed at all.
|
||||||
|
*/
|
||||||
|
public boolean hasSomeData() throws IOException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @return a string representation of the formattable item, suitable
|
||||||
|
* for display to the user inside a prompt
|
||||||
|
*/
|
||||||
|
public String toString();
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Get common storage fields.
|
* Get common storage fields.
|
||||||
* Should be overloaded if additional fields need to be get.
|
* Should be overloaded if additional fields need to be get.
|
||||||
|
@ -22,6 +22,7 @@
|
|||||||
|
|
||||||
import org.apache.hadoop.hdfs.server.protocol.JournalInfo;
|
import org.apache.hadoop.hdfs.server.protocol.JournalInfo;
|
||||||
import org.apache.hadoop.hdfs.server.protocol.NamenodeRegistration;
|
import org.apache.hadoop.hdfs.server.protocol.NamenodeRegistration;
|
||||||
|
import org.apache.hadoop.hdfs.server.protocol.NamespaceInfo;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* A JournalManager implementation that uses RPCs to log transactions
|
* A JournalManager implementation that uses RPCs to log transactions
|
||||||
@ -38,6 +39,20 @@ class BackupJournalManager implements JournalManager {
|
|||||||
this.bnReg = bnReg;
|
this.bnReg = bnReg;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void format(NamespaceInfo nsInfo) {
|
||||||
|
// format() should only get called at startup, before any BNs
|
||||||
|
// can register with the NN.
|
||||||
|
throw new UnsupportedOperationException(
|
||||||
|
"BackupNode journal should never get formatted");
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean hasSomeData() {
|
||||||
|
throw new UnsupportedOperationException();
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public EditLogOutputStream startLogSegment(long txId) throws IOException {
|
public EditLogOutputStream startLogSegment(long txId) throws IOException {
|
||||||
EditLogBackupOutputStream stm = new EditLogBackupOutputStream(bnReg,
|
EditLogBackupOutputStream stm = new EditLogBackupOutputStream(bnReg,
|
||||||
|
@ -41,6 +41,7 @@
|
|||||||
import static org.apache.hadoop.util.ExitUtil.terminate;
|
import static org.apache.hadoop.util.ExitUtil.terminate;
|
||||||
|
|
||||||
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NamenodeRole;
|
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NamenodeRole;
|
||||||
|
import org.apache.hadoop.hdfs.server.common.Storage.FormatConfirmable;
|
||||||
import org.apache.hadoop.hdfs.server.common.Storage.StorageDirectory;
|
import org.apache.hadoop.hdfs.server.common.Storage.StorageDirectory;
|
||||||
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.AddOp;
|
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.AddOp;
|
||||||
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.CancelDelegationTokenOp;
|
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.CancelDelegationTokenOp;
|
||||||
@ -333,6 +334,39 @@ synchronized void close() {
|
|||||||
state = State.CLOSED;
|
state = State.CLOSED;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Format all configured journals which are not file-based.
|
||||||
|
*
|
||||||
|
* File-based journals are skipped, since they are formatted by the
|
||||||
|
* Storage format code.
|
||||||
|
*/
|
||||||
|
void formatNonFileJournals(NamespaceInfo nsInfo) throws IOException {
|
||||||
|
Preconditions.checkState(state == State.BETWEEN_LOG_SEGMENTS,
|
||||||
|
"Bad state: %s", state);
|
||||||
|
|
||||||
|
for (JournalManager jm : journalSet.getJournalManagers()) {
|
||||||
|
if (!(jm instanceof FileJournalManager)) {
|
||||||
|
jm.format(nsInfo);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
List<FormatConfirmable> getFormatConfirmables() {
|
||||||
|
Preconditions.checkState(state == State.BETWEEN_LOG_SEGMENTS,
|
||||||
|
"Bad state: %s", state);
|
||||||
|
|
||||||
|
List<FormatConfirmable> ret = Lists.newArrayList();
|
||||||
|
for (final JournalManager jm : journalSet.getJournalManagers()) {
|
||||||
|
// The FJMs are confirmed separately since they are also
|
||||||
|
// StorageDirectories
|
||||||
|
if (!(jm instanceof FileJournalManager)) {
|
||||||
|
ret.add(jm);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Write an operation to the edit log. Do not sync to persistent
|
* Write an operation to the edit log. Do not sync to persistent
|
||||||
* store yet.
|
* store yet.
|
||||||
|
@ -40,10 +40,12 @@
|
|||||||
import org.apache.hadoop.hdfs.protocol.LayoutVersion.Feature;
|
import org.apache.hadoop.hdfs.protocol.LayoutVersion.Feature;
|
||||||
import org.apache.hadoop.hdfs.server.common.InconsistentFSStateException;
|
import org.apache.hadoop.hdfs.server.common.InconsistentFSStateException;
|
||||||
import org.apache.hadoop.hdfs.server.common.Storage;
|
import org.apache.hadoop.hdfs.server.common.Storage;
|
||||||
|
import org.apache.hadoop.hdfs.server.common.Storage.FormatConfirmable;
|
||||||
import org.apache.hadoop.hdfs.server.common.Storage.StorageDirectory;
|
import org.apache.hadoop.hdfs.server.common.Storage.StorageDirectory;
|
||||||
import org.apache.hadoop.hdfs.server.common.Storage.StorageState;
|
import org.apache.hadoop.hdfs.server.common.Storage.StorageState;
|
||||||
import org.apache.hadoop.hdfs.server.common.Util;
|
import org.apache.hadoop.hdfs.server.common.Util;
|
||||||
import static org.apache.hadoop.util.Time.now;
|
import static org.apache.hadoop.util.Time.now;
|
||||||
|
|
||||||
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NamenodeRole;
|
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NamenodeRole;
|
||||||
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption;
|
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption;
|
||||||
|
|
||||||
@ -139,10 +141,33 @@ void format(FSNamesystem fsn, String clusterId) throws IOException {
|
|||||||
fileCount + " files");
|
fileCount + " files");
|
||||||
NamespaceInfo ns = NNStorage.newNamespaceInfo();
|
NamespaceInfo ns = NNStorage.newNamespaceInfo();
|
||||||
ns.clusterID = clusterId;
|
ns.clusterID = clusterId;
|
||||||
|
|
||||||
storage.format(ns);
|
storage.format(ns);
|
||||||
|
editLog.formatNonFileJournals(ns);
|
||||||
saveFSImageInAllDirs(fsn, 0);
|
saveFSImageInAllDirs(fsn, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Check whether the storage directories and non-file journals exist.
|
||||||
|
* If running in interactive mode, will prompt the user for each
|
||||||
|
* directory to allow them to format anyway. Otherwise, returns
|
||||||
|
* false, unless 'force' is specified.
|
||||||
|
*
|
||||||
|
* @param force format regardless of whether dirs exist
|
||||||
|
* @param interactive prompt the user when a dir exists
|
||||||
|
* @return true if formatting should proceed
|
||||||
|
* @throws IOException if some storage cannot be accessed
|
||||||
|
*/
|
||||||
|
boolean confirmFormat(boolean force, boolean interactive) throws IOException {
|
||||||
|
List<FormatConfirmable> confirms = Lists.newArrayList();
|
||||||
|
for (StorageDirectory sd : storage.dirIterable(null)) {
|
||||||
|
confirms.add(sd);
|
||||||
|
}
|
||||||
|
|
||||||
|
confirms.addAll(editLog.getFormatConfirmables());
|
||||||
|
return Storage.confirmFormat(confirms, force, interactive);
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Analyze storage directories.
|
* Analyze storage directories.
|
||||||
* Recover from previous transitions if required.
|
* Recover from previous transitions if required.
|
||||||
|
@ -36,6 +36,7 @@
|
|||||||
import org.apache.hadoop.hdfs.server.namenode.NNStorageRetentionManager.StoragePurger;
|
import org.apache.hadoop.hdfs.server.namenode.NNStorageRetentionManager.StoragePurger;
|
||||||
import org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.EditLogValidation;
|
import org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.EditLogValidation;
|
||||||
import org.apache.hadoop.hdfs.server.namenode.NNStorage.NameNodeFile;
|
import org.apache.hadoop.hdfs.server.namenode.NNStorage.NameNodeFile;
|
||||||
|
import org.apache.hadoop.hdfs.server.protocol.NamespaceInfo;
|
||||||
import org.apache.hadoop.hdfs.server.protocol.RemoteEditLog;
|
import org.apache.hadoop.hdfs.server.protocol.RemoteEditLog;
|
||||||
|
|
||||||
import com.google.common.annotations.VisibleForTesting;
|
import com.google.common.annotations.VisibleForTesting;
|
||||||
@ -77,6 +78,22 @@ public FileJournalManager(StorageDirectory sd,
|
|||||||
@Override
|
@Override
|
||||||
public void close() throws IOException {}
|
public void close() throws IOException {}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void format(NamespaceInfo ns) {
|
||||||
|
// Formatting file journals is done by the StorageDirectory
|
||||||
|
// format code, since they may share their directory with
|
||||||
|
// checkpoints, etc.
|
||||||
|
throw new UnsupportedOperationException();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean hasSomeData() {
|
||||||
|
// Formatting file journals is done by the StorageDirectory
|
||||||
|
// format code, since they may share their directory with
|
||||||
|
// checkpoints, etc.
|
||||||
|
throw new UnsupportedOperationException();
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
synchronized public EditLogOutputStream startLogSegment(long txid)
|
synchronized public EditLogOutputStream startLogSegment(long txid)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
|
@ -23,6 +23,8 @@
|
|||||||
|
|
||||||
import org.apache.hadoop.classification.InterfaceAudience;
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.classification.InterfaceStability;
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
import org.apache.hadoop.hdfs.server.common.Storage.FormatConfirmable;
|
||||||
|
import org.apache.hadoop.hdfs.server.protocol.NamespaceInfo;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* A JournalManager is responsible for managing a single place of storing
|
* A JournalManager is responsible for managing a single place of storing
|
||||||
@ -33,7 +35,14 @@
|
|||||||
*/
|
*/
|
||||||
@InterfaceAudience.Private
|
@InterfaceAudience.Private
|
||||||
@InterfaceStability.Evolving
|
@InterfaceStability.Evolving
|
||||||
public interface JournalManager extends Closeable {
|
public interface JournalManager extends Closeable, FormatConfirmable {
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Format the underlying storage, removing any previously
|
||||||
|
* stored data.
|
||||||
|
*/
|
||||||
|
void format(NamespaceInfo ns);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Begin writing to a new segment of the log stream, which starts at
|
* Begin writing to a new segment of the log stream, which starts at
|
||||||
* the given transaction ID.
|
* the given transaction ID.
|
||||||
|
@ -33,6 +33,7 @@
|
|||||||
|
|
||||||
import static org.apache.hadoop.util.ExitUtil.terminate;
|
import static org.apache.hadoop.util.ExitUtil.terminate;
|
||||||
|
|
||||||
|
import org.apache.hadoop.hdfs.server.protocol.NamespaceInfo;
|
||||||
import org.apache.hadoop.hdfs.server.protocol.RemoteEditLog;
|
import org.apache.hadoop.hdfs.server.protocol.RemoteEditLog;
|
||||||
import org.apache.hadoop.hdfs.server.protocol.RemoteEditLogManifest;
|
import org.apache.hadoop.hdfs.server.protocol.RemoteEditLogManifest;
|
||||||
|
|
||||||
@ -172,6 +173,20 @@ public boolean isRequired() {
|
|||||||
this.minimumRedundantJournals = minimumRedundantResources;
|
this.minimumRedundantJournals = minimumRedundantResources;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void format(NamespaceInfo nsInfo) {
|
||||||
|
// The iteration is done by FSEditLog itself
|
||||||
|
throw new UnsupportedOperationException();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean hasSomeData() throws IOException {
|
||||||
|
// This is called individually on the underlying journals,
|
||||||
|
// not on the JournalSet.
|
||||||
|
throw new UnsupportedOperationException();
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public EditLogOutputStream startLogSegment(final long txId) throws IOException {
|
public EditLogOutputStream startLogSegment(final long txId) throws IOException {
|
||||||
mapJournalsAndReportErrors(new JournalClosure() {
|
mapJournalsAndReportErrors(new JournalClosure() {
|
||||||
|
@ -28,7 +28,6 @@
|
|||||||
import java.util.HashMap;
|
import java.util.HashMap;
|
||||||
import java.util.Iterator;
|
import java.util.Iterator;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.Map;
|
|
||||||
import java.util.Properties;
|
import java.util.Properties;
|
||||||
import java.util.UUID;
|
import java.util.UUID;
|
||||||
import java.util.concurrent.CopyOnWriteArrayList;
|
import java.util.concurrent.CopyOnWriteArrayList;
|
||||||
@ -58,7 +57,6 @@
|
|||||||
import com.google.common.base.Preconditions;
|
import com.google.common.base.Preconditions;
|
||||||
import com.google.common.annotations.VisibleForTesting;
|
import com.google.common.annotations.VisibleForTesting;
|
||||||
import com.google.common.collect.Lists;
|
import com.google.common.collect.Lists;
|
||||||
import com.google.common.collect.Maps;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* NNStorage is responsible for management of the StorageDirectories used by
|
* NNStorage is responsible for management of the StorageDirectories used by
|
||||||
|
@ -67,6 +67,7 @@
|
|||||||
import org.apache.hadoop.hdfs.server.protocol.NamenodeProtocol;
|
import org.apache.hadoop.hdfs.server.protocol.NamenodeProtocol;
|
||||||
import org.apache.hadoop.hdfs.server.protocol.NamenodeProtocols;
|
import org.apache.hadoop.hdfs.server.protocol.NamenodeProtocols;
|
||||||
import org.apache.hadoop.hdfs.server.protocol.NamenodeRegistration;
|
import org.apache.hadoop.hdfs.server.protocol.NamenodeRegistration;
|
||||||
|
import org.apache.hadoop.hdfs.server.protocol.NamespaceInfo;
|
||||||
import org.apache.hadoop.hdfs.util.AtomicFileOutputStream;
|
import org.apache.hadoop.hdfs.util.AtomicFileOutputStream;
|
||||||
import org.apache.hadoop.io.IOUtils;
|
import org.apache.hadoop.io.IOUtils;
|
||||||
import org.apache.hadoop.ipc.Server;
|
import org.apache.hadoop.ipc.Server;
|
||||||
@ -709,9 +710,6 @@ private static boolean format(Configuration conf, boolean force,
|
|||||||
dirsToPrompt.addAll(sharedDirs);
|
dirsToPrompt.addAll(sharedDirs);
|
||||||
List<URI> editDirsToFormat =
|
List<URI> editDirsToFormat =
|
||||||
FSNamesystem.getNamespaceEditsDirs(conf);
|
FSNamesystem.getNamespaceEditsDirs(conf);
|
||||||
if (!confirmFormat(dirsToPrompt, force, isInteractive)) {
|
|
||||||
return true; // aborted
|
|
||||||
}
|
|
||||||
|
|
||||||
// if clusterID is not provided - see if you can find the current one
|
// if clusterID is not provided - see if you can find the current one
|
||||||
String clusterId = StartupOption.FORMAT.getClusterId();
|
String clusterId = StartupOption.FORMAT.getClusterId();
|
||||||
@ -723,62 +721,16 @@ private static boolean format(Configuration conf, boolean force,
|
|||||||
|
|
||||||
FSImage fsImage = new FSImage(conf, nameDirsToFormat, editDirsToFormat);
|
FSImage fsImage = new FSImage(conf, nameDirsToFormat, editDirsToFormat);
|
||||||
FSNamesystem fsn = new FSNamesystem(conf, fsImage);
|
FSNamesystem fsn = new FSNamesystem(conf, fsImage);
|
||||||
|
fsImage.getEditLog().initJournalsForWrite();
|
||||||
|
|
||||||
|
if (!fsImage.confirmFormat(force, isInteractive)) {
|
||||||
|
return true; // aborted
|
||||||
|
}
|
||||||
|
|
||||||
fsImage.format(fsn, clusterId);
|
fsImage.format(fsn, clusterId);
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Check whether the given storage directories already exist.
|
|
||||||
* If running in interactive mode, will prompt the user for each
|
|
||||||
* directory to allow them to format anyway. Otherwise, returns
|
|
||||||
* false, unless 'force' is specified.
|
|
||||||
*
|
|
||||||
* @param dirsToFormat the dirs to check
|
|
||||||
* @param force format regardless of whether dirs exist
|
|
||||||
* @param interactive prompt the user when a dir exists
|
|
||||||
* @return true if formatting should proceed
|
|
||||||
* @throws IOException
|
|
||||||
*/
|
|
||||||
public static boolean confirmFormat(Collection<URI> dirsToFormat,
|
|
||||||
boolean force, boolean interactive)
|
|
||||||
throws IOException {
|
|
||||||
for(Iterator<URI> it = dirsToFormat.iterator(); it.hasNext();) {
|
|
||||||
URI dirUri = it.next();
|
|
||||||
if (!dirUri.getScheme().equals(NNStorage.LOCAL_URI_SCHEME)) {
|
|
||||||
System.err.println("Skipping format for directory \"" + dirUri
|
|
||||||
+ "\". Can only format local directories with scheme \""
|
|
||||||
+ NNStorage.LOCAL_URI_SCHEME + "\".");
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
// To validate only file based schemes are formatted
|
|
||||||
assert dirUri.getScheme().equals(NNStorage.LOCAL_URI_SCHEME) :
|
|
||||||
"formatting is not supported for " + dirUri;
|
|
||||||
|
|
||||||
File curDir = new File(dirUri.getPath());
|
|
||||||
// Its alright for a dir not to exist, or to exist (properly accessible)
|
|
||||||
// and be completely empty.
|
|
||||||
if (!curDir.exists() ||
|
|
||||||
(curDir.isDirectory() && FileUtil.listFiles(curDir).length == 0))
|
|
||||||
continue;
|
|
||||||
if (force) { // Don't confirm, always format.
|
|
||||||
System.err.println(
|
|
||||||
"Storage directory exists in " + curDir + ". Formatting anyway.");
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
if (!interactive) { // Don't ask - always don't format
|
|
||||||
System.err.println(
|
|
||||||
"Running in non-interactive mode, and image appears to exist in " +
|
|
||||||
curDir + ". Not formatting.");
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
if (!confirmPrompt("Re-format filesystem in " + curDir + " ?")) {
|
|
||||||
System.err.println("Format aborted in " + curDir);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
public static void checkAllowFormat(Configuration conf) throws IOException {
|
public static void checkAllowFormat(Configuration conf) throws IOException {
|
||||||
if (!conf.getBoolean(DFS_NAMENODE_SUPPORT_ALLOW_FORMAT_KEY,
|
if (!conf.getBoolean(DFS_NAMENODE_SUPPORT_ALLOW_FORMAT_KEY,
|
||||||
DFS_NAMENODE_SUPPORT_ALLOW_FORMAT_DEFAULT)) {
|
DFS_NAMENODE_SUPPORT_ALLOW_FORMAT_DEFAULT)) {
|
||||||
@ -822,16 +774,25 @@ private static boolean initializeSharedEdits(Configuration conf,
|
|||||||
FSNamesystem.getNamespaceEditsDirs(conf, false));
|
FSNamesystem.getNamespaceEditsDirs(conf, false));
|
||||||
|
|
||||||
existingStorage = fsns.getFSImage().getStorage();
|
existingStorage = fsns.getFSImage().getStorage();
|
||||||
|
NamespaceInfo nsInfo = existingStorage.getNamespaceInfo();
|
||||||
|
|
||||||
Collection<URI> sharedEditsDirs = FSNamesystem.getSharedEditsDirs(conf);
|
List<URI> sharedEditsDirs = FSNamesystem.getSharedEditsDirs(conf);
|
||||||
if (!confirmFormat(sharedEditsDirs, force, interactive)) {
|
|
||||||
return true; // aborted
|
FSImage sharedEditsImage = new FSImage(conf,
|
||||||
}
|
|
||||||
NNStorage newSharedStorage = new NNStorage(conf,
|
|
||||||
Lists.<URI>newArrayList(),
|
Lists.<URI>newArrayList(),
|
||||||
sharedEditsDirs);
|
sharedEditsDirs);
|
||||||
|
sharedEditsImage.getEditLog().initJournalsForWrite();
|
||||||
|
|
||||||
newSharedStorage.format(existingStorage.getNamespaceInfo());
|
if (!sharedEditsImage.confirmFormat(force, interactive)) {
|
||||||
|
return true; // abort
|
||||||
|
}
|
||||||
|
|
||||||
|
NNStorage newSharedStorage = sharedEditsImage.getStorage();
|
||||||
|
// Call Storage.format instead of FSImage.format here, since we don't
|
||||||
|
// actually want to save a checkpoint - just prime the dirs with
|
||||||
|
// the existing namespace info
|
||||||
|
newSharedStorage.format(nsInfo);
|
||||||
|
sharedEditsImage.getEditLog().formatNonFileJournals(nsInfo);
|
||||||
|
|
||||||
// Need to make sure the edit log segments are in good shape to initialize
|
// Need to make sure the edit log segments are in good shape to initialize
|
||||||
// the shared edits dir.
|
// the shared edits dir.
|
||||||
|
@ -37,6 +37,7 @@
|
|||||||
import org.apache.hadoop.hdfs.HAUtil;
|
import org.apache.hadoop.hdfs.HAUtil;
|
||||||
import org.apache.hadoop.hdfs.NameNodeProxies;
|
import org.apache.hadoop.hdfs.NameNodeProxies;
|
||||||
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
||||||
|
import org.apache.hadoop.hdfs.server.common.Storage;
|
||||||
import org.apache.hadoop.hdfs.server.namenode.EditLogInputStream;
|
import org.apache.hadoop.hdfs.server.namenode.EditLogInputStream;
|
||||||
import org.apache.hadoop.hdfs.server.namenode.FSImage;
|
import org.apache.hadoop.hdfs.server.namenode.FSImage;
|
||||||
import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
|
import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
|
||||||
@ -55,7 +56,6 @@
|
|||||||
|
|
||||||
import com.google.common.base.Joiner;
|
import com.google.common.base.Joiner;
|
||||||
import com.google.common.base.Preconditions;
|
import com.google.common.base.Preconditions;
|
||||||
import com.google.common.collect.Sets;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Tool which allows the standby node's storage directories to be bootstrapped
|
* Tool which allows the standby node's storage directories to be bootstrapped
|
||||||
@ -171,19 +171,18 @@ private int doRun() throws IOException {
|
|||||||
" Layout version: " + nsInfo.getLayoutVersion() + "\n" +
|
" Layout version: " + nsInfo.getLayoutVersion() + "\n" +
|
||||||
"=====================================================");
|
"=====================================================");
|
||||||
|
|
||||||
|
long imageTxId = proxy.getMostRecentCheckpointTxId();
|
||||||
|
long curTxId = proxy.getTransactionID();
|
||||||
|
|
||||||
|
NNStorage storage = new NNStorage(conf, dirsToFormat, editUrisToFormat);
|
||||||
|
|
||||||
// Check with the user before blowing away data.
|
// Check with the user before blowing away data.
|
||||||
if (!NameNode.confirmFormat(
|
if (!Storage.confirmFormat(storage.dirIterable(null),
|
||||||
Sets.union(Sets.newHashSet(dirsToFormat),
|
|
||||||
Sets.newHashSet(editUrisToFormat)),
|
|
||||||
force, interactive)) {
|
force, interactive)) {
|
||||||
return ERR_CODE_ALREADY_FORMATTED;
|
return ERR_CODE_ALREADY_FORMATTED;
|
||||||
}
|
}
|
||||||
|
|
||||||
long imageTxId = proxy.getMostRecentCheckpointTxId();
|
|
||||||
long curTxId = proxy.getTransactionID();
|
|
||||||
|
|
||||||
// Format the storage (writes VERSION file)
|
// Format the storage (writes VERSION file)
|
||||||
NNStorage storage = new NNStorage(conf, dirsToFormat, editUrisToFormat);
|
|
||||||
storage.format(nsInfo);
|
storage.format(nsInfo);
|
||||||
|
|
||||||
// Load the newly formatted image, using all of the directories (including shared
|
// Load the newly formatted image, using all of the directories (including shared
|
||||||
|
@ -17,9 +17,7 @@
|
|||||||
*/
|
*/
|
||||||
package org.apache.hadoop.hdfs.server.namenode;
|
package org.apache.hadoop.hdfs.server.namenode;
|
||||||
|
|
||||||
import static org.junit.Assert.assertEquals;
|
import static org.junit.Assert.*;
|
||||||
import static org.junit.Assert.assertNotNull;
|
|
||||||
import static org.junit.Assert.fail;
|
|
||||||
import static org.mockito.Mockito.mock;
|
import static org.mockito.Mockito.mock;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
@ -29,6 +27,7 @@
|
|||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||||
|
import org.apache.hadoop.hdfs.server.common.Storage.FormatConfirmable;
|
||||||
import org.apache.hadoop.hdfs.server.protocol.NamespaceInfo;
|
import org.apache.hadoop.hdfs.server.protocol.NamespaceInfo;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
|
|
||||||
@ -125,6 +124,8 @@ public void testDummyJournalManager() throws Exception {
|
|||||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(0).build();
|
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(0).build();
|
||||||
cluster.waitActive();
|
cluster.waitActive();
|
||||||
|
|
||||||
|
assertTrue(DummyJournalManager.shouldPromptCalled);
|
||||||
|
assertTrue(DummyJournalManager.formatCalled);
|
||||||
assertNotNull(DummyJournalManager.conf);
|
assertNotNull(DummyJournalManager.conf);
|
||||||
assertEquals(new URI(DUMMY_URI), DummyJournalManager.uri);
|
assertEquals(new URI(DUMMY_URI), DummyJournalManager.uri);
|
||||||
assertNotNull(DummyJournalManager.nsInfo);
|
assertNotNull(DummyJournalManager.nsInfo);
|
||||||
@ -141,6 +142,8 @@ public static class DummyJournalManager implements JournalManager {
|
|||||||
static Configuration conf = null;
|
static Configuration conf = null;
|
||||||
static URI uri = null;
|
static URI uri = null;
|
||||||
static NamespaceInfo nsInfo = null;
|
static NamespaceInfo nsInfo = null;
|
||||||
|
static boolean formatCalled = false;
|
||||||
|
static boolean shouldPromptCalled = false;
|
||||||
|
|
||||||
public DummyJournalManager(Configuration conf, URI u,
|
public DummyJournalManager(Configuration conf, URI u,
|
||||||
NamespaceInfo nsInfo) {
|
NamespaceInfo nsInfo) {
|
||||||
@ -150,6 +153,11 @@ public DummyJournalManager(Configuration conf, URI u,
|
|||||||
DummyJournalManager.nsInfo = nsInfo;
|
DummyJournalManager.nsInfo = nsInfo;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void format(NamespaceInfo nsInfo) {
|
||||||
|
formatCalled = true;
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public EditLogOutputStream startLogSegment(long txId) throws IOException {
|
public EditLogOutputStream startLogSegment(long txId) throws IOException {
|
||||||
return mock(EditLogOutputStream.class);
|
return mock(EditLogOutputStream.class);
|
||||||
@ -178,6 +186,12 @@ public void recoverUnfinalizedSegments() throws IOException {}
|
|||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void close() throws IOException {}
|
public void close() throws IOException {}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean hasSomeData() throws IOException {
|
||||||
|
shouldPromptCalled = true;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
public static class BadConstructorJournalManager extends DummyJournalManager {
|
public static class BadConstructorJournalManager extends DummyJournalManager {
|
||||||
|
Loading…
Reference in New Issue
Block a user