Revert "HDFS-11201. Spelling errors in the logging, help, assertions and exception messages. Contributed by Grant Sohn."

This reverts commit b9522e86a5.
This commit is contained in:
Wei-Chiu Chuang 2016-12-05 10:54:43 -08:00
parent b9522e86a5
commit 1b5cceaffb
13 changed files with 15 additions and 15 deletions

View File

@ -1052,7 +1052,7 @@ void actualGetFromOneDataNode(final DNAddrPair datanode, LocatedBlock block,
reader.getNetworkDistance(), nread); reader.getNetworkDistance(), nread);
if (nread != len) { if (nread != len) {
throw new IOException("truncated return from reader.read(): " + throw new IOException("truncated return from reader.read(): " +
"expected " + len + ", got " + nread); "excpected " + len + ", got " + nread);
} }
DFSClientFaultInjector.get().readFromDatanodeDelay(); DFSClientFaultInjector.get().readFromDatanodeDelay();
return; return;

View File

@ -57,11 +57,11 @@ public long retrieve(long record) {
public long combine(long value, long record) { public long combine(long value, long record) {
if (value < MIN) { if (value < MIN) {
throw new IllegalArgumentException( throw new IllegalArgumentException(
"Illegal value: " + NAME + " = " + value + " < MIN = " + MIN); "Illagal value: " + NAME + " = " + value + " < MIN = " + MIN);
} }
if (value > MAX) { if (value > MAX) {
throw new IllegalArgumentException( throw new IllegalArgumentException(
"Illegal value: " + NAME + " = " + value + " > MAX = " + MAX); "Illagal value: " + NAME + " = " + value + " > MAX = " + MAX);
} }
return (record & ~MASK) | (value << OFFSET); return (record & ~MASK) | (value << OFFSET);
} }

View File

@ -38,7 +38,7 @@ public static enum ERROR implements XException.ERROR {
S04("Service [{0}] does not implement declared interface [{1}]"), S04("Service [{0}] does not implement declared interface [{1}]"),
S05("[{0}] is not a file"), S05("[{0}] is not a file"),
S06("Could not load file [{0}], {1}"), S06("Could not load file [{0}], {1}"),
S07("Could not instantiate service class [{0}], {1}"), S07("Could not instanciate service class [{0}], {1}"),
S08("Could not load service classes, {0}"), S08("Could not load service classes, {0}"),
S09("Could not set service [{0}] programmatically -server shutting down-, {1}"), S09("Could not set service [{0}] programmatically -server shutting down-, {1}"),
S10("Service [{0}] requires service [{1}]"), S10("Service [{0}] requires service [{1}]"),

View File

@ -130,7 +130,7 @@ boolean put(FileHandle h, OpenFileCtx context) {
} }
toEvict = openFileMap.remove(pairs.getKey()); toEvict = openFileMap.remove(pairs.getKey());
Preconditions.checkState(toEvict == pairs.getValue(), Preconditions.checkState(toEvict == pairs.getValue(),
"The deleted entry is not the same as oldest found."); "The deleted entry is not the same as odlest found.");
} }
} }
openFileMap.put(h, context); openFileMap.put(h, context);

View File

@ -1364,7 +1364,7 @@ public static HttpConfig.Policy getHttpPolicy(Configuration conf) {
DFSConfigKeys.DFS_HTTP_POLICY_DEFAULT); DFSConfigKeys.DFS_HTTP_POLICY_DEFAULT);
HttpConfig.Policy policy = HttpConfig.Policy.fromString(policyStr); HttpConfig.Policy policy = HttpConfig.Policy.fromString(policyStr);
if (policy == null) { if (policy == null) {
throw new HadoopIllegalArgumentException("Unrecognized value '" throw new HadoopIllegalArgumentException("Unregonized value '"
+ policyStr + "' for " + DFSConfigKeys.DFS_HTTP_POLICY_KEY); + policyStr + "' for " + DFSConfigKeys.DFS_HTTP_POLICY_KEY);
} }

View File

@ -2772,7 +2772,7 @@ private void processQueuedMessages(Iterable<ReportedBlockInfo> rbis)
throws IOException { throws IOException {
for (ReportedBlockInfo rbi : rbis) { for (ReportedBlockInfo rbi : rbis) {
if (LOG.isDebugEnabled()) { if (LOG.isDebugEnabled()) {
LOG.debug("Processing previously queued message " + rbi); LOG.debug("Processing previouly queued message " + rbi);
} }
if (rbi.getReportedState() == null) { if (rbi.getReportedState() == null) {
// This is a DELETE_BLOCK request // This is a DELETE_BLOCK request

View File

@ -836,7 +836,7 @@ private boolean readReplicasFromCache(ReplicaMap volumeMap,
} catch (Exception e) { } catch (Exception e) {
// Any exception we need to revert back to read from disk // Any exception we need to revert back to read from disk
// Log the error and return false // Log the error and return false
LOG.info("Exception occurred while reading the replicas cache file: " LOG.info("Exception occured while reading the replicas cache file: "
+ replicaFile.getPath(), e ); + replicaFile.getPath(), e );
return false; return false;
} }

View File

@ -1292,7 +1292,7 @@ private LinkedList<ScanInfo> compileReport(File bpFinalizedDir,
try { try {
fileNames = IOUtils.listDirectory(dir, BlockDirFilter.INSTANCE); fileNames = IOUtils.listDirectory(dir, BlockDirFilter.INSTANCE);
} catch (IOException ioe) { } catch (IOException ioe) {
LOG.warn("Exception occurred while compiling report: ", ioe); LOG.warn("Exception occured while compiling report: ", ioe);
// Initiate a check on disk failure. // Initiate a check on disk failure.
dataset.datanode.checkDiskErrorAsync(); dataset.datanode.checkDiskErrorAsync();
// Ignore this directory and proceed. // Ignore this directory and proceed.

View File

@ -97,7 +97,7 @@ public void printHelp() {
String header = "Query Plan queries a given data node about the " + String header = "Query Plan queries a given data node about the " +
"current state of disk balancer execution.\n\n"; "current state of disk balancer execution.\n\n";
String footer = "\nQuery command retrieves the plan ID and the current " + String footer = "\nQuery command retrievs the plan ID and the current " +
"running state. "; "running state. ";
HelpFormatter helpFormatter = new HelpFormatter(); HelpFormatter helpFormatter = new HelpFormatter();

View File

@ -134,7 +134,7 @@ static TruncateResult truncate(final FSNamesystem fsn, final String srcArg,
if (!onBlockBoundary) { if (!onBlockBoundary) {
// Open file for write, but don't log into edits // Open file for write, but don't log into edits
long lastBlockDelta = file.computeFileSize() - newLength; long lastBlockDelta = file.computeFileSize() - newLength;
assert lastBlockDelta > 0 : "delta is 0 only if on block boundary"; assert lastBlockDelta > 0 : "delta is 0 only if on block bounday";
truncateBlock = prepareFileForTruncate(fsn, iip, clientName, truncateBlock = prepareFileForTruncate(fsn, iip, clientName,
clientMachine, lastBlockDelta, null); clientMachine, lastBlockDelta, null);
} }

View File

@ -1082,7 +1082,7 @@ private void updateBlocks(FSDirectory fsDir, BlockListUpdatingOp op,
boolean removed = FSDirWriteFileOp.unprotectedRemoveBlock( boolean removed = FSDirWriteFileOp.unprotectedRemoveBlock(
fsDir, path, iip, file, oldBlock); fsDir, path, iip, file, oldBlock);
if (!removed && !(op instanceof UpdateBlocksOp)) { if (!removed && !(op instanceof UpdateBlocksOp)) {
throw new IOException("Trying to delete non-existent block " + oldBlock); throw new IOException("Trying to delete non-existant block " + oldBlock);
} }
} else if (newBlocks.length > oldBlocks.length) { } else if (newBlocks.length > oldBlocks.length) {
final boolean isStriped = ecPolicy != null; final boolean isStriped = ecPolicy != null;

View File

@ -1025,7 +1025,7 @@ private void setBlockPoolID(File storage, String bpid)
if (!blockpoolID.equals("") && !blockpoolID.equals(bpid)) { if (!blockpoolID.equals("") && !blockpoolID.equals(bpid)) {
throw new InconsistentFSStateException(storage, throw new InconsistentFSStateException(storage,
"Unexpected blockpoolID " + bpid + " . Expected " + blockpoolID); "Unexepcted blockpoolID " + bpid + " . Expected " + blockpoolID);
} }
setBlockPoolID(bpid); setBlockPoolID(bpid);
} }

View File

@ -219,7 +219,7 @@ static DatanodeInfo chooseDatanode(final NameNode namenode,
final String remoteAddr) throws IOException { final String remoteAddr) throws IOException {
FSNamesystem fsn = namenode.getNamesystem(); FSNamesystem fsn = namenode.getNamesystem();
if (fsn == null) { if (fsn == null) {
throw new IOException("Namesystem has not been initialized yet."); throw new IOException("Namesystem has not been intialized yet.");
} }
final BlockManager bm = fsn.getBlockManager(); final BlockManager bm = fsn.getBlockManager();