MAPREDUCE-6997. Moving logging APIs over to slf4j in hadoop-mapreduce-client-hs. Contributed by Gergely Novák.
This commit is contained in:
parent
410d0319cf
commit
ffee10b68e
@ -30,8 +30,6 @@
|
||||
import com.google.common.cache.LoadingCache;
|
||||
import com.google.common.cache.Weigher;
|
||||
import com.google.common.util.concurrent.UncheckedExecutionException;
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.mapreduce.v2.api.records.JobId;
|
||||
import org.apache.hadoop.mapreduce.v2.api.records.JobReport;
|
||||
@ -45,13 +43,16 @@
|
||||
import org.apache.hadoop.yarn.exceptions.YarnRuntimeException;
|
||||
|
||||
import com.google.common.annotations.VisibleForTesting;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* Manages an in memory cache of parsed Job History files.
|
||||
*/
|
||||
public class CachedHistoryStorage extends AbstractService implements
|
||||
HistoryStorage {
|
||||
private static final Log LOG = LogFactory.getLog(CachedHistoryStorage.class);
|
||||
private static final Logger LOG =
|
||||
LoggerFactory.getLogger(CachedHistoryStorage.class);
|
||||
|
||||
private LoadingCache<JobId, Job> loadedJobCache = null;
|
||||
private int loadedJobCacheSize;
|
||||
|
@ -31,8 +31,6 @@
|
||||
import java.util.concurrent.locks.Lock;
|
||||
import java.util.concurrent.locks.ReentrantLock;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.mapred.JobACLsManager;
|
||||
@ -63,6 +61,8 @@
|
||||
import org.apache.hadoop.yarn.api.records.Priority;
|
||||
import org.apache.hadoop.yarn.exceptions.YarnRuntimeException;
|
||||
import org.apache.hadoop.yarn.util.Records;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
|
||||
/**
|
||||
@ -75,7 +75,7 @@ public class CompletedJob implements org.apache.hadoop.mapreduce.v2.app.job.Job
|
||||
// so we count it as 0
|
||||
private static final int UNDEFINED_VALUE = -1;
|
||||
|
||||
static final Log LOG = LogFactory.getLog(CompletedJob.class);
|
||||
private static final Logger LOG = LoggerFactory.getLogger(CompletedJob.class);
|
||||
private final Configuration conf;
|
||||
private final JobId jobId; //Can be picked from JobInfo with a conversion.
|
||||
private final String user; //Can be picked up from JobInfo
|
||||
|
@ -20,14 +20,15 @@
|
||||
|
||||
import java.net.InetAddress;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.classification.InterfaceAudience.Private;
|
||||
import org.apache.hadoop.ipc.Server;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
@Private
|
||||
public class HSAuditLogger {
|
||||
private static final Log LOG = LogFactory.getLog(HSAuditLogger.class);
|
||||
private static final Logger LOG =
|
||||
LoggerFactory.getLogger(HSAuditLogger.class);
|
||||
|
||||
enum Keys {
|
||||
USER, OPERATION, TARGET, RESULT, IP, PERMISSIONS, DESCRIPTION
|
||||
|
@ -21,8 +21,6 @@
|
||||
import java.io.IOException;
|
||||
import java.net.InetSocketAddress;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.classification.InterfaceAudience.Private;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.ipc.ProtobufRpcEngine;
|
||||
@ -38,11 +36,13 @@
|
||||
import org.apache.hadoop.tools.GetUserMappingsProtocol;
|
||||
import org.apache.hadoop.tools.protocolPB.GetUserMappingsProtocolClientSideTranslatorPB;
|
||||
import org.apache.hadoop.tools.protocolPB.GetUserMappingsProtocolPB;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
@Private
|
||||
public class HSProxies {
|
||||
|
||||
private static final Log LOG = LogFactory.getLog(HSProxies.class);
|
||||
private static final Logger LOG = LoggerFactory.getLogger(HSProxies.class);
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
public static <T> T createProxy(Configuration conf, InetSocketAddress hsaddr,
|
||||
|
@ -26,8 +26,6 @@
|
||||
import java.util.Collection;
|
||||
import java.util.EnumSet;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.classification.InterfaceAudience.Private;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.CommonConfigurationKeysPublic;
|
||||
@ -90,6 +88,8 @@
|
||||
import org.apache.hadoop.yarn.webapp.WebApps;
|
||||
|
||||
import com.google.common.annotations.VisibleForTesting;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* This module is responsible for talking to the
|
||||
@ -98,7 +98,8 @@
|
||||
*/
|
||||
public class HistoryClientService extends AbstractService {
|
||||
|
||||
private static final Log LOG = LogFactory.getLog(HistoryClientService.class);
|
||||
private static final Logger LOG =
|
||||
LoggerFactory.getLogger(HistoryClientService.class);
|
||||
|
||||
private HSClientProtocol protocolHandler;
|
||||
private Server server;
|
||||
|
@ -40,8 +40,6 @@
|
||||
import java.util.concurrent.TimeUnit;
|
||||
import java.util.concurrent.atomic.AtomicInteger;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
import org.apache.hadoop.classification.InterfaceStability;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
@ -76,6 +74,8 @@
|
||||
import com.google.common.util.concurrent.ThreadFactoryBuilder;
|
||||
import org.apache.hadoop.yarn.util.Clock;
|
||||
import org.apache.hadoop.yarn.util.SystemClock;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* This class provides a way to interact with history files in a thread safe
|
||||
@ -84,8 +84,10 @@
|
||||
@InterfaceAudience.Public
|
||||
@InterfaceStability.Unstable
|
||||
public class HistoryFileManager extends AbstractService {
|
||||
private static final Log LOG = LogFactory.getLog(HistoryFileManager.class);
|
||||
private static final Log SUMMARY_LOG = LogFactory.getLog(JobSummary.class);
|
||||
private static final Logger LOG =
|
||||
LoggerFactory.getLogger(HistoryFileManager.class);
|
||||
private static final Logger SUMMARY_LOG =
|
||||
LoggerFactory.getLogger(JobSummary.class);
|
||||
|
||||
private enum HistoryInfoState {
|
||||
IN_INTERMEDIATE, IN_DONE, DELETED, MOVE_FAILED
|
||||
|
@ -27,8 +27,6 @@
|
||||
import java.util.HashSet;
|
||||
import java.util.Set;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.classification.InterfaceAudience.Private;
|
||||
import org.apache.hadoop.classification.InterfaceStability.Unstable;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
@ -44,6 +42,8 @@
|
||||
import org.apache.hadoop.mapreduce.v2.jobhistory.JHAdminConfig;
|
||||
import org.apache.hadoop.security.token.delegation.DelegationKey;
|
||||
import org.apache.hadoop.util.Shell;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
@Private
|
||||
@Unstable
|
||||
@ -54,8 +54,8 @@
|
||||
public class HistoryServerFileSystemStateStoreService
|
||||
extends HistoryServerStateStoreService {
|
||||
|
||||
public static final Log LOG =
|
||||
LogFactory.getLog(HistoryServerFileSystemStateStoreService.class);
|
||||
public static final Logger LOG =
|
||||
LoggerFactory.getLogger(HistoryServerFileSystemStateStoreService.class);
|
||||
|
||||
private static final String ROOT_STATE_DIR_NAME = "HistoryServerState";
|
||||
private static final String TOKEN_STATE_DIR_NAME = "tokens";
|
||||
@ -192,7 +192,7 @@ public void storeTokenMasterKey(DelegationKey key) throws IOException {
|
||||
dataStream.close();
|
||||
dataStream = null;
|
||||
} finally {
|
||||
IOUtils.cleanup(LOG, dataStream);
|
||||
IOUtils.cleanupWithLogger(LOG, dataStream);
|
||||
}
|
||||
|
||||
createNewFile(keyPath, memStream.toByteArray());
|
||||
@ -265,7 +265,7 @@ private void writeFile(Path file, byte[] data) throws IOException {
|
||||
out.close();
|
||||
out = null;
|
||||
} finally {
|
||||
IOUtils.cleanup(LOG, out);
|
||||
IOUtils.cleanupWithLogger(LOG, out);
|
||||
}
|
||||
} catch (IOException e) {
|
||||
fs.delete(file, false);
|
||||
@ -279,7 +279,7 @@ private byte[] readFile(Path file, long numBytes) throws IOException {
|
||||
try {
|
||||
in.readFully(data);
|
||||
} finally {
|
||||
IOUtils.cleanup(LOG, in);
|
||||
IOUtils.cleanupWithLogger(LOG, in);
|
||||
}
|
||||
return data;
|
||||
}
|
||||
@ -306,7 +306,7 @@ private byte[] buildTokenData(MRDelegationTokenIdentifier tokenId,
|
||||
dataStream.close();
|
||||
dataStream = null;
|
||||
} finally {
|
||||
IOUtils.cleanup(LOG, dataStream);
|
||||
IOUtils.cleanupWithLogger(LOG, dataStream);
|
||||
}
|
||||
return memStream.toByteArray();
|
||||
}
|
||||
@ -320,7 +320,7 @@ private void loadTokenMasterKey(HistoryServerState state, Path keyFile,
|
||||
try {
|
||||
key.readFields(in);
|
||||
} finally {
|
||||
IOUtils.cleanup(LOG, in);
|
||||
IOUtils.cleanupWithLogger(LOG, in);
|
||||
}
|
||||
state.tokenMasterKeyState.add(key);
|
||||
}
|
||||
@ -349,7 +349,7 @@ private MRDelegationTokenIdentifier loadToken(HistoryServerState state,
|
||||
tokenId.readFields(in);
|
||||
renewDate = in.readLong();
|
||||
} finally {
|
||||
IOUtils.cleanup(LOG, in);
|
||||
IOUtils.cleanupWithLogger(LOG, in);
|
||||
}
|
||||
state.tokenState.put(tokenId, renewDate);
|
||||
return tokenId;
|
||||
|
@ -29,8 +29,6 @@
|
||||
import java.io.IOException;
|
||||
import java.util.Map.Entry;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
@ -47,8 +45,9 @@
|
||||
import org.fusesource.leveldbjni.internal.NativeDB;
|
||||
import org.iq80.leveldb.DB;
|
||||
import org.iq80.leveldb.DBException;
|
||||
import org.iq80.leveldb.Logger;
|
||||
import org.iq80.leveldb.Options;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
public class HistoryServerLeveldbStateStoreService extends
|
||||
HistoryServerStateStoreService {
|
||||
@ -63,8 +62,8 @@ public class HistoryServerLeveldbStateStoreService extends
|
||||
|
||||
private DB db;
|
||||
|
||||
public static final Log LOG =
|
||||
LogFactory.getLog(HistoryServerLeveldbStateStoreService.class);
|
||||
public static final Logger LOG =
|
||||
LoggerFactory.getLogger(HistoryServerLeveldbStateStoreService.class);
|
||||
|
||||
@Override
|
||||
protected void initStorage(Configuration conf) throws IOException {
|
||||
@ -158,7 +157,7 @@ private void loadTokenMasterKey(HistoryServerState state, byte[] data)
|
||||
try {
|
||||
key.readFields(in);
|
||||
} finally {
|
||||
IOUtils.cleanup(LOG, in);
|
||||
IOUtils.cleanupWithLogger(LOG, in);
|
||||
}
|
||||
state.tokenMasterKeyState.add(key);
|
||||
}
|
||||
@ -204,7 +203,7 @@ private void loadToken(HistoryServerState state, byte[] data)
|
||||
tokenId.readFields(in);
|
||||
renewDate = in.readLong();
|
||||
} finally {
|
||||
IOUtils.cleanup(LOG, in);
|
||||
IOUtils.cleanupWithLogger(LOG, in);
|
||||
}
|
||||
state.tokenState.put(tokenId, renewDate);
|
||||
}
|
||||
@ -224,7 +223,7 @@ public void storeToken(MRDelegationTokenIdentifier tokenId, Long renewDate)
|
||||
dataStream.close();
|
||||
dataStream = null;
|
||||
} finally {
|
||||
IOUtils.cleanup(LOG, dataStream);
|
||||
IOUtils.cleanupWithLogger(LOG, dataStream);
|
||||
}
|
||||
|
||||
String dbKey = getTokenDatabaseKey(tokenId);
|
||||
@ -270,7 +269,7 @@ public void storeTokenMasterKey(DelegationKey masterKey)
|
||||
dataStream.close();
|
||||
dataStream = null;
|
||||
} finally {
|
||||
IOUtils.cleanup(LOG, dataStream);
|
||||
IOUtils.cleanupWithLogger(LOG, dataStream);
|
||||
}
|
||||
|
||||
String dbKey = getTokenMasterKeyDatabaseKey(masterKey);
|
||||
@ -368,8 +367,9 @@ private void checkVersion() throws IOException {
|
||||
}
|
||||
}
|
||||
|
||||
private static class LeveldbLogger implements Logger {
|
||||
private static final Log LOG = LogFactory.getLog(LeveldbLogger.class);
|
||||
private static class LeveldbLogger implements org.iq80.leveldb.Logger {
|
||||
private static final Logger LOG =
|
||||
LoggerFactory.getLogger(LeveldbLogger.class);
|
||||
|
||||
@Override
|
||||
public void log(String message) {
|
||||
|
@ -21,14 +21,14 @@
|
||||
import java.io.IOException;
|
||||
import java.util.Map.Entry;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
import org.apache.hadoop.classification.InterfaceStability;
|
||||
import org.apache.hadoop.mapreduce.v2.api.MRDelegationTokenIdentifier;
|
||||
import org.apache.hadoop.mapreduce.v2.hs.HistoryServerStateStoreService.HistoryServerState;
|
||||
import org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager;
|
||||
import org.apache.hadoop.security.token.delegation.DelegationKey;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* A MapReduce specific delegation token secret manager.
|
||||
@ -40,7 +40,7 @@
|
||||
public class JHSDelegationTokenSecretManager
|
||||
extends AbstractDelegationTokenSecretManager<MRDelegationTokenIdentifier> {
|
||||
|
||||
private static final Log LOG = LogFactory.getLog(
|
||||
private static final Logger LOG = LoggerFactory.getLogger(
|
||||
JHSDelegationTokenSecretManager.class);
|
||||
|
||||
private HistoryServerStateStoreService store;
|
||||
|
@ -27,8 +27,6 @@
|
||||
import java.util.concurrent.TimeUnit;
|
||||
import java.util.regex.Pattern;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.mapreduce.JobID;
|
||||
import org.apache.hadoop.mapreduce.MRJobConfig;
|
||||
@ -56,12 +54,14 @@
|
||||
|
||||
import com.google.common.annotations.VisibleForTesting;
|
||||
import com.google.common.util.concurrent.ThreadFactoryBuilder;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* Loads and manages the Job history cache.
|
||||
*/
|
||||
public class JobHistory extends AbstractService implements HistoryContext {
|
||||
private static final Log LOG = LogFactory.getLog(JobHistory.class);
|
||||
private static final Logger LOG = LoggerFactory.getLogger(JobHistory.class);
|
||||
|
||||
public static final Pattern CONF_FILENAME_REGEX = Pattern.compile("("
|
||||
+ JobID.JOBID_REGEX + ")_conf.xml(?:\\.[0-9]+\\.old)?");
|
||||
|
@ -21,8 +21,6 @@
|
||||
import java.io.IOException;
|
||||
import java.net.InetSocketAddress;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.classification.InterfaceAudience.Private;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.mapred.JobConf;
|
||||
@ -47,6 +45,8 @@
|
||||
import org.apache.hadoop.yarn.logaggregation.AggregatedLogDeletionService;
|
||||
|
||||
import com.google.common.annotations.VisibleForTesting;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/******************************************************************
|
||||
* {@link JobHistoryServer} is responsible for servicing all job history
|
||||
@ -62,7 +62,8 @@ public class JobHistoryServer extends CompositeService {
|
||||
|
||||
public static final long historyServerTimeStamp = System.currentTimeMillis();
|
||||
|
||||
private static final Log LOG = LogFactory.getLog(JobHistoryServer.class);
|
||||
private static final Logger LOG =
|
||||
LoggerFactory.getLogger(JobHistoryServer.class);
|
||||
protected HistoryContext historyContext;
|
||||
private HistoryClientService clientService;
|
||||
private JobHistory jobHistoryService;
|
||||
@ -225,7 +226,7 @@ static JobHistoryServer launchJobHistoryServer(String[] args) {
|
||||
jobHistoryServer.init(conf);
|
||||
jobHistoryServer.start();
|
||||
} catch (Throwable t) {
|
||||
LOG.fatal("Error starting JobHistoryServer", t);
|
||||
LOG.error("Error starting JobHistoryServer", t);
|
||||
ExitUtil.terminate(-1, "Error starting JobHistoryServer");
|
||||
}
|
||||
return jobHistoryServer;
|
||||
|
@ -21,8 +21,6 @@
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.mapred.TaskCompletionEvent;
|
||||
@ -41,10 +39,12 @@
|
||||
import org.apache.hadoop.security.authorize.AccessControlList;
|
||||
import org.apache.hadoop.yarn.api.records.Priority;
|
||||
import org.apache.hadoop.yarn.factory.providers.RecordFactoryProvider;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
|
||||
public class PartialJob implements org.apache.hadoop.mapreduce.v2.app.job.Job {
|
||||
private static final Log LOG = LogFactory.getLog(PartialJob.class);
|
||||
private static final Logger LOG = LoggerFactory.getLogger(PartialJob.class);
|
||||
|
||||
private JobIndexInfo jobIndexInfo = null;
|
||||
private JobId jobId = null;
|
||||
|
@ -22,8 +22,6 @@
|
||||
import java.net.InetSocketAddress;
|
||||
import java.security.PrivilegedExceptionAction;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.classification.InterfaceAudience.Private;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.CommonConfigurationKeysPublic;
|
||||
@ -54,11 +52,14 @@
|
||||
|
||||
import com.google.common.annotations.VisibleForTesting;
|
||||
import com.google.protobuf.BlockingService;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
@Private
|
||||
public class HSAdminServer extends AbstractService implements HSAdminProtocol {
|
||||
|
||||
private static final Log LOG = LogFactory.getLog(HSAdminServer.class);
|
||||
private static final Logger LOG =
|
||||
LoggerFactory.getLogger(HSAdminServer.class);
|
||||
private AccessControlList adminAcl;
|
||||
private AggregatedLogDeletionService aggLogDelService = null;
|
||||
|
||||
|
@ -22,8 +22,6 @@
|
||||
|
||||
import org.junit.Assert;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.mapreduce.TypeConverter;
|
||||
import org.apache.hadoop.mapreduce.jobhistory.JobHistoryEvent;
|
||||
@ -44,9 +42,12 @@
|
||||
import org.apache.hadoop.yarn.api.records.ContainerId;
|
||||
import org.apache.hadoop.yarn.event.EventHandler;
|
||||
import org.junit.Test;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
public class TestJobHistoryEvents {
|
||||
private static final Log LOG = LogFactory.getLog(TestJobHistoryEvents.class);
|
||||
private static final Logger LOG =
|
||||
LoggerFactory.getLogger(TestJobHistoryEvents.class);
|
||||
|
||||
@Test
|
||||
public void testHistoryEvents() throws Exception {
|
||||
|
@ -38,8 +38,6 @@
|
||||
|
||||
import org.junit.Assert;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.FSDataInputStream;
|
||||
import org.apache.hadoop.fs.FileContext;
|
||||
@ -96,9 +94,12 @@
|
||||
import org.mockito.Mockito;
|
||||
import org.mockito.invocation.InvocationOnMock;
|
||||
import org.mockito.stubbing.Answer;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
public class TestJobHistoryParsing {
|
||||
private static final Log LOG = LogFactory.getLog(TestJobHistoryParsing.class);
|
||||
private static final Logger LOG =
|
||||
LoggerFactory.getLogger(TestJobHistoryParsing.class);
|
||||
|
||||
private static final String RACK_NAME = "/MyRackName";
|
||||
|
||||
|
@ -35,8 +35,6 @@
|
||||
import java.util.HashMap;
|
||||
import java.util.Map;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.mapreduce.v2.api.records.JobId;
|
||||
import org.apache.hadoop.mapreduce.v2.app.AppContext;
|
||||
@ -54,9 +52,11 @@
|
||||
|
||||
import com.google.inject.AbstractModule;
|
||||
import com.google.inject.Injector;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
public class TestHSWebApp {
|
||||
private static final Log LOG = LogFactory.getLog(TestHSWebApp.class);
|
||||
private static final Logger LOG = LoggerFactory.getLogger(TestHSWebApp.class);
|
||||
|
||||
@Test public void testAppControllerIndex() {
|
||||
MockAppContext ctx = new MockAppContext(0, 1, 1, 1);
|
||||
|
Loading…
Reference in New Issue
Block a user