HDFS-17439. Support -nonSuperUser for NNThroughputBenchmark: useful for testing auth frameworks such as Ranger (#6677)
This commit is contained in:
parent
2fbbfe3cc9
commit
90024d8cb1
@ -46,6 +46,7 @@ The following are all supported command options:
|
|||||||
|`-logLevel` | Specify the logging level when the benchmark runs. The default logging level is ERROR. |
|
|`-logLevel` | Specify the logging level when the benchmark runs. The default logging level is ERROR. |
|
||||||
|`-UGCacheRefreshCount` | After every specified number of operations, the benchmark purges the name-node's user group cache. By default the refresh is never called. |
|
|`-UGCacheRefreshCount` | After every specified number of operations, the benchmark purges the name-node's user group cache. By default the refresh is never called. |
|
||||||
|`-keepResults` | If specified, do not clean up the name-space after execution. By default the name-space will be removed after test. |
|
|`-keepResults` | If specified, do not clean up the name-space after execution. By default the name-space will be removed after test. |
|
||||||
|
|`-nonSuperUser` | If specified, non super user can use the tool and can be helpful for bringing authorization time into benchmarking calculations. |
|
||||||
|
|
||||||
##### Operations Supported
|
##### Operations Supported
|
||||||
|
|
||||||
|
@ -89,10 +89,10 @@
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Main class for a series of name-node benchmarks.
|
* Main class for a series of name-node benchmarks.
|
||||||
*
|
*
|
||||||
* Each benchmark measures throughput and average execution time
|
* Each benchmark measures throughput and average execution time
|
||||||
* of a specific name-node operation, e.g. file creation or block reports.
|
* of a specific name-node operation, e.g. file creation or block reports.
|
||||||
*
|
*
|
||||||
* The benchmark does not involve any other hadoop components
|
* The benchmark does not involve any other hadoop components
|
||||||
* except for the name-node. Each operation is executed
|
* except for the name-node. Each operation is executed
|
||||||
* by calling directly the respective name-node method.
|
* by calling directly the respective name-node method.
|
||||||
@ -107,7 +107,7 @@ public class NNThroughputBenchmark implements Tool {
|
|||||||
LoggerFactory.getLogger(NNThroughputBenchmark.class);
|
LoggerFactory.getLogger(NNThroughputBenchmark.class);
|
||||||
private static final int BLOCK_SIZE = 16;
|
private static final int BLOCK_SIZE = 16;
|
||||||
private static final String GENERAL_OPTIONS_USAGE =
|
private static final String GENERAL_OPTIONS_USAGE =
|
||||||
"[-keepResults] | [-logLevel L] | [-UGCacheRefreshCount G]";
|
"[-keepResults] | [-logLevel L] | [-UGCacheRefreshCount G] [-nonSuperUser]";
|
||||||
|
|
||||||
static Configuration config;
|
static Configuration config;
|
||||||
static NameNode nameNode;
|
static NameNode nameNode;
|
||||||
@ -157,7 +157,7 @@ static void setNameNodeLoggingLevel(Level logLevel) {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Base class for collecting operation statistics.
|
* Base class for collecting operation statistics.
|
||||||
*
|
*
|
||||||
* Overload this class in order to run statistics for a
|
* Overload this class in order to run statistics for a
|
||||||
* specific name-node operation.
|
* specific name-node operation.
|
||||||
*/
|
*/
|
||||||
@ -175,6 +175,7 @@ abstract class OperationStatsBase {
|
|||||||
protected long cumulativeTime = 0; // sum of times for each op
|
protected long cumulativeTime = 0; // sum of times for each op
|
||||||
protected long elapsedTime = 0; // time from start to finish
|
protected long elapsedTime = 0; // time from start to finish
|
||||||
protected boolean keepResults = false;// don't clean base directory on exit
|
protected boolean keepResults = false;// don't clean base directory on exit
|
||||||
|
protected boolean nonSuperUser = false; // enter/exit safe mode
|
||||||
protected Level logLevel; // logging level, ERROR by default
|
protected Level logLevel; // logging level, ERROR by default
|
||||||
protected int ugcRefreshCount = 0; // user group cache refresh count
|
protected int ugcRefreshCount = 0; // user group cache refresh count
|
||||||
|
|
||||||
@ -187,7 +188,7 @@ abstract class OperationStatsBase {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Parse command line arguments.
|
* Parse command line arguments.
|
||||||
*
|
*
|
||||||
* @param args arguments
|
* @param args arguments
|
||||||
* @throws IOException
|
* @throws IOException
|
||||||
*/
|
*/
|
||||||
@ -195,7 +196,7 @@ abstract class OperationStatsBase {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Generate inputs for each daemon thread.
|
* Generate inputs for each daemon thread.
|
||||||
*
|
*
|
||||||
* @param opsPerThread number of inputs for each thread.
|
* @param opsPerThread number of inputs for each thread.
|
||||||
* @throws IOException
|
* @throws IOException
|
||||||
*/
|
*/
|
||||||
@ -205,7 +206,7 @@ abstract class OperationStatsBase {
|
|||||||
* This corresponds to the arg1 argument of
|
* This corresponds to the arg1 argument of
|
||||||
* {@link #executeOp(int, int, String)}, which can have different meanings
|
* {@link #executeOp(int, int, String)}, which can have different meanings
|
||||||
* depending on the operation performed.
|
* depending on the operation performed.
|
||||||
*
|
*
|
||||||
* @param daemonId id of the daemon calling this method
|
* @param daemonId id of the daemon calling this method
|
||||||
* @return the argument
|
* @return the argument
|
||||||
*/
|
*/
|
||||||
@ -213,7 +214,7 @@ abstract class OperationStatsBase {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Execute name-node operation.
|
* Execute name-node operation.
|
||||||
*
|
*
|
||||||
* @param daemonId id of the daemon calling this method.
|
* @param daemonId id of the daemon calling this method.
|
||||||
* @param inputIdx serial index of the operation called by the deamon.
|
* @param inputIdx serial index of the operation called by the deamon.
|
||||||
* @param arg1 operation specific argument.
|
* @param arg1 operation specific argument.
|
||||||
@ -247,7 +248,7 @@ void benchmark() throws IOException {
|
|||||||
return;
|
return;
|
||||||
int tIdx = 0; // thread index < nrThreads
|
int tIdx = 0; // thread index < nrThreads
|
||||||
int opsPerThread[] = new int[numThreads];
|
int opsPerThread[] = new int[numThreads];
|
||||||
for(int opsScheduled = 0; opsScheduled < numOpsRequired;
|
for(int opsScheduled = 0; opsScheduled < numOpsRequired;
|
||||||
opsScheduled += opsPerThread[tIdx++]) {
|
opsScheduled += opsPerThread[tIdx++]) {
|
||||||
// execute in a separate thread
|
// execute in a separate thread
|
||||||
opsPerThread[tIdx] = (numOpsRequired-opsScheduled)/(numThreads-tIdx);
|
opsPerThread[tIdx] = (numOpsRequired-opsScheduled)/(numThreads-tIdx);
|
||||||
@ -285,14 +286,30 @@ private boolean isInProgress() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
void cleanUp() throws IOException {
|
void cleanUp() throws IOException {
|
||||||
clientProto.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_LEAVE,
|
if (!nonSuperUser) {
|
||||||
false);
|
try {
|
||||||
|
clientProto.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_LEAVE,
|
||||||
|
false);
|
||||||
|
} catch (Exception e){
|
||||||
|
LOG.error("Potentially insufficient permission: try running the tool" +
|
||||||
|
"with -nonSuperUser argument or login as super user");
|
||||||
|
throw e;
|
||||||
|
}
|
||||||
|
}
|
||||||
if(!keepResults)
|
if(!keepResults)
|
||||||
clientProto.delete(getBaseDir(), true);
|
clientProto.delete(getBaseDir(), true);
|
||||||
else {
|
else {
|
||||||
clientProto.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_ENTER,
|
if (!nonSuperUser) {
|
||||||
true);
|
try {
|
||||||
clientProto.saveNamespace(0, 0);
|
clientProto.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_ENTER,
|
||||||
|
true);
|
||||||
|
clientProto.saveNamespace(0, 0);
|
||||||
|
} catch (Exception e){
|
||||||
|
LOG.error("Potentially insufficient permission: try running the tool" +
|
||||||
|
" with -nonSuperUser argument or login as super user");
|
||||||
|
throw e;
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
public String getBaseDirName() {
|
public String getBaseDirName() {
|
||||||
@ -341,7 +358,7 @@ void incrementStats(int ops, long time) {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Parse first 2 arguments, corresponding to the "-op" option.
|
* Parse first 2 arguments, corresponding to the "-op" option.
|
||||||
*
|
*
|
||||||
* @param args argument list
|
* @param args argument list
|
||||||
* @return true if operation is all, which means that options not related
|
* @return true if operation is all, which means that options not related
|
||||||
* to this operation should be ignored, or false otherwise, meaning
|
* to this operation should be ignored, or false otherwise, meaning
|
||||||
@ -358,6 +375,12 @@ protected boolean verifyOpArgument(List<String> args) {
|
|||||||
args.remove(krIndex);
|
args.remove(krIndex);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
int nonSuperUserIndex = args.indexOf("-nonSuperUser");
|
||||||
|
nonSuperUser = (nonSuperUserIndex >= 0);
|
||||||
|
if(nonSuperUser) {
|
||||||
|
args.remove(nonSuperUserIndex);
|
||||||
|
}
|
||||||
|
|
||||||
int llIndex = args.indexOf("-logLevel");
|
int llIndex = args.indexOf("-logLevel");
|
||||||
if(llIndex >= 0) {
|
if(llIndex >= 0) {
|
||||||
if(args.size() <= llIndex + 1)
|
if(args.size() <= llIndex + 1)
|
||||||
@ -422,7 +445,7 @@ public void run() {
|
|||||||
try {
|
try {
|
||||||
benchmarkOne();
|
benchmarkOne();
|
||||||
} catch(IOException ex) {
|
} catch(IOException ex) {
|
||||||
LOG.error("StatsDaemon " + daemonId + " failed: \n"
|
LOG.error("StatsDaemon " + daemonId + " failed: \n"
|
||||||
+ StringUtils.stringifyException(ex));
|
+ StringUtils.stringifyException(ex));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -499,10 +522,18 @@ String getExecutionArgument(int daemonId) {
|
|||||||
* Remove entire benchmark directory.
|
* Remove entire benchmark directory.
|
||||||
*/
|
*/
|
||||||
@Override
|
@Override
|
||||||
long executeOp(int daemonId, int inputIdx, String ignore)
|
long executeOp(int daemonId, int inputIdx, String ignore)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
clientProto.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_LEAVE,
|
if (!nonSuperUser) {
|
||||||
false);
|
try{
|
||||||
|
clientProto.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_LEAVE,
|
||||||
|
false);
|
||||||
|
} catch (Exception e){
|
||||||
|
LOG.error("Potentially insufficient permission: try running the tool" +
|
||||||
|
" with -nonSuperUser argument or login as super user");
|
||||||
|
throw e;
|
||||||
|
}
|
||||||
|
}
|
||||||
long start = Time.now();
|
long start = Time.now();
|
||||||
clientProto.delete(getBaseDirName(), true);
|
clientProto.delete(getBaseDirName(), true);
|
||||||
long end = Time.now();
|
long end = Time.now();
|
||||||
@ -519,7 +550,7 @@ void printResults() {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* File creation statistics.
|
* File creation statistics.
|
||||||
*
|
*
|
||||||
* Each thread creates the same (+ or -1) number of files.
|
* Each thread creates the same (+ or -1) number of files.
|
||||||
* File names are pre-generated during initialization.
|
* File names are pre-generated during initialization.
|
||||||
* The created files do not have blocks.
|
* The created files do not have blocks.
|
||||||
@ -578,9 +609,17 @@ void parseArguments(List<String> args) {
|
|||||||
|
|
||||||
@Override
|
@Override
|
||||||
void generateInputs(int[] opsPerThread) throws IOException {
|
void generateInputs(int[] opsPerThread) throws IOException {
|
||||||
assert opsPerThread.length == numThreads : "Error opsPerThread.length";
|
assert opsPerThread.length == numThreads : "Error opsPerThread.length";
|
||||||
clientProto.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_LEAVE,
|
if (!nonSuperUser) {
|
||||||
false);
|
try{
|
||||||
|
clientProto.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_LEAVE,
|
||||||
|
false);
|
||||||
|
} catch (Exception e){
|
||||||
|
LOG.error("Potentially insufficient permission: try running the tool" +
|
||||||
|
"with -nonSuperUser argument or login as super user");
|
||||||
|
throw e;
|
||||||
|
}
|
||||||
|
}
|
||||||
// int generatedFileIdx = 0;
|
// int generatedFileIdx = 0;
|
||||||
LOG.info("Generate " + numOpsRequired + " intputs for " + getOpName());
|
LOG.info("Generate " + numOpsRequired + " intputs for " + getOpName());
|
||||||
LOG.info("basedir: " + getBaseDir());
|
LOG.info("basedir: " + getBaseDir());
|
||||||
@ -614,7 +653,7 @@ String getExecutionArgument(int daemonId) {
|
|||||||
* Do file create.
|
* Do file create.
|
||||||
*/
|
*/
|
||||||
@Override
|
@Override
|
||||||
long executeOp(int daemonId, int inputIdx, String clientName)
|
long executeOp(int daemonId, int inputIdx, String clientName)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
long start = Time.now();
|
long start = Time.now();
|
||||||
clientProto.create(fileNames[daemonId][inputIdx],
|
clientProto.create(fileNames[daemonId][inputIdx],
|
||||||
@ -695,8 +734,16 @@ void parseArguments(List<String> args) {
|
|||||||
@Override
|
@Override
|
||||||
void generateInputs(int[] opsPerThread) throws IOException {
|
void generateInputs(int[] opsPerThread) throws IOException {
|
||||||
assert opsPerThread.length == numThreads : "Error opsPerThread.length";
|
assert opsPerThread.length == numThreads : "Error opsPerThread.length";
|
||||||
clientProto.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_LEAVE,
|
if (!nonSuperUser) {
|
||||||
false);
|
try {
|
||||||
|
clientProto.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_LEAVE,
|
||||||
|
false);
|
||||||
|
} catch (Exception e){
|
||||||
|
LOG.error("Potentially insufficient permission: try running the tool" +
|
||||||
|
" with -nonSuperUser argument or login as super user");
|
||||||
|
throw e;
|
||||||
|
}
|
||||||
|
}
|
||||||
LOG.info("Generate " + numOpsRequired + " inputs for " + getOpName());
|
LOG.info("Generate " + numOpsRequired + " inputs for " + getOpName());
|
||||||
dirPaths = new String[numThreads][];
|
dirPaths = new String[numThreads][];
|
||||||
try {
|
try {
|
||||||
@ -750,14 +797,14 @@ void printResults() {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Open file statistics.
|
* Open file statistics.
|
||||||
*
|
*
|
||||||
* Measure how many open calls (getBlockLocations())
|
* Measure how many open calls (getBlockLocations())
|
||||||
* the name-node can handle per second.
|
* the name-node can handle per second.
|
||||||
*/
|
*/
|
||||||
class OpenFileStats extends CreateFileStats {
|
class OpenFileStats extends CreateFileStats {
|
||||||
// Operation types
|
// Operation types
|
||||||
static final String OP_OPEN_NAME = "open";
|
static final String OP_OPEN_NAME = "open";
|
||||||
static final String OP_USAGE_ARGS =
|
static final String OP_USAGE_ARGS =
|
||||||
" [-threads T] [-files N] [-blockSize S] [-filesPerDir P]"
|
" [-threads T] [-files N] [-blockSize S] [-filesPerDir P]"
|
||||||
+ " [-useExisting] [-baseDirName D]";
|
+ " [-useExisting] [-baseDirName D]";
|
||||||
static final String OP_OPEN_USAGE =
|
static final String OP_OPEN_USAGE =
|
||||||
@ -796,13 +843,16 @@ void generateInputs(int[] opsPerThread) throws IOException {
|
|||||||
String.valueOf(nameGenerator.getFilesPerDirectory()),
|
String.valueOf(nameGenerator.getFilesPerDirectory()),
|
||||||
"-baseDirName", getBaseDirName(),
|
"-baseDirName", getBaseDirName(),
|
||||||
"-close"};
|
"-close"};
|
||||||
CreateFileStats opCreate = new CreateFileStats(Arrays.asList(createArgs));
|
List<String> createArgsList = new ArrayList<String>(Arrays.asList(createArgs));
|
||||||
|
if (this.nonSuperUser){
|
||||||
|
createArgsList.add("-nonSuperUser");
|
||||||
|
}
|
||||||
|
CreateFileStats opCreate = new CreateFileStats(createArgsList);
|
||||||
if(!useExisting) { // create files if they were not created before
|
if(!useExisting) { // create files if they were not created before
|
||||||
opCreate.benchmark();
|
opCreate.benchmark();
|
||||||
LOG.info("Created " + numOpsRequired + " files.");
|
LOG.info("Created " + numOpsRequired + " files.");
|
||||||
} else {
|
} else {
|
||||||
LOG.info("useExisting = true. Assuming "
|
LOG.info("useExisting = true. Assuming "
|
||||||
+ numOpsRequired + " files have been created before.");
|
+ numOpsRequired + " files have been created before.");
|
||||||
}
|
}
|
||||||
// use the same files for open
|
// use the same files for open
|
||||||
@ -820,7 +870,7 @@ void generateInputs(int[] opsPerThread) throws IOException {
|
|||||||
* Do file open.
|
* Do file open.
|
||||||
*/
|
*/
|
||||||
@Override
|
@Override
|
||||||
long executeOp(int daemonId, int inputIdx, String ignore)
|
long executeOp(int daemonId, int inputIdx, String ignore)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
long start = Time.now();
|
long start = Time.now();
|
||||||
clientProto.getBlockLocations(fileNames[daemonId][inputIdx], 0L,
|
clientProto.getBlockLocations(fileNames[daemonId][inputIdx], 0L,
|
||||||
@ -832,13 +882,13 @@ long executeOp(int daemonId, int inputIdx, String ignore)
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Delete file statistics.
|
* Delete file statistics.
|
||||||
*
|
*
|
||||||
* Measure how many delete calls the name-node can handle per second.
|
* Measure how many delete calls the name-node can handle per second.
|
||||||
*/
|
*/
|
||||||
class DeleteFileStats extends OpenFileStats {
|
class DeleteFileStats extends OpenFileStats {
|
||||||
// Operation types
|
// Operation types
|
||||||
static final String OP_DELETE_NAME = "delete";
|
static final String OP_DELETE_NAME = "delete";
|
||||||
static final String OP_DELETE_USAGE =
|
static final String OP_DELETE_USAGE =
|
||||||
"-op " + OP_DELETE_NAME + OP_USAGE_ARGS;
|
"-op " + OP_DELETE_NAME + OP_USAGE_ARGS;
|
||||||
|
|
||||||
DeleteFileStats(List<String> args) {
|
DeleteFileStats(List<String> args) {
|
||||||
@ -851,7 +901,7 @@ String getOpName() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
long executeOp(int daemonId, int inputIdx, String ignore)
|
long executeOp(int daemonId, int inputIdx, String ignore)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
long start = Time.now();
|
long start = Time.now();
|
||||||
clientProto.delete(fileNames[daemonId][inputIdx], false);
|
clientProto.delete(fileNames[daemonId][inputIdx], false);
|
||||||
@ -909,13 +959,13 @@ long executeOp(int daemonId, int inputIdx, String ignore)
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* List file status statistics.
|
* List file status statistics.
|
||||||
*
|
*
|
||||||
* Measure how many get-file-status calls the name-node can handle per second.
|
* Measure how many get-file-status calls the name-node can handle per second.
|
||||||
*/
|
*/
|
||||||
class FileStatusStats extends OpenFileStats {
|
class FileStatusStats extends OpenFileStats {
|
||||||
// Operation types
|
// Operation types
|
||||||
static final String OP_FILE_STATUS_NAME = "fileStatus";
|
static final String OP_FILE_STATUS_NAME = "fileStatus";
|
||||||
static final String OP_FILE_STATUS_USAGE =
|
static final String OP_FILE_STATUS_USAGE =
|
||||||
"-op " + OP_FILE_STATUS_NAME + OP_USAGE_ARGS;
|
"-op " + OP_FILE_STATUS_NAME + OP_USAGE_ARGS;
|
||||||
|
|
||||||
FileStatusStats(List<String> args) {
|
FileStatusStats(List<String> args) {
|
||||||
@ -928,7 +978,7 @@ String getOpName() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
long executeOp(int daemonId, int inputIdx, String ignore)
|
long executeOp(int daemonId, int inputIdx, String ignore)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
long start = Time.now();
|
long start = Time.now();
|
||||||
clientProto.getFileInfo(fileNames[daemonId][inputIdx]);
|
clientProto.getFileInfo(fileNames[daemonId][inputIdx]);
|
||||||
@ -939,13 +989,13 @@ long executeOp(int daemonId, int inputIdx, String ignore)
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Rename file statistics.
|
* Rename file statistics.
|
||||||
*
|
*
|
||||||
* Measure how many rename calls the name-node can handle per second.
|
* Measure how many rename calls the name-node can handle per second.
|
||||||
*/
|
*/
|
||||||
class RenameFileStats extends OpenFileStats {
|
class RenameFileStats extends OpenFileStats {
|
||||||
// Operation types
|
// Operation types
|
||||||
static final String OP_RENAME_NAME = "rename";
|
static final String OP_RENAME_NAME = "rename";
|
||||||
static final String OP_RENAME_USAGE =
|
static final String OP_RENAME_USAGE =
|
||||||
"-op " + OP_RENAME_NAME + OP_USAGE_ARGS;
|
"-op " + OP_RENAME_NAME + OP_USAGE_ARGS;
|
||||||
|
|
||||||
protected String[][] destNames;
|
protected String[][] destNames;
|
||||||
@ -972,7 +1022,7 @@ void generateInputs(int[] opsPerThread) throws IOException {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
long executeOp(int daemonId, int inputIdx, String ignore)
|
long executeOp(int daemonId, int inputIdx, String ignore)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
long start = Time.now();
|
long start = Time.now();
|
||||||
clientProto.rename(fileNames[daemonId][inputIdx],
|
clientProto.rename(fileNames[daemonId][inputIdx],
|
||||||
@ -988,10 +1038,10 @@ long executeOp(int daemonId, int inputIdx, String ignore)
|
|||||||
private static class TinyDatanode implements Comparable<String> {
|
private static class TinyDatanode implements Comparable<String> {
|
||||||
private static final long DF_CAPACITY = 100*1024*1024;
|
private static final long DF_CAPACITY = 100*1024*1024;
|
||||||
private static final long DF_USED = 0;
|
private static final long DF_USED = 0;
|
||||||
|
|
||||||
NamespaceInfo nsInfo;
|
NamespaceInfo nsInfo;
|
||||||
DatanodeRegistration dnRegistration;
|
DatanodeRegistration dnRegistration;
|
||||||
DatanodeStorage storage; //only one storage
|
DatanodeStorage storage; //only one storage
|
||||||
final List<BlockReportReplica> blocks;
|
final List<BlockReportReplica> blocks;
|
||||||
int nrBlocks; // actual number of blocks
|
int nrBlocks; // actual number of blocks
|
||||||
BlockListAsLongs blockReportList;
|
BlockListAsLongs blockReportList;
|
||||||
@ -1124,7 +1174,7 @@ int replicateBlocks() throws IOException {
|
|||||||
* Just report on behalf of the other data-node
|
* Just report on behalf of the other data-node
|
||||||
* that the blocks have been received.
|
* that the blocks have been received.
|
||||||
*/
|
*/
|
||||||
private int transferBlocks( Block blocks[],
|
private int transferBlocks( Block blocks[],
|
||||||
DatanodeInfo xferTargets[][],
|
DatanodeInfo xferTargets[][],
|
||||||
String targetStorageIDs[][]
|
String targetStorageIDs[][]
|
||||||
) throws IOException {
|
) throws IOException {
|
||||||
@ -1152,7 +1202,7 @@ private int transferBlocks( Block blocks[],
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Block report statistics.
|
* Block report statistics.
|
||||||
*
|
*
|
||||||
* Each thread here represents its own data-node.
|
* Each thread here represents its own data-node.
|
||||||
* Data-nodes send the same block report each time.
|
* Data-nodes send the same block report each time.
|
||||||
* The block report may contain missing or non-existing blocks.
|
* The block report may contain missing or non-existing blocks.
|
||||||
@ -1224,7 +1274,7 @@ void parseArguments(List<String> args) {
|
|||||||
@Override
|
@Override
|
||||||
void generateInputs(int[] ignore) throws IOException {
|
void generateInputs(int[] ignore) throws IOException {
|
||||||
int nrDatanodes = getNumDatanodes();
|
int nrDatanodes = getNumDatanodes();
|
||||||
int nrBlocks = (int)Math.ceil((double)blocksPerReport * nrDatanodes
|
int nrBlocks = (int)Math.ceil((double)blocksPerReport * nrDatanodes
|
||||||
/ replication);
|
/ replication);
|
||||||
int nrFiles = (int)Math.ceil((double)nrBlocks / blocksPerFile);
|
int nrFiles = (int)Math.ceil((double)nrBlocks / blocksPerFile);
|
||||||
datanodes = new TinyDatanode[nrDatanodes];
|
datanodes = new TinyDatanode[nrDatanodes];
|
||||||
@ -1235,13 +1285,21 @@ void generateInputs(int[] ignore) throws IOException {
|
|||||||
datanodes[idx].sendHeartbeat();
|
datanodes[idx].sendHeartbeat();
|
||||||
}
|
}
|
||||||
|
|
||||||
// create files
|
// create files
|
||||||
LOG.info("Creating " + nrFiles + " files with " + blocksPerFile + " blocks each.");
|
LOG.info("Creating " + nrFiles + " files with " + blocksPerFile + " blocks each.");
|
||||||
FileNameGenerator nameGenerator;
|
FileNameGenerator nameGenerator;
|
||||||
nameGenerator = new FileNameGenerator(getBaseDir(), 100);
|
nameGenerator = new FileNameGenerator(getBaseDir(), 100);
|
||||||
String clientName = getClientName(007);
|
String clientName = getClientName(007);
|
||||||
clientProto.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_LEAVE,
|
if (!nonSuperUser) {
|
||||||
false);
|
try {
|
||||||
|
clientProto.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_LEAVE,
|
||||||
|
false);
|
||||||
|
} catch (Exception e){
|
||||||
|
LOG.error("Potentially insufficient permission: try running the tool" +
|
||||||
|
" with -nonSuperUser argument or login as super user");
|
||||||
|
throw e;
|
||||||
|
}
|
||||||
|
}
|
||||||
for(int idx=0; idx < nrFiles; idx++) {
|
for(int idx=0; idx < nrFiles; idx++) {
|
||||||
String fileName = nameGenerator.getNextFileName("ThroughputBench");
|
String fileName = nameGenerator.getNextFileName("ThroughputBench");
|
||||||
clientProto.create(fileName, FsPermission.getDefault(), clientName,
|
clientProto.create(fileName, FsPermission.getDefault(), clientName,
|
||||||
@ -1375,7 +1433,7 @@ void printResults() {
|
|||||||
*/
|
*/
|
||||||
class ReplicationStats extends OperationStatsBase {
|
class ReplicationStats extends OperationStatsBase {
|
||||||
static final String OP_REPLICATION_NAME = "replication";
|
static final String OP_REPLICATION_NAME = "replication";
|
||||||
static final String OP_REPLICATION_USAGE =
|
static final String OP_REPLICATION_USAGE =
|
||||||
"-op replication [-datanodes T] [-nodesToDecommission D] " +
|
"-op replication [-datanodes T] [-nodesToDecommission D] " +
|
||||||
"[-nodeReplicationLimit C] [-totalBlocks B] [-blockSize S] "
|
"[-nodeReplicationLimit C] [-totalBlocks B] [-blockSize S] "
|
||||||
+ "[-replication R] [-baseDirName D]";
|
+ "[-replication R] [-baseDirName D]";
|
||||||
|
@ -100,7 +100,30 @@ public void testNNThroughputAgainstRemoteNN() throws Exception {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
/**
|
||||||
|
* This test runs {@link NNThroughputBenchmark} against a mini DFS cluster with
|
||||||
|
* nonSuperUser option (useful when testing any authorization framework e.g.
|
||||||
|
* Ranger since only super user e.g. hdfs can enter/exit safemode
|
||||||
|
* but any request from super user is not sent for authorization).
|
||||||
|
*/
|
||||||
|
@Test(timeout = 120000)
|
||||||
|
public void testNNThroughputAgainstRemoteNNNonSuperUser() throws Exception {
|
||||||
|
final Configuration conf = new HdfsConfiguration();
|
||||||
|
conf.setInt(DFSConfigKeys.DFS_NAMENODE_MIN_BLOCK_SIZE_KEY, 16);
|
||||||
|
MiniDFSCluster cluster = null;
|
||||||
|
try {
|
||||||
|
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(0).build();
|
||||||
|
cluster.waitActive();
|
||||||
|
final Configuration benchConf = new HdfsConfiguration();
|
||||||
|
benchConf.setInt(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, 16);
|
||||||
|
FileSystem.setDefaultUri(benchConf, cluster.getURI());
|
||||||
|
NNThroughputBenchmark.runBenchmark(benchConf, new String[]{"-op", "all", "-nonSuperUser"});
|
||||||
|
} finally {
|
||||||
|
if (cluster != null) {
|
||||||
|
cluster.shutdown();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
/**
|
/**
|
||||||
* This test runs {@link NNThroughputBenchmark} against a mini DFS cluster
|
* This test runs {@link NNThroughputBenchmark} against a mini DFS cluster
|
||||||
* with explicit -fs option.
|
* with explicit -fs option.
|
||||||
|
Loading…
Reference in New Issue
Block a user