HDFS-10403. DiskBalancer: Add cancel command. Contributed by Anu Engineer.
This commit is contained in:
parent
9e5fcb5e40
commit
43eee50966
@ -0,0 +1,137 @@
|
|||||||
|
/*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.hdfs.server.diskbalancer.command;
|
||||||
|
|
||||||
|
import com.google.common.base.Preconditions;
|
||||||
|
import org.apache.commons.cli.CommandLine;
|
||||||
|
import org.apache.commons.codec.digest.DigestUtils;
|
||||||
|
import org.apache.commons.io.IOUtils;
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.fs.FSDataInputStream;
|
||||||
|
import org.apache.hadoop.hdfs.protocol.ClientDatanodeProtocol;
|
||||||
|
import org.apache.hadoop.hdfs.server.diskbalancer.DiskBalancerException;
|
||||||
|
import org.apache.hadoop.hdfs.server.diskbalancer.planner.NodePlan;
|
||||||
|
import org.apache.hadoop.hdfs.tools.DiskBalancer;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Cancels a running plan.
|
||||||
|
*/
|
||||||
|
public class CancelCommand extends Command {
|
||||||
|
/**
|
||||||
|
* Contructs a cancel Command.
|
||||||
|
*
|
||||||
|
* @param conf - Conf
|
||||||
|
*/
|
||||||
|
public CancelCommand(Configuration conf) {
|
||||||
|
super(conf);
|
||||||
|
addValidCommandParameters(DiskBalancer.CANCEL, "Cancels a running plan.");
|
||||||
|
addValidCommandParameters(DiskBalancer.NODE, "Node to run the command " +
|
||||||
|
"against in node:port format.");
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Executes the Client Calls.
|
||||||
|
*
|
||||||
|
* @param cmd - CommandLine
|
||||||
|
*/
|
||||||
|
@Override
|
||||||
|
public void execute(CommandLine cmd) throws Exception {
|
||||||
|
LOG.info("Executing \"Cancel plan\" command.");
|
||||||
|
Preconditions.checkState(cmd.hasOption(DiskBalancer.CANCEL));
|
||||||
|
verifyCommandOptions(DiskBalancer.CANCEL, cmd);
|
||||||
|
|
||||||
|
// We can cancel a plan using datanode address and plan ID
|
||||||
|
// that you can read from a datanode using queryStatus
|
||||||
|
if(cmd.hasOption(DiskBalancer.NODE)) {
|
||||||
|
String nodeAddress = cmd.getOptionValue(DiskBalancer.NODE);
|
||||||
|
String planHash = cmd.getOptionValue(DiskBalancer.CANCEL);
|
||||||
|
cancelPlanUsingHash(nodeAddress, planHash);
|
||||||
|
} else {
|
||||||
|
// Or you can cancel a plan using the plan file. If the user
|
||||||
|
// points us to the plan file, we can compute the hash as well as read
|
||||||
|
// the address of the datanode from the plan file.
|
||||||
|
String planFile = cmd.getOptionValue(DiskBalancer.CANCEL);
|
||||||
|
Preconditions.checkArgument(planFile == null || planFile.isEmpty(),
|
||||||
|
"Invalid plan file specified.");
|
||||||
|
String planData = null;
|
||||||
|
try (FSDataInputStream plan = open(planFile)) {
|
||||||
|
planData = IOUtils.toString(plan);
|
||||||
|
}
|
||||||
|
cancelPlan(planData);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Cancels a running plan.
|
||||||
|
*
|
||||||
|
* @param planData - Plan data.
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
private void cancelPlan(String planData) throws IOException {
|
||||||
|
Preconditions.checkNotNull(planData);
|
||||||
|
NodePlan plan = readPlan(planData);
|
||||||
|
String dataNodeAddress = plan.getNodeName() + ":" + plan.getPort();
|
||||||
|
Preconditions.checkNotNull(dataNodeAddress);
|
||||||
|
ClientDatanodeProtocol dataNode = getDataNodeProxy(dataNodeAddress);
|
||||||
|
String planHash = DigestUtils.sha512Hex(planData);
|
||||||
|
try {
|
||||||
|
dataNode.cancelDiskBalancePlan(planHash);
|
||||||
|
} catch (DiskBalancerException ex) {
|
||||||
|
LOG.error("Cancelling plan on {} failed. Result: {}, Message: {}",
|
||||||
|
plan.getNodeName(), ex.getResult().toString(), ex.getMessage());
|
||||||
|
throw ex;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Cancels a running plan.
|
||||||
|
* @param nodeAddress - Address of the data node.
|
||||||
|
* @param hash - Sha512 hash of the plan, which can be read from datanode
|
||||||
|
* using query status command.
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
private void cancelPlanUsingHash(String nodeAddress, String hash) throws
|
||||||
|
IOException {
|
||||||
|
Preconditions.checkNotNull(nodeAddress);
|
||||||
|
Preconditions.checkNotNull(hash);
|
||||||
|
ClientDatanodeProtocol dataNode = getDataNodeProxy(nodeAddress);
|
||||||
|
try {
|
||||||
|
dataNode.cancelDiskBalancePlan(hash);
|
||||||
|
} catch (DiskBalancerException ex) {
|
||||||
|
LOG.error("Cancelling plan on {} failed. Result: {}, Message: {}",
|
||||||
|
nodeAddress, ex.getResult().toString(), ex.getMessage());
|
||||||
|
throw ex;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Gets extended help for this command.
|
||||||
|
*
|
||||||
|
* @return Help Message
|
||||||
|
*/
|
||||||
|
@Override
|
||||||
|
protected String getHelp() {
|
||||||
|
return "Cancels a running command. e.g -cancel <PlanFile> or -cancel " +
|
||||||
|
"<planID> -node <datanode>";
|
||||||
|
}
|
||||||
|
}
|
@ -31,6 +31,7 @@
|
|||||||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||||
import org.apache.hadoop.hdfs.DFSUtilClient;
|
import org.apache.hadoop.hdfs.DFSUtilClient;
|
||||||
import org.apache.hadoop.hdfs.protocol.ClientDatanodeProtocol;
|
import org.apache.hadoop.hdfs.protocol.ClientDatanodeProtocol;
|
||||||
|
import org.apache.hadoop.hdfs.server.diskbalancer.planner.NodePlan;
|
||||||
import org.apache.hadoop.hdfs.tools.DiskBalancer;
|
import org.apache.hadoop.hdfs.tools.DiskBalancer;
|
||||||
import org.apache.hadoop.hdfs.server.diskbalancer.connectors.ClusterConnector;
|
import org.apache.hadoop.hdfs.server.diskbalancer.connectors.ClusterConnector;
|
||||||
import org.apache.hadoop.hdfs.server.diskbalancer.connectors.ConnectorFactory;
|
import org.apache.hadoop.hdfs.server.diskbalancer.connectors.ConnectorFactory;
|
||||||
@ -39,6 +40,7 @@
|
|||||||
|
|
||||||
import org.apache.hadoop.net.NetUtils;
|
import org.apache.hadoop.net.NetUtils;
|
||||||
import org.apache.hadoop.security.UserGroupInformation;
|
import org.apache.hadoop.security.UserGroupInformation;
|
||||||
|
import org.apache.htrace.fasterxml.jackson.databind.ObjectMapper;
|
||||||
import org.slf4j.Logger;
|
import org.slf4j.Logger;
|
||||||
import org.slf4j.LoggerFactory;
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
@ -392,4 +394,16 @@ protected void addValidCommandParameters(String key, String desc) {
|
|||||||
protected DiskBalancerCluster getCluster() {
|
protected DiskBalancerCluster getCluster() {
|
||||||
return cluster;
|
return cluster;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns a plan from the Json Data.
|
||||||
|
*
|
||||||
|
* @param planData - Json String
|
||||||
|
* @return NodePlan
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
protected NodePlan readPlan(String planData) throws IOException {
|
||||||
|
ObjectMapper mapper = new ObjectMapper();
|
||||||
|
return mapper.readValue(planData, NodePlan.class);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
@ -27,9 +27,9 @@
|
|||||||
import org.apache.hadoop.fs.FSDataInputStream;
|
import org.apache.hadoop.fs.FSDataInputStream;
|
||||||
import org.apache.hadoop.hdfs.protocol.ClientDatanodeProtocol;
|
import org.apache.hadoop.hdfs.protocol.ClientDatanodeProtocol;
|
||||||
import org.apache.hadoop.hdfs.server.diskbalancer.DiskBalancerException;
|
import org.apache.hadoop.hdfs.server.diskbalancer.DiskBalancerException;
|
||||||
import org.apache.hadoop.hdfs.tools.DiskBalancer;
|
|
||||||
import org.apache.hadoop.hdfs.server.diskbalancer.planner.NodePlan;
|
import org.apache.hadoop.hdfs.server.diskbalancer.planner.NodePlan;
|
||||||
import org.apache.htrace.fasterxml.jackson.databind.ObjectMapper;
|
import org.apache.hadoop.hdfs.tools.DiskBalancer;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
|
|
||||||
|
|
||||||
@ -94,17 +94,7 @@ private void submitPlan(String planData) throws IOException {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Returns a plan from the Json Data.
|
|
||||||
*
|
|
||||||
* @param planData - Json String
|
|
||||||
* @return NodePlan
|
|
||||||
* @throws IOException
|
|
||||||
*/
|
|
||||||
private NodePlan readPlan(String planData) throws IOException {
|
|
||||||
ObjectMapper mapper = new ObjectMapper();
|
|
||||||
return mapper.readValue(planData, NodePlan.class);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Gets extended help for this command.
|
* Gets extended help for this command.
|
||||||
|
@ -24,6 +24,7 @@
|
|||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.conf.Configured;
|
import org.apache.hadoop.conf.Configured;
|
||||||
import org.apache.hadoop.hdfs.HdfsConfiguration;
|
import org.apache.hadoop.hdfs.HdfsConfiguration;
|
||||||
|
import org.apache.hadoop.hdfs.server.diskbalancer.command.CancelCommand;
|
||||||
import org.apache.hadoop.hdfs.server.diskbalancer.command.Command;
|
import org.apache.hadoop.hdfs.server.diskbalancer.command.Command;
|
||||||
import org.apache.hadoop.hdfs.server.diskbalancer.command.ExecuteCommand;
|
import org.apache.hadoop.hdfs.server.diskbalancer.command.ExecuteCommand;
|
||||||
import org.apache.hadoop.hdfs.server.diskbalancer.command.PlanCommand;
|
import org.apache.hadoop.hdfs.server.diskbalancer.command.PlanCommand;
|
||||||
@ -105,7 +106,10 @@ public class DiskBalancer extends Configured implements Tool {
|
|||||||
* Reports the status of disk balancer operation.
|
* Reports the status of disk balancer operation.
|
||||||
*/
|
*/
|
||||||
public static final String QUERY = "query";
|
public static final String QUERY = "query";
|
||||||
|
/**
|
||||||
|
* Cancels a running plan.
|
||||||
|
*/
|
||||||
|
public static final String CANCEL = "cancel";
|
||||||
/**
|
/**
|
||||||
* Template for the Before File. It is node.before.json.
|
* Template for the Before File. It is node.before.json.
|
||||||
*/
|
*/
|
||||||
@ -168,6 +172,7 @@ private Options getOpts() {
|
|||||||
addPlanCommands(opts);
|
addPlanCommands(opts);
|
||||||
addExecuteCommands(opts);
|
addExecuteCommands(opts);
|
||||||
addQueryCommands(opts);
|
addQueryCommands(opts);
|
||||||
|
addCancelCommands(opts);
|
||||||
return opts;
|
return opts;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -233,6 +238,19 @@ private void addQueryCommands(Options opt) {
|
|||||||
opt.addOption(query);
|
opt.addOption(query);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Adds cancel command options.
|
||||||
|
* @param opt Options
|
||||||
|
*/
|
||||||
|
private void addCancelCommands(Options opt) {
|
||||||
|
Option cancel = new Option(CANCEL, true, "Cancels a running plan. -cancel" +
|
||||||
|
" <planFile> or -cancel <planID> -node <datanode:port>");
|
||||||
|
opt.addOption(cancel);
|
||||||
|
Option node = new Option(NODE, true, "Name of the datanode in name:port " +
|
||||||
|
"format");
|
||||||
|
opt.addOption(node);
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* This function parses all command line arguments and returns the appropriate
|
* This function parses all command line arguments and returns the appropriate
|
||||||
* values.
|
* values.
|
||||||
@ -271,6 +289,10 @@ private int dispatch(CommandLine cmd, Options opts)
|
|||||||
currentCommand = new QueryCommand(getConf());
|
currentCommand = new QueryCommand(getConf());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if(cmd.hasOption(DiskBalancer.CANCEL)) {
|
||||||
|
currentCommand = new CancelCommand(getConf());
|
||||||
|
}
|
||||||
|
|
||||||
if(currentCommand == null) {
|
if(currentCommand == null) {
|
||||||
HelpFormatter helpFormatter = new HelpFormatter();
|
HelpFormatter helpFormatter = new HelpFormatter();
|
||||||
helpFormatter.printHelp(80, "hdfs diskbalancer -uri [args]",
|
helpFormatter.printHelp(80, "hdfs diskbalancer -uri [args]",
|
||||||
|
Loading…
Reference in New Issue
Block a user