HADOOP-14250. Correct spelling of 'separate' and variants. Contributed by Doris Gu
This commit is contained in:
parent
c8bd5fc7a8
commit
7297e82cb4
@ -45,7 +45,7 @@ public NodeBase() {
|
||||
|
||||
/** Construct a node from its path
|
||||
* @param path
|
||||
* a concatenation of this node's location, the path seperator, and its name
|
||||
* a concatenation of this node's location, the path separator, and its name
|
||||
*/
|
||||
public NodeBase(String path) {
|
||||
path = normalize(path);
|
||||
|
@ -372,8 +372,8 @@ public static String getFormattedTimeWithDiff(String formattedFinishTime,
|
||||
|
||||
/**
|
||||
* Returns an arraylist of strings.
|
||||
* @param str the comma seperated string values
|
||||
* @return the arraylist of the comma seperated string values
|
||||
* @param str the comma separated string values
|
||||
* @return the arraylist of the comma separated string values
|
||||
*/
|
||||
public static String[] getStrings(String str){
|
||||
String delim = ",";
|
||||
@ -384,7 +384,7 @@ public static String[] getStrings(String str){
|
||||
* Returns an arraylist of strings.
|
||||
* @param str the string values
|
||||
* @param delim delimiter to separate the values
|
||||
* @return the arraylist of the seperated string values
|
||||
* @return the arraylist of the separated string values
|
||||
*/
|
||||
public static String[] getStrings(String str, String delim){
|
||||
Collection<String> values = getStringCollection(str, delim);
|
||||
@ -396,7 +396,7 @@ public static String[] getStrings(String str, String delim){
|
||||
|
||||
/**
|
||||
* Returns a collection of strings.
|
||||
* @param str comma seperated string values
|
||||
* @param str comma separated string values
|
||||
* @return an <code>ArrayList</code> of string values
|
||||
*/
|
||||
public static Collection<String> getStringCollection(String str){
|
||||
|
@ -358,7 +358,7 @@ public static class FSConcat implements FileSystemAccess.FileSystemExecutor<Void
|
||||
* Creates a Concat executor.
|
||||
*
|
||||
* @param path target path to concat to.
|
||||
* @param sources comma seperated absolute paths to use as sources.
|
||||
* @param sources comma separated absolute paths to use as sources.
|
||||
*/
|
||||
public FSConcat(String path, String[] sources) {
|
||||
this.sources = new Path[sources.length];
|
||||
|
@ -159,7 +159,7 @@ private void parseOutputKeyValueSpec() {
|
||||
}
|
||||
|
||||
public void configure(JobConf job) {
|
||||
this.fieldSeparator = job.get(FieldSelectionHelper.DATA_FIELD_SEPERATOR,
|
||||
this.fieldSeparator = job.get(FieldSelectionHelper.DATA_FIELD_SEPARATOR,
|
||||
"\t");
|
||||
this.mapOutputKeyValueSpec = job.get(
|
||||
FieldSelectionHelper.MAP_OUTPUT_KEY_VALUE_SPEC, "0-:");
|
||||
|
@ -37,7 +37,7 @@
|
||||
* of the field); if omitted from pos2, it defaults to 0 (the end of the
|
||||
* field). opts are ordering options (any of 'nr' as described above).
|
||||
* We assume that the fields in the key are separated by
|
||||
* {@link JobContext#MAP_OUTPUT_KEY_FIELD_SEPERATOR}
|
||||
* {@link JobContext#MAP_OUTPUT_KEY_FIELD_SEPARATOR}
|
||||
*/
|
||||
@InterfaceAudience.Public
|
||||
@InterfaceStability.Stable
|
||||
|
@ -318,7 +318,13 @@ public interface MRJobConfig {
|
||||
|
||||
public static final String MAP_OUTPUT_VALUE_CLASS = "mapreduce.map.output.value.class";
|
||||
|
||||
public static final String MAP_OUTPUT_KEY_FIELD_SEPERATOR = "mapreduce.map.output.key.field.separator";
|
||||
public static final String MAP_OUTPUT_KEY_FIELD_SEPARATOR = "mapreduce.map.output.key.field.separator";
|
||||
|
||||
/**
|
||||
* @deprecated Use {@link #MAP_OUTPUT_KEY_FIELD_SEPARATOR}
|
||||
*/
|
||||
@Deprecated
|
||||
public static final String MAP_OUTPUT_KEY_FIELD_SEPERATOR = MAP_OUTPUT_KEY_FIELD_SEPARATOR;
|
||||
|
||||
public static final String MAP_LOG_LEVEL = "mapreduce.map.log.level";
|
||||
|
||||
|
@ -60,8 +60,13 @@
|
||||
public class FieldSelectionHelper {
|
||||
|
||||
public static Text emptyText = new Text("");
|
||||
public static final String DATA_FIELD_SEPERATOR =
|
||||
public static final String DATA_FIELD_SEPARATOR =
|
||||
"mapreduce.fieldsel.data.field.separator";
|
||||
/**
|
||||
* @deprecated Use {@link #DATA_FIELD_SEPARATOR}
|
||||
*/
|
||||
@Deprecated
|
||||
public static final String DATA_FIELD_SEPERATOR = DATA_FIELD_SEPARATOR;
|
||||
public static final String MAP_OUTPUT_KEY_VALUE_SPEC =
|
||||
"mapreduce.fieldsel.map.output.key.value.fields.spec";
|
||||
public static final String REDUCE_OUTPUT_KEY_VALUE_SPEC =
|
||||
|
@ -79,7 +79,7 @@ public void setup(Context context)
|
||||
throws IOException, InterruptedException {
|
||||
Configuration conf = context.getConfiguration();
|
||||
this.fieldSeparator =
|
||||
conf.get(FieldSelectionHelper.DATA_FIELD_SEPERATOR, "\t");
|
||||
conf.get(FieldSelectionHelper.DATA_FIELD_SEPARATOR, "\t");
|
||||
this.mapOutputKeyValueSpec =
|
||||
conf.get(FieldSelectionHelper.MAP_OUTPUT_KEY_VALUE_SPEC, "0-:");
|
||||
try {
|
||||
|
@ -77,7 +77,7 @@ public void setup(Context context)
|
||||
Configuration conf = context.getConfiguration();
|
||||
|
||||
this.fieldSeparator =
|
||||
conf.get(FieldSelectionHelper.DATA_FIELD_SEPERATOR, "\t");
|
||||
conf.get(FieldSelectionHelper.DATA_FIELD_SEPARATOR, "\t");
|
||||
|
||||
this.reduceOutputKeyValueSpec =
|
||||
conf.get(FieldSelectionHelper.REDUCE_OUTPUT_KEY_VALUE_SPEC, "0-:");
|
||||
|
@ -37,9 +37,14 @@
|
||||
@InterfaceAudience.Public
|
||||
@InterfaceStability.Stable
|
||||
public class KeyValueLineRecordReader extends RecordReader<Text, Text> {
|
||||
public static final String KEY_VALUE_SEPERATOR =
|
||||
"mapreduce.input.keyvaluelinerecordreader.key.value.separator";
|
||||
|
||||
public static final String KEY_VALUE_SEPARATOR =
|
||||
"mapreduce.input.keyvaluelinerecordreader.key.value.separator";
|
||||
/**
|
||||
* @deprecated Use {@link #KEY_VALUE_SEPARATOR}
|
||||
*/
|
||||
@Deprecated
|
||||
public static final String KEY_VALUE_SEPERATOR = KEY_VALUE_SEPARATOR;
|
||||
|
||||
private final LineRecordReader lineRecordReader;
|
||||
|
||||
private byte separator = (byte) '\t';
|
||||
@ -56,7 +61,7 @@ public KeyValueLineRecordReader(Configuration conf)
|
||||
throws IOException {
|
||||
|
||||
lineRecordReader = new LineRecordReader();
|
||||
String sepStr = conf.get(KEY_VALUE_SEPERATOR, "\t");
|
||||
String sepStr = conf.get(KEY_VALUE_SEPARATOR, "\t");
|
||||
this.separator = (byte) sepStr.charAt(0);
|
||||
}
|
||||
|
||||
|
@ -42,7 +42,12 @@
|
||||
@InterfaceAudience.Public
|
||||
@InterfaceStability.Stable
|
||||
public class TextOutputFormat<K, V> extends FileOutputFormat<K, V> {
|
||||
public static String SEPERATOR = "mapreduce.output.textoutputformat.separator";
|
||||
public static String SEPARATOR = "mapreduce.output.textoutputformat.separator";
|
||||
/**
|
||||
* @deprecated Use {@link #SEPARATOR}
|
||||
*/
|
||||
@Deprecated
|
||||
public static String SEPERATOR = SEPARATOR;
|
||||
protected static class LineRecordWriter<K, V>
|
||||
extends RecordWriter<K, V> {
|
||||
private static final byte[] NEWLINE =
|
||||
@ -107,7 +112,7 @@ void close(TaskAttemptContext context) throws IOException {
|
||||
) throws IOException, InterruptedException {
|
||||
Configuration conf = job.getConfiguration();
|
||||
boolean isCompressed = getCompressOutput(job);
|
||||
String keyValueSeparator= conf.get(SEPERATOR, "\t");
|
||||
String keyValueSeparator= conf.get(SEPARATOR, "\t");
|
||||
CompressionCodec codec = null;
|
||||
String extension = "";
|
||||
if (isCompressed) {
|
||||
|
@ -46,7 +46,7 @@
|
||||
* of the field); if omitted from pos2, it defaults to 0 (the end of the
|
||||
* field). opts are ordering options (any of 'nr' as described above).
|
||||
* We assume that the fields in the key are separated by
|
||||
* {@link JobContext#MAP_OUTPUT_KEY_FIELD_SEPERATOR}.
|
||||
* {@link JobContext#MAP_OUTPUT_KEY_FIELD_SEPARATOR}.
|
||||
*/
|
||||
@InterfaceAudience.Public
|
||||
@InterfaceStability.Stable
|
||||
@ -62,7 +62,7 @@ public class KeyFieldBasedComparator<K, V> extends WritableComparator
|
||||
public void setConf(Configuration conf) {
|
||||
this.conf = conf;
|
||||
String option = conf.get(COMPARATOR_OPTIONS);
|
||||
String keyFieldSeparator = conf.get(MRJobConfig.MAP_OUTPUT_KEY_FIELD_SEPERATOR,"\t");
|
||||
String keyFieldSeparator = conf.get(MRJobConfig.MAP_OUTPUT_KEY_FIELD_SEPARATOR,"\t");
|
||||
keyFieldHelper.setKeyFieldSeparator(keyFieldSeparator);
|
||||
keyFieldHelper.parseOption(option);
|
||||
}
|
||||
|
@ -65,7 +65,7 @@ public void setConf(Configuration conf) {
|
||||
this.conf = conf;
|
||||
keyFieldHelper = new KeyFieldHelper();
|
||||
String keyFieldSeparator =
|
||||
conf.get(MRJobConfig.MAP_OUTPUT_KEY_FIELD_SEPERATOR, "\t");
|
||||
conf.get(MRJobConfig.MAP_OUTPUT_KEY_FIELD_SEPARATOR, "\t");
|
||||
keyFieldHelper.setKeyFieldSeparator(keyFieldSeparator);
|
||||
if (conf.get("num.key.fields.for.partition") != null) {
|
||||
LOG.warn("Using deprecated num.key.fields.for.partition. " +
|
||||
|
@ -290,7 +290,7 @@ private static void addDeprecatedKeys() {
|
||||
new DeprecationDelta("mapred.mapoutput.value.class",
|
||||
MRJobConfig.MAP_OUTPUT_VALUE_CLASS),
|
||||
new DeprecationDelta("map.output.key.field.separator",
|
||||
MRJobConfig.MAP_OUTPUT_KEY_FIELD_SEPERATOR),
|
||||
MRJobConfig.MAP_OUTPUT_KEY_FIELD_SEPARATOR),
|
||||
new DeprecationDelta("mapred.map.child.log.level",
|
||||
MRJobConfig.MAP_LOG_LEVEL),
|
||||
new DeprecationDelta("mapred.inmem.merge.threshold",
|
||||
@ -412,7 +412,7 @@ private static void addDeprecatedKeys() {
|
||||
ControlledJob.CREATE_DIR),
|
||||
new DeprecationDelta("mapred.data.field.separator",
|
||||
org.apache.hadoop.mapreduce.lib.fieldsel.
|
||||
FieldSelectionHelper.DATA_FIELD_SEPERATOR),
|
||||
FieldSelectionHelper.DATA_FIELD_SEPARATOR),
|
||||
new DeprecationDelta("map.output.key.value.fields.spec",
|
||||
org.apache.hadoop.mapreduce.lib.fieldsel.
|
||||
FieldSelectionHelper.MAP_OUTPUT_KEY_VALUE_SPEC),
|
||||
@ -427,7 +427,7 @@ private static void addDeprecatedKeys() {
|
||||
CombineFileInputFormat.SPLIT_MINSIZE_PERRACK),
|
||||
new DeprecationDelta("key.value.separator.in.input.line",
|
||||
org.apache.hadoop.mapreduce.lib.input.
|
||||
KeyValueLineRecordReader.KEY_VALUE_SEPERATOR),
|
||||
KeyValueLineRecordReader.KEY_VALUE_SEPARATOR),
|
||||
new DeprecationDelta("mapred.linerecordreader.maxlength",
|
||||
org.apache.hadoop.mapreduce.lib.input.
|
||||
LineRecordReader.MAX_LINE_LENGTH),
|
||||
@ -436,7 +436,7 @@ private static void addDeprecatedKeys() {
|
||||
LazyOutputFormat.OUTPUT_FORMAT),
|
||||
new DeprecationDelta("mapred.textoutputformat.separator",
|
||||
org.apache.hadoop.mapreduce.lib.output.
|
||||
TextOutputFormat.SEPERATOR),
|
||||
TextOutputFormat.SEPARATOR),
|
||||
new DeprecationDelta("mapred.join.expr",
|
||||
org.apache.hadoop.mapreduce.lib.join.
|
||||
CompositeInputFormat.JOIN_EXPR),
|
||||
|
@ -65,7 +65,7 @@ static enum OutputType {
|
||||
int place = key.indexOf(TYPE_SEP);
|
||||
if (place == -1) {
|
||||
throw new IllegalArgumentException(
|
||||
"Invalid key format - no type seperator - " + TYPE_SEP);
|
||||
"Invalid key format - no type separator - " + TYPE_SEP);
|
||||
}
|
||||
try {
|
||||
dataType = OutputType.valueOf(
|
||||
@ -78,7 +78,7 @@ static enum OutputType {
|
||||
place = key.indexOf(MEASUREMENT_SEP);
|
||||
if (place == -1) {
|
||||
throw new IllegalArgumentException(
|
||||
"Invalid key format - no measurement seperator - " + MEASUREMENT_SEP);
|
||||
"Invalid key format - no measurement separator - " + MEASUREMENT_SEP);
|
||||
}
|
||||
opType = key.substring(0, place);
|
||||
measurementType = key.substring(place + 1);
|
||||
|
@ -77,7 +77,7 @@ public static void launch() throws Exception {
|
||||
job.setOutputFormat(TextOutputFormat.class);
|
||||
job.setNumReduceTasks(1);
|
||||
|
||||
job.set(FieldSelectionHelper.DATA_FIELD_SEPERATOR, "-");
|
||||
job.set(FieldSelectionHelper.DATA_FIELD_SEPARATOR, "-");
|
||||
job.set(FieldSelectionHelper.MAP_OUTPUT_KEY_VALUE_SPEC, "6,5,1-3:0-");
|
||||
job.set(FieldSelectionHelper.REDUCE_OUTPUT_KEY_VALUE_SPEC, ":4,3,2,1,0,0-");
|
||||
|
||||
|
@ -63,7 +63,7 @@ public TestKeyFieldBasedComparator() throws IOException {
|
||||
super(HadoopTestCase.LOCAL_MR, HadoopTestCase.LOCAL_FS, 1, 1);
|
||||
conf = createJobConf();
|
||||
localConf = createJobConf();
|
||||
localConf.set(JobContext.MAP_OUTPUT_KEY_FIELD_SEPERATOR, " ");
|
||||
localConf.set(JobContext.MAP_OUTPUT_KEY_FIELD_SEPARATOR, " ");
|
||||
}
|
||||
|
||||
public void configure(String keySpec, int expect) throws Exception {
|
||||
@ -85,7 +85,7 @@ public void configure(String keySpec, int expect) throws Exception {
|
||||
conf.setOutputKeyComparatorClass(KeyFieldBasedComparator.class);
|
||||
conf.setKeyFieldComparatorOptions(keySpec);
|
||||
conf.setKeyFieldPartitionerOptions("-k1.1,1.1");
|
||||
conf.set(JobContext.MAP_OUTPUT_KEY_FIELD_SEPERATOR, " ");
|
||||
conf.set(JobContext.MAP_OUTPUT_KEY_FIELD_SEPARATOR, " ");
|
||||
conf.setMapperClass(InverseMapper.class);
|
||||
conf.setReducerClass(IdentityReducer.class);
|
||||
if (!fs.mkdirs(testdir)) {
|
||||
|
@ -57,7 +57,7 @@ public static void launch() throws Exception {
|
||||
StringBuffer expectedOutput = new StringBuffer();
|
||||
constructInputOutputData(inputData, expectedOutput, numOfInputLines);
|
||||
|
||||
conf.set(FieldSelectionHelper.DATA_FIELD_SEPERATOR, "-");
|
||||
conf.set(FieldSelectionHelper.DATA_FIELD_SEPARATOR, "-");
|
||||
conf.set(FieldSelectionHelper.MAP_OUTPUT_KEY_VALUE_SPEC, "6,5,1-3:0-");
|
||||
conf.set(
|
||||
FieldSelectionHelper.REDUCE_OUTPUT_KEY_VALUE_SPEC, ":4,3,2,1,0,0-");
|
||||
|
@ -50,7 +50,7 @@ public class TestMRKeyFieldBasedComparator extends HadoopTestCase {
|
||||
public TestMRKeyFieldBasedComparator() throws IOException {
|
||||
super(HadoopTestCase.LOCAL_MR, HadoopTestCase.LOCAL_FS, 1, 1);
|
||||
conf = createJobConf();
|
||||
conf.set(MRJobConfig.MAP_OUTPUT_KEY_FIELD_SEPERATOR, " ");
|
||||
conf.set(MRJobConfig.MAP_OUTPUT_KEY_FIELD_SEPARATOR, " ");
|
||||
}
|
||||
|
||||
private void testComparator(String keySpec, int expect)
|
||||
@ -61,7 +61,7 @@ private void testComparator(String keySpec, int expect)
|
||||
|
||||
conf.set("mapreduce.partition.keycomparator.options", keySpec);
|
||||
conf.set("mapreduce.partition.keypartitioner.options", "-k1.1,1.1");
|
||||
conf.set(MRJobConfig.MAP_OUTPUT_KEY_FIELD_SEPERATOR, " ");
|
||||
conf.set(MRJobConfig.MAP_OUTPUT_KEY_FIELD_SEPARATOR, " ");
|
||||
|
||||
Job job = MapReduceTestUtil.createJob(conf, inDir, outDir, 1, 1,
|
||||
line1 +"\n" + line2 + "\n");
|
||||
|
@ -1897,7 +1897,7 @@ public static boolean isAclEnabled(Configuration conf) {
|
||||
public static final float DEFAULT_TIMELINE_SERVICE_VERSION = 1.0f;
|
||||
|
||||
/**
|
||||
* Comma seperated list of names for UIs hosted in the timeline server
|
||||
* Comma separated list of names for UIs hosted in the timeline server
|
||||
* (For pluggable UIs).
|
||||
*/
|
||||
public static final String TIMELINE_SERVICE_UI_NAMES =
|
||||
|
Loading…
Reference in New Issue
Block a user