diff --git a/hadoop-tools/hadoop-aws/src/site/markdown/tools/hadoop-aws/committers.md b/hadoop-tools/hadoop-aws/src/site/markdown/tools/hadoop-aws/committers.md index b92d8fa187..9f3ea58355 100644 --- a/hadoop-tools/hadoop-aws/src/site/markdown/tools/hadoop-aws/committers.md +++ b/hadoop-tools/hadoop-aws/src/site/markdown/tools/hadoop-aws/committers.md @@ -51,7 +51,7 @@ obsolete. ## Introduction: The Commit Problem Apache Hadoop MapReduce (and behind the scenes, Apache Spark) often write -the output of their work to filesystems +the output of their work to filesystems. Normally, Hadoop uses the `FileOutputFormatCommitter` to manage the promotion of files created in a single task attempt to the final output of @@ -68,37 +68,37 @@ process across the cluster may rename a file or directory to the same path. If the rename fails for any reason, either the data is at the original location, or it is at the destination, -in which case the rename actually succeeded. -**The S3 object store and the `s3a://` filesystem client cannot meet these requirements.* +_The S3 object store and the `s3a://` filesystem client cannot meet these requirements._ -Although S3A is (now) consistent, the S3A client still mimics `rename()` +Although S3 is (now) consistent, the S3A client still mimics `rename()` by copying files and then deleting the originals. This can fail partway through, and there is nothing to prevent any other process in the cluster attempting a rename at the same time. As a result, -* If a rename fails, the data is left in an unknown state. +* If a 'rename' fails, the data is left in an unknown state. * If more than one process attempts to commit work simultaneously, the output directory may contain the results of both processes: it is no longer an exclusive operation. -*. Commit time is still -proportional to the amount of data created. It still can't handle task failure. +* Commit time is still proportional to the amount of data created. +It still can't handle task failure. **Using the "classic" `FileOutputCommmitter` to commit work to Amazon S3 risks -loss or corruption of generated data** +loss or corruption of generated data**. -To address these problems there is now explicit support in the `hadop-aws` -module for committing work to Amazon S3 via the S3A filesystem client, -*the S3A Committers* +To address these problems there is now explicit support in the `hadoop-aws` +module for committing work to Amazon S3 via the S3A filesystem client: +*the S3A Committers*. For safe, as well as high-performance output of work to S3, -we need use "a committer" explicitly written to work with S3, treating it as -an object store with special features. +we need to use "a committer" explicitly written to work with S3, +treating it as an object store with special features. -### Background : Hadoop's "Commit Protocol" +### Background: Hadoop's "Commit Protocol" How exactly is work written to its final destination? That is accomplished by a "commit protocol" between the workers and the job manager. @@ -106,10 +106,10 @@ a "commit protocol" between the workers and the job manager. This protocol is implemented in Hadoop MapReduce, with a similar but extended version in Apache Spark: -1. A "Job" is the entire query, with inputs to outputs +1. The "Job" is the entire query. It takes a given set of input and produces some output. 1. The "Job Manager" is the process in charge of choreographing the execution of the job. It may perform some of the actual computation too. -1. The job has "workers", which are processes which work the actual data +1. The job has "workers", which are processes which work with the actual data and write the results. 1. Workers execute "Tasks", which are fractions of the job, a job whose input has been *partitioned* into units of work which can be executed independently. @@ -126,7 +126,7 @@ this "speculation" delivers speedup as it can address the "straggler problem". When multiple workers are working on the same data, only one worker is allowed to write the final output. 1. The entire job may fail (often from the failure of the Job Manager (MR Master, Spark Driver, ...)). -1, The network may partition, with workers isolated from each other or +1. The network may partition, with workers isolated from each other or the process managing the entire commit. 1. Restarted jobs may recover from a failure by reusing the output of all completed tasks (MapReduce with the "v1" algorithm), or just by rerunning everything @@ -137,34 +137,34 @@ What is "the commit protocol" then? It is the requirements on workers as to when their data is made visible, where, for a filesystem, "visible" means "can be seen in the destination directory of the query." -* There is a destination directory of work, "the output directory." -* The final output of tasks must be in this directory *or paths underneath it*. +* There is a destination directory of work: "the output directory". +The final output of tasks must be in this directory *or paths underneath it*. * The intermediate output of a task must not be visible in the destination directory. That is: they must not write directly to the destination. * The final output of a task *may* be visible under the destination. -* The Job Manager makes the decision as to whether a task's data is to be "committed", -be it directly to the final directory or to some intermediate store.. -* Individual workers communicate with the Job manager to manage the commit process: -whether the output is to be *committed* or *aborted* +* Individual workers communicate with the Job manager to manage the commit process. +* The Job Manager makes the decision on if a task's output data is to be "committed", +be it directly to the final directory or to some intermediate store. * When a worker commits the output of a task, it somehow promotes its intermediate work to becoming final. * When a worker aborts a task's output, that output must not become visible (i.e. it is not committed). * Jobs themselves may be committed/aborted (the nature of "when" is not covered here). * After a Job is committed, all its work must be visible. -* And a file `_SUCCESS` may be written to the output directory. +A file named `_SUCCESS` may be written to the output directory. * After a Job is aborted, all its intermediate data is lost. * Jobs may also fail. When restarted, the successor job must be able to clean up all the intermediate and committed work of its predecessor(s). * Task and Job processes measure the intervals between communications with their Application Master and YARN respectively. -When the interval has grown too large they must conclude +When the interval has grown too large, they must conclude that the network has partitioned and that they must abort their work. That's "essentially" it. When working with HDFS and similar filesystems, directory `rename()` is the mechanism used to commit the work of tasks and jobs. + * Tasks write data to task attempt directories under the directory `_temporary` underneath the final destination directory. * When a task is committed, these files are renamed to the destination directory @@ -180,20 +180,19 @@ and restarting the job. whose output is in the job attempt directory, *and only rerunning all uncommitted tasks*. -This algorithm does not works safely or swiftly with AWS S3 storage because -tenames go from being fast, atomic operations to slow operations which can fail partway through. +This algorithm does not work safely or swiftly with AWS S3 storage because +renames go from being fast, atomic operations to slow operations which can fail partway through. This then is the problem which the S3A committers address: - -*How to safely and reliably commit work to Amazon S3 or compatible object store* +*How to safely and reliably commit work to Amazon S3 or compatible object store.* ## Meet the S3A Committers Since Hadoop 3.1, the S3A FileSystem has been accompanied by classes -designed to integrate with the Hadoop and Spark job commit protocols, classes -which interact with the S3A filesystem to reliably commit work work to S3: -*The S3A Committers* +designed to integrate with the Hadoop and Spark job commit protocols, +classes which interact with the S3A filesystem to reliably commit work to S3: +*The S3A Committers*. The underlying architecture of this process is very complex, and covered in [the committer architecture documentation](./committer_architecture.html). @@ -219,8 +218,8 @@ conflict with existing files is resolved. | feature | staging | magic | |--------|---------|---| -| task output destination | local disk | S3A *without completing the write* | -| task commit process | upload data from disk to S3 | list all pending uploads on s3 and write details to job attempt directory | +| task output destination | write to local disk | upload to S3 *without completing the write* | +| task commit process | upload data from disk to S3 *without completing the write* | list all pending uploads on S3 and write details to job attempt directory | | task abort process | delete local disk data | list all pending uploads and abort them | | job commit | list & complete pending uploads | list & complete pending uploads | @@ -228,33 +227,30 @@ The other metric is "maturity". There, the fact that the Staging committers are based on Netflix's production code counts in its favor. -### The Staging Committer +### The Staging Committers -This is based on work from Netflix. It "stages" data into the local filesystem. -It also requires the cluster to have HDFS, so that +This is based on work from Netflix. +It "stages" data into the local filesystem, using URLs with `file://` schemas. -Tasks write to URLs with `file://` schemas. When a task is committed, -its files are listed, uploaded to S3 as incompleted Multipart Uploads. +When a task is committed, its files are listed and uploaded to S3 as incomplete Multipart Uploads. The information needed to complete the uploads is saved to HDFS where it is committed through the standard "v1" commit algorithm. When the Job is committed, the Job Manager reads the lists of pending writes from its HDFS Job destination directory and completes those uploads. -Canceling a task is straightforward: the local directory is deleted with -its staged data. Canceling a job is achieved by reading in the lists of +Canceling a _task_ is straightforward: the local directory is deleted with its staged data. +Canceling a _job_ is achieved by reading in the lists of pending writes from the HDFS job attempt directory, and aborting those uploads. For extra safety, all outstanding multipart writes to the destination directory are aborted. -The staging committer comes in two slightly different forms, with slightly -different conflict resolution policies: +There are two staging committers with slightly different conflict resolution behaviors: - -* **Directory**: the entire directory tree of data is written or overwritten, +* **Directory Committer**: the entire directory tree of data is written or overwritten, as normal. -* **Partitioned**: special handling of partitioned directory trees of the form +* **Partitioned Committer**: special handling of partitioned directory trees of the form `YEAR=2017/MONTH=09/DAY=19`: conflict resolution is limited to the partitions being updated. @@ -265,13 +261,16 @@ directories containing new data. It is intended for use with Apache Spark only. -## Conflict Resolution in the Staging Committers +#### Conflict Resolution in the Staging Committers The Staging committers offer the ability to replace the conflict policy of the execution engine with policy designed to work with the tree of data. This is based on the experience and needs of Netflix, where efficiently adding new data to an existing partitioned directory tree is a common operation. +An XML configuration is shown below. +The default conflict mode if unset would be `append`. + ```xml fs.s3a.committer.staging.conflict-mode @@ -283,40 +282,37 @@ new data to an existing partitioned directory tree is a common operation. ``` -**replace** : when the job is committed (and not before), delete files in +The _Directory Committer_ uses the entire directory tree for conflict resolution. +For this committer, the behavior of each conflict mode is shown below: + +- **replace**: When the job is committed (and not before), delete files in directories into which new data will be written. -**fail**: when there are existing files in the destination, fail the job. +- **fail**: When there are existing files in the destination, fail the job. -**append**: Add new data to the directories at the destination; overwriting +- **append**: Add new data to the directories at the destination; overwriting any with the same name. Reliable use requires unique names for generated files, which the committers generate by default. -The difference between the two staging committers are as follows: +The _Partitioned Committer_ calculates the partitions into which files are added, +the final directories in the tree, and uses that in its conflict resolution process. +For the _Partitioned Committer_, the behavior of each mode is as follows: -The Directory Committer uses the entire directory tree for conflict resolution. -If any file exists at the destination it will fail in job setup; if the resolution -mechanism is "replace" then all existing files will be deleted. - -The partitioned committer calculates the partitions into which files are added, -the final directories in the tree, and uses that in its conflict resolution -process: - - -**replace** : delete all data in the destination partition before committing +- **replace**: Delete all data in the destination _partition_ before committing the new files. -**fail**: fail if there is data in the destination partition, ignoring the state +- **fail**: Fail if there is data in the destination _partition_, ignoring the state of any parallel partitions. -**append**: add the new data. +- **append**: Add the new data to the destination _partition_, + overwriting any files with the same name. -It's intended for use in Apache Spark Dataset operations, rather +The _Partitioned Committer_ is intended for use in Apache Spark Dataset operations, rather than Hadoop's original MapReduce engine, and only in jobs where adding new data to an existing dataset is the desired goal. -Prerequisites for successful work +Prerequisites for success with the _Partitioned Committer_: 1. The output is written into partitions via `PARTITIONED BY` or `partitionedBy()` instructions. @@ -356,19 +352,20 @@ task commit. However, it has extra requirements of the filesystem -1. [Obsolete] It requires a consistent object store. +1. The object store must be consistent. 1. The S3A client must be configured to recognize interactions -with the magic directories and treat them specially. +with the magic directories and treat them as a special case. -Now that Amazon S3 is consistent, the magic committer is enabled by default. +Now that [Amazon S3 is consistent](https://aws.amazon.com/s3/consistency/), +the magic directory path rewriting is enabled by default. -It's also not been field tested to the extent of Netflix's committer; consider -it the least mature of the committers. +The Magic Committer has not been field tested to the extent of Netflix's committer; +consider it the least mature of the committers. -#### Which Committer to Use? +### Which Committer to Use? -1. If you want to create or update existing partitioned data trees in Spark, use thee +1. If you want to create or update existing partitioned data trees in Spark, use the Partitioned Committer. Make sure you have enough hard disk capacity for all staged data. Do not use it in other situations. @@ -398,8 +395,8 @@ This is done in `mapred-default.xml` ``` -What is missing is an explicit choice of committer to use in the property -`fs.s3a.committer.name`; so the classic (and unsafe) file committer is used. +You must also choose which of the S3A committers to use with the `fs.s3a.committer.name` property. +Otherwise, the classic (and unsafe) file committer is used. | `fs.s3a.committer.name` | Committer | |--------|---------| @@ -408,9 +405,7 @@ What is missing is an explicit choice of committer to use in the property | `magic` | the "magic" committer | | `file` | the original and unsafe File committer; (default) | - - -## Using the Directory and Partitioned Staging Committers +## Using the Staging Committers Generated files are initially written to a local directory underneath one of the temporary directories listed in `fs.s3a.buffer.dir`. @@ -422,16 +417,14 @@ The staging committer needs a path in the cluster filesystem Temporary files are saved in HDFS (or other cluster filesystem) under the path `${fs.s3a.committer.staging.tmp.path}/${user}` where `user` is the name of the user running the job. The default value of `fs.s3a.committer.staging.tmp.path` is `tmp/staging`, -Which will be converted at run time to a path under the current user's home directory, -essentially `~/tmp/staging` - so the temporary directory +resulting in the HDFS directory `~/tmp/staging/${user}`. The application attempt ID is used to create a unique path under this directory, resulting in a path `~/tmp/staging/${user}/${application-attempt-id}/` under which summary data of each task's pending commits are managed using the standard `FileOutputFormat` committer. -When a task is committed the data is uploaded under the destination directory. +When a task is committed, the data is uploaded under the destination directory. The policy of how to react if the destination exists is defined by the `fs.s3a.committer.staging.conflict-mode` setting. @@ -442,9 +435,9 @@ the `fs.s3a.committer.staging.conflict-mode` setting. | `append` | Add the new files to the existing directory tree | -## The "Partitioned" Staging Committer +### The "Partitioned" Staging Committer -This committer an extension of the "Directory" committer which has a special conflict resolution +This committer is an extension of the "Directory" committer which has a special conflict resolution policy designed to support operations which insert new data into a directory tree structured using Hive's partitioning strategy: different levels of the tree represent different columns. @@ -471,10 +464,10 @@ logs/YEAR=2017/MONTH=04/ A partitioned structure like this allows for queries using Hive or Spark to filter out files which do not contain relevant data. -What the partitioned committer does is, where the tooling permits, allows callers -to add data to an existing partitioned layout*. +The partitioned committer allows callers to add new data to an existing partitioned layout, +where the application supports it. -More specifically, it does this by having a conflict resolution options which +More specifically, it does this by reducing the scope of conflict resolution to only act on individual partitions, rather than across the entire output tree. | `fs.s3a.committer.staging.conflict-mode` | Meaning | @@ -492,18 +485,18 @@ was written. With the policy of `append`, the new file would be added to the existing set of files. -### Notes +### Notes on using Staging Committers 1. A deep partition tree can itself be a performance problem in S3 and the s3a client, -or, more specifically. a problem with applications which use recursive directory tree +or more specifically a problem with applications which use recursive directory tree walks to work with data. 1. The outcome if you have more than one job trying simultaneously to write data to the same destination with any policy other than "append" is undefined. 1. In the `append` operation, there is no check for conflict with file names. -If, in the example above, the file `log-20170228.avro` already existed, -it would be overridden. Set `fs.s3a.committer.staging.unique-filenames` to `true` +If the file `log-20170228.avro` in the example above already existed, it would be overwritten. +Set `fs.s3a.committer.staging.unique-filenames` to `true` to ensure that a UUID is included in every filename to avoid this. @@ -514,7 +507,11 @@ performance. ### FileSystem client setup -1. Turn the magic on by `fs.s3a.committer.magic.enabled"` +The S3A connector can recognize files created under paths with `__magic/` as a parent directory. +This allows it to handle those files in a special way, such as uploading to a different location +and storing the information needed to complete pending multipart uploads. + +Turn the magic on by setting `fs.s3a.committer.magic.enabled` to `true`: ```xml @@ -526,22 +523,24 @@ performance. ``` - - ### Enabling the committer +Set the committer used by S3A's committer factory to `magic`: + ```xml fs.s3a.committer.name magic - ``` Conflict management is left to the execution engine itself. -## Common Committer Options +## Committer Options Reference +### Common S3A Committer Options + +The table below provides a summary of each option. | Option | Meaning | Default | |--------|---------|---------| @@ -553,19 +552,7 @@ Conflict management is left to the execution engine itself. | `fs.s3a.committer.generate.uuid` | Generate a Job UUID if none is passed down from Spark | `false` | | `fs.s3a.committer.require.uuid` |Require the Job UUID to be passed down from Spark | `false` | - -## Staging committer (Directory and Partitioned) options - - -| Option | Meaning | Default | -|--------|---------|---------| -| `fs.s3a.committer.staging.conflict-mode` | Conflict resolution: `fail`, `append` or `replace`| `append` | -| `fs.s3a.committer.staging.tmp.path` | Path in the cluster filesystem for temporary data. | `tmp/staging` | -| `fs.s3a.committer.staging.unique-filenames` | Generate unique filenames. | `true` | -| `fs.s3a.committer.staging.abort.pending.uploads` | Deprecated; replaced by `fs.s3a.committer.abort.pending.uploads`. | `(false)` | - - -### Common Committer Options +The examples below shows how these options can be configured in XML. ```xml @@ -628,8 +615,8 @@ Conflict management is left to the execution engine itself. fs.s3a.committer.require.uuid false - Should the committer fail to initialize if a unique ID isn't set in - "spark.sql.sources.writeJobUUID" or fs.s3a.committer.staging.uuid + Require the committer fail to initialize if a unique ID is not set in + "spark.sql.sources.writeJobUUID" or "fs.s3a.committer.uuid". This helps guarantee that unique IDs for jobs are being passed down in spark applications. @@ -650,7 +637,14 @@ Conflict management is left to the execution engine itself. ``` -### Staging Committer Options +### Staging committer (Directory and Partitioned) options + +| Option | Meaning | Default | +|--------|---------|---------| +| `fs.s3a.committer.staging.conflict-mode` | Conflict resolution: `fail`, `append`, or `replace`.| `append` | +| `fs.s3a.committer.staging.tmp.path` | Path in the cluster filesystem for temporary data. | `tmp/staging` | +| `fs.s3a.committer.staging.unique-filenames` | Generate unique filenames. | `true` | +| `fs.s3a.committer.staging.abort.pending.uploads` | Deprecated; replaced by `fs.s3a.committer.abort.pending.uploads`. | `(false)` | ```xml @@ -672,7 +666,7 @@ Conflict management is left to the execution engine itself. true Option for final files to have a unique name through job attempt info, - or the value of fs.s3a.committer.staging.uuid + or the value of fs.s3a.committer.uuid. When writing data with the "append" conflict option, this guarantees that new data will not overwrite any existing data. @@ -696,10 +690,9 @@ The magic committer recognizes when files are created under paths with `__magic/ and redirects the upload to a different location, adding the information needed to complete the upload in the job commit operation. -If, for some reason, you *do not* want these paths to be redirected and not manifest until later, +If, for some reason, you *do not* want these paths to be redirected and completed later, the feature can be disabled by setting `fs.s3a.committer.magic.enabled` to false. - -By default it is true. +By default, it is enabled. ```xml @@ -711,6 +704,8 @@ By default it is true. ``` +You will not be able to use the Magic Committer if this option is disabled. + ## Concurrent Jobs writing to the same destination It is sometimes possible for multiple jobs to simultaneously write to the same destination path. @@ -730,7 +725,7 @@ be creating files with paths/filenames unique to the specific job. It is not enough for them to be unique by task `part-00000.snappy.parquet`, because each job will have tasks with the same name, so generate files with conflicting operations. -For the staging committers, setting `fs.s3a.committer.staging.unique-filenames` to ensure unique names are +For the staging committers, enable `fs.s3a.committer.staging.unique-filenames` to ensure unique names are generated during the upload. Otherwise, use what configuration options are available in the specific `FileOutputFormat`. Note: by default, the option `mapreduce.output.basename` sets the base name for files; @@ -757,13 +752,12 @@ org.apache.hadoop.fs.s3a.commit.PathCommitException: `s3a://landsat-pds': Filesy in configuration option fs.s3a.committer.magic.enabled ``` -The Job is configured to use the magic committer, but the S3A bucket has not been explicitly -declared as supporting it. +The Job is configured to use the magic committer, +but the S3A bucket has not been explicitly declared as supporting it. -The Job is configured to use the magic committer, but the S3A bucket has not been explicitly declared as supporting it. - -As this is now true by default, this error will only surface with a configuration which has explicitly disabled it. -Remove all global/per-bucket declarations of `fs.s3a.bucket.magic.enabled` or set them to `true` +Magic Committer support within the S3A filesystem has been enabled by default since Hadoop 3.3.1. +This error will only surface with a configuration which has explicitly disabled it. +Remove all global/per-bucket declarations of `fs.s3a.bucket.magic.enabled` or set them to `true`. ```xml @@ -846,7 +840,7 @@ the failure happen at the start of a job. (Setting this option will not interfere with the Staging Committers' use of HDFS, as it explicitly sets the algorithm to "2" for that part of its work). -The other way to check which committer to use is to examine the `_SUCCESS` file. +The other way to check which committer was used is to examine the `_SUCCESS` file. If it is 0-bytes long, the classic `FileOutputCommitter` committed the job. The S3A committers all write a non-empty JSON file; the `committer` field lists the committer used. @@ -862,7 +856,7 @@ all committers registered for the s3a:// schema. 1. The output format has overridden `FileOutputFormat.getOutputCommitter()` and is returning its own committer -one which is a subclass of `FileOutputCommitter`. -That final cause. *the output format is returning its own committer*, is not +The final cause "the output format is returning its own committer" is not easily fixed; it may be that the custom committer performs critical work during its lifecycle, and contains assumptions about the state of the written data during task and job commit (i.e. it is in the destination filesystem).