
By default, the mapreduce manifest committer is used for jobs working with abfs and gcs. Hadoop mapreduce will pick this up automatically; for Spark it is a bit complicated: read the docs to see the steps required.
By default, the mapreduce manifest committer is used for jobs working with abfs and gcs. Hadoop mapreduce will pick this up automatically; for Spark it is a bit complicated: read the docs to see the steps required.