All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.hadoop.mapreduce.OutputCommitter Maven / Gradle / Ivy

The newest version!
/**
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

package org.apache.hadoop.mapreduce;

import java.io.IOException;

import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.classification.InterfaceStability;
/**
 * OutputCommitter describes the commit of task output for a 
 * Map-Reduce job.
 *
 * 

The Map-Reduce framework relies on the OutputCommitter of * the job to:

*

    *
  1. * Setup the job during initialization. For example, create the temporary * output directory for the job during the initialization of the job. *
  2. *
  3. * Cleanup the job after the job completion. For example, remove the * temporary output directory after the job completion. *
  4. *
  5. * Setup the task temporary output. *
  6. *
  7. * Check whether a task needs a commit. This is to avoid the commit * procedure if a task does not need commit. *
  8. *
  9. * Commit of the task output. *
  10. *
  11. * Discard the task commit. *
  12. *
* The methods in this class can be called from several different processes and * from several different contexts. It is important to know which process and * which context each is called from. Each method should be marked accordingly * in its documentation. It is also important to note that not all methods are * guaranteed to be called once and only once. If a method is not guaranteed to * have this property the output committer needs to handle this appropriately. * Also note it will only be in rare situations where they may be called * multiple times for the same task. * * @see org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter * @see JobContext * @see TaskAttemptContext */ @InterfaceAudience.Public @InterfaceStability.Stable public abstract class OutputCommitter { /** * For the framework to setup the job output during initialization. This is * called from the application master process for the entire job. This will be * called multiple times, once per job attempt. * * @param jobContext Context of the job whose output is being written. * @throws IOException if temporary output could not be created */ public abstract void setupJob(JobContext jobContext) throws IOException; /** * For cleaning up the job's output after job completion. This is called * from the application master process for the entire job. This may be called * multiple times. * * @param jobContext Context of the job whose output is being written. * @throws IOException * @deprecated Use {@link #commitJob(JobContext)} and * {@link #abortJob(JobContext, JobStatus.State)} instead. */ @Deprecated public void cleanupJob(JobContext jobContext) throws IOException { } /** * For committing job's output after successful job completion. Note that this * is invoked for jobs with final runstate as SUCCESSFUL. This is called * from the application master process for the entire job. This is guaranteed * to only be called once. If it throws an exception the entire job will * fail. * * @param jobContext Context of the job whose output is being written. * @throws IOException */ public void commitJob(JobContext jobContext) throws IOException { cleanupJob(jobContext); } /** * For aborting an unsuccessful job's output. Note that this is invoked for * jobs with final runstate as {@link JobStatus.State#FAILED} or * {@link JobStatus.State#KILLED}. This is called from the application * master process for the entire job. This may be called multiple times. * * @param jobContext Context of the job whose output is being written. * @param state final runstate of the job * @throws IOException */ public void abortJob(JobContext jobContext, JobStatus.State state) throws IOException { cleanupJob(jobContext); } /** * Sets up output for the task. This is called from each individual task's * process that will output to HDFS, and it is called just for that task. This * may be called multiple times for the same task, but for different task * attempts. * * @param taskContext Context of the task whose output is being written. * @throws IOException */ public abstract void setupTask(TaskAttemptContext taskContext) throws IOException; /** * Check whether task needs a commit. This is called from each individual * task's process that will output to HDFS, and it is called just for that * task. * * @param taskContext * @return true/false * @throws IOException */ public abstract boolean needsTaskCommit(TaskAttemptContext taskContext) throws IOException; /** * To promote the task's temporary output to final output location. * If {@link #needsTaskCommit(TaskAttemptContext)} returns true and this * task is the task that the AM determines finished first, this method * is called to commit an individual task's output. This is to mark * that tasks output as complete, as {@link #commitJob(JobContext)} will * also be called later on if the entire job finished successfully. This * is called from a task's process. This may be called multiple times for the * same task, but different task attempts. It should be very rare for this to * be called multiple times and requires odd networking failures to make this * happen. In the future the Hadoop framework may eliminate this race. * * @param taskContext Context of the task whose output is being written. * @throws IOException if commit is not successful. */ public abstract void commitTask(TaskAttemptContext taskContext) throws IOException; /** * Discard the task output. This is called from a task's process to clean * up a single task's output that can not yet been committed. This may be * called multiple times for the same task, but for different task attempts. * * @param taskContext * @throws IOException */ public abstract void abortTask(TaskAttemptContext taskContext) throws IOException; /** * Is task output recovery supported for restarting jobs? * * If task output recovery is supported, job restart can be done more * efficiently. * * @return true if task output recovery is supported, * false otherwise * @see #recoverTask(TaskAttemptContext) * @deprecated Use {@link #isRecoverySupported(JobContext)} instead. */ @Deprecated public boolean isRecoverySupported() { return false; } /** * Returns true if an in-progress job commit can be retried. If the MR AM is * re-run then it will check this value to determine if it can retry an * in-progress commit that was started by a previous version. * Note that in rare scenarios, the previous AM version might still be running * at that time, due to system anomalies. Hence if this method returns true * then the retry commit operation should be able to run concurrently with * the previous operation. * * If repeatable job commit is supported, job restart can tolerate previous * AM failures during job commit. * * By default, it is not supported. Extended classes (like: * FileOutputCommitter) should explicitly override it if provide support. * * @param jobContext * Context of the job whose output is being written. * @return true repeatable job commit is supported, * false otherwise * @throws IOException */ public boolean isCommitJobRepeatable(JobContext jobContext) throws IOException { return false; } /** * Is task output recovery supported for restarting jobs? * * If task output recovery is supported, job restart can be done more * efficiently. * * @param jobContext * Context of the job whose output is being written. * @return true if task output recovery is supported, * false otherwise * @throws IOException * @see #recoverTask(TaskAttemptContext) */ public boolean isRecoverySupported(JobContext jobContext) throws IOException { return isRecoverySupported(); } /** * Recover the task output. * * The retry-count for the job will be passed via the * {@link MRJobConfig#APPLICATION_ATTEMPT_ID} key in * {@link TaskAttemptContext#getConfiguration()} for the * OutputCommitter. This is called from the application master * process, but it is called individually for each task. * * If an exception is thrown the task will be attempted again. * * This may be called multiple times for the same task. But from different * application attempts. * * @param taskContext Context of the task whose output is being recovered * @throws IOException */ public void recoverTask(TaskAttemptContext taskContext) throws IOException {} }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy