All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.ontology2.bakemono.mapred.RealMultipleOutputs Maven / Gradle / Ivy

There is a newer version: 3.2
Show newest version
//
// copied from Hadoop 1.1.2's "MultipleOutputs" class and delicately
// hacked
//

package com.ontology2.bakemono.mapred;

import org.apache.hadoop.fs.Path;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.classification.InterfaceStability;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.mapreduce.*;
import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
import org.apache.hadoop.util.ReflectionUtils;

import java.io.IOException;
import java.util.*;

/* If you want Javadoc,  this is different from the "real" MultipleOutputs class
 * but overall similar in character.  That javadoc could be copied over and then
 * pasted up in here 
 * 
 */

@InterfaceAudience.Public
@InterfaceStability.Stable
public class RealMultipleOutputs {

  private static final String MULTIPLE_OUTPUTS = "mapreduce.realmultipleoutputs";

  private static final String MO_PREFIX = 
    MULTIPLE_OUTPUTS+".namedOutput.";

  private static final String FORMAT = ".format";
  private static final String KEY = ".key";
  private static final String VALUE = ".value";
  private static final String HDFS_PATH = ".hdfsPath";
  private static final String COUNTERS_ENABLED = 
    MULTIPLE_OUTPUTS+".counters";
  
  //
  // copied from FileOutputFormat,  used later in inlining of package-visible
  // code from that class
  //
  
  protected static final String BASE_OUTPUT_NAME = "mapreduce.output.basename";

  /**
   * Counters group used by the counters of MultipleOutputs.
   */
  private static final String COUNTERS_GROUP = RealMultipleOutputs.class.getName();

  /**
   * Cache for the taskContexts
   */
  private Map taskContexts = new HashMap();
  /**
   * Cached TaskAttemptContext which uses the job's configured settings
   */
  private TaskAttemptContext jobOutputFormatContext;

  /**
   * Checks if a named output name is valid token.
   *
   * @param namedOutput named output Name
   * @throws IllegalArgumentException if the output name is not valid.
   */
  private static void checkTokenName(String namedOutput) {
    if (namedOutput == null || namedOutput.length() == 0) {
      throw new IllegalArgumentException(
        "Name cannot be NULL or emtpy");
    }
    for (char ch : namedOutput.toCharArray()) {
      if ((ch >= 'A') && (ch <= 'Z')) {
        continue;
      }
      if ((ch >= 'a') && (ch <= 'z')) {
        continue;
      }
      if ((ch >= '0') && (ch <= '9')) {
        continue;
      }
      throw new IllegalArgumentException(
        "Name cannot be have a '" + ch + "' char");
    }
  }

  /**
   * Checks if output name is valid.
   *
   * name cannot be the name used for the default output
   * @param outputPath base output Name
   * @throws IllegalArgumentException if the output name is not valid.
   */
  private static void checkBaseOutputPath(String outputPath) {
    if (outputPath.equals("part")) {
      throw new IllegalArgumentException("output name cannot be 'part'");
    }
  }
  
  /**
   * Checks if a named output name is valid.
   *
   * @param namedOutput named output Name
   * @throws IllegalArgumentException if the output name is not valid.
   */
  private static void checkNamedOutputName(JobContext job,
      String namedOutput, boolean alreadyDefined) {
    checkTokenName(namedOutput);
    checkBaseOutputPath(namedOutput);
    List definedChannels = getNamedOutputsList(job);
    if (alreadyDefined && definedChannels.contains(namedOutput)) {
      throw new IllegalArgumentException("Named output '" + namedOutput +
        "' already alreadyDefined");
    } else if (!alreadyDefined && !definedChannels.contains(namedOutput)) {
      throw new IllegalArgumentException("Named output '" + namedOutput +
        "' not defined");
    }
  }

  // Returns list of channel names -- infovore made this package scope so this
  // can be seen when we are initializating the committer
  
  static List getNamedOutputsList(JobContext job) {
    List names = new ArrayList();
    StringTokenizer st = new StringTokenizer(
      job.getConfiguration().get(MULTIPLE_OUTPUTS, ""), " ");
    while (st.hasMoreTokens()) {
      names.add(st.nextToken());
    }
    return names;
  }

  // Returns the named output OutputFormat.
  @SuppressWarnings("unchecked")
  private static Class> getNamedOutputFormatClass(
    JobContext job, String namedOutput) {
    return (Class>)
      job.getConfiguration().getClass(MO_PREFIX + namedOutput + FORMAT, null,
      OutputFormat.class);
  }

  // Returns the key class for a named output.
  private static Class getNamedOutputKeyClass(JobContext job,
                                                String namedOutput) {
    return job.getConfiguration().getClass(MO_PREFIX + namedOutput + KEY, null,
      Object.class);
  }

  // Returns the value class for a named output.
  private static Class getNamedOutputValueClass(
      JobContext job, String namedOutput) {
    return job.getConfiguration().getClass(MO_PREFIX + namedOutput + VALUE,
      null, Object.class);
  }
  
  // package scope in Infovore
  static String getHdfsPath(JobContext job, String namedOutput) {
      return job.getConfiguration().get(MO_PREFIX + namedOutput + HDFS_PATH);
  };

  /**
   * Adds a named output for the job.
   * 

* * @param job job to add the named output * @param namedOutput named output name, it has to be a word, letters * and numbers only, cannot be the word 'part' as * that is reserved for the default output. * * The named output is a key used internally that * references the hdfsPath give as the next argument * * @param Path to output in HDFS * @param outputFormatClass OutputFormat class. * @param keyClass key class * @param valueClass value class */ @SuppressWarnings("unchecked") public static void addNamedOutput(Job job, String namedOutput, Path hdfsPath, Class outputFormatClass, Class keyClass, Class valueClass) { checkNamedOutputName(job, namedOutput, true); Configuration conf = job.getConfiguration(); conf.set(MULTIPLE_OUTPUTS, conf.get(MULTIPLE_OUTPUTS, "") + " " + namedOutput); conf.setClass(MO_PREFIX + namedOutput + FORMAT, outputFormatClass, OutputFormat.class); conf.setClass(MO_PREFIX + namedOutput + KEY, keyClass, Object.class); conf.setClass(MO_PREFIX + namedOutput + VALUE, valueClass, Object.class); conf.set(MO_PREFIX+namedOutput+HDFS_PATH, hdfsPath.toString()); } /** * Enables or disables counters for the named outputs. * * The counters group is the {@link MultipleOutputs} class name. * The names of the counters are the same as the named outputs. These * counters count the number records written to each output name. * By default these counters are disabled. * * @param job job to enable counters * @param enabled indicates if the counters will be enabled or not. */ public static void setCountersEnabled(Job job, boolean enabled) { job.getConfiguration().setBoolean(COUNTERS_ENABLED, enabled); } /** * Returns if the counters for the named outputs are enabled or not. * By default these counters are disabled. * * @param job the job * @return TRUE if the counters are enabled, FALSE if they are disabled. */ public static boolean getCountersEnabled(JobContext job) { return job.getConfiguration().getBoolean(COUNTERS_ENABLED, false); } /** * Wraps RecordWriter to increment counters. */ @SuppressWarnings("unchecked") private static class RecordWriterWithCounter extends RecordWriter { private RecordWriter writer; private String counterName; private TaskInputOutputContext context; public RecordWriterWithCounter(RecordWriter writer, String counterName, TaskInputOutputContext context) { this.writer = writer; this.counterName = counterName; this.context = context; } @SuppressWarnings({"unchecked"}) public void write(Object key, Object value) throws IOException, InterruptedException { context.getCounter(COUNTERS_GROUP, counterName).increment(1); writer.write(key, value); } public void close(TaskAttemptContext context) throws IOException, InterruptedException { writer.close(context); } } // instance code, to be used from Mapper/Reducer code private TaskInputOutputContext context; private Set namedOutputs; private Map> recordWriters; private boolean countersEnabled; /** * Creates and initializes multiple outputs support, * it should be instantiated in the Mapper/Reducer setup method. * * @param context the TaskInputOutputContext object */ public RealMultipleOutputs( TaskInputOutputContext context) { this.context = context; namedOutputs = Collections.unmodifiableSet( new HashSet(getNamedOutputsList(context))); recordWriters = new HashMap>(); countersEnabled = getCountersEnabled(context); } /** * Write key and value to the namedOutput. * * Output path is a unique file generated for the namedOutput. * For example, {namedOutput}-(m|r)-{part-number} * * @param namedOutput the named output name * @param key the key * @param value the value */ @SuppressWarnings("unchecked") public void write(String namedOutput, K key, V value) throws IOException, InterruptedException { write(namedOutput, key, value, namedOutput); } /** * Write key and value to baseOutputPath using the namedOutput. * * @param namedOutput the named output name * @param key the key * @param value the value * @param baseOutputPath base-output path to write the record to. * Note: Framework will generate unique filename for the baseOutputPath */ @SuppressWarnings("unchecked") public void write(String namedOutput, K key, V value, String baseOutputPath) throws IOException, InterruptedException { checkNamedOutputName(context, namedOutput, false); checkBaseOutputPath(baseOutputPath); if (!namedOutputs.contains(namedOutput)) { throw new IllegalArgumentException("Undefined named output '" + namedOutput + "'"); } TaskAttemptContext taskContext = getContext(namedOutput); getRecordWriter(taskContext, baseOutputPath).write(key, value); } /** * Write key value to an output file name. * * Gets the record writer from job's output format. * Job's output format should be a FileOutputFormat. * * @param key the key * @param value the value * @param baseOutputPath base-output path to write the record to. * Note: Framework will generate unique filename for the baseOutputPath */ @SuppressWarnings("unchecked") public void write(KEYOUT key, VALUEOUT value, String baseOutputPath) throws IOException, InterruptedException { checkBaseOutputPath(baseOutputPath); if (jobOutputFormatContext == null) { jobOutputFormatContext = new TaskAttemptContext(context.getConfiguration(), context.getTaskAttemptID()); } getRecordWriter(jobOutputFormatContext, baseOutputPath).write(key, value); } // by being synchronized MultipleOutputTask can be use with a // MultithreadedMapper. @SuppressWarnings("unchecked") private synchronized RecordWriter getRecordWriter( TaskAttemptContext taskContext, String baseFileName) throws IOException, InterruptedException { // look for record-writer in the cache RecordWriter writer = recordWriters.get(baseFileName); // If not in cache, create a new one if (writer == null) { // in MultipleOutputs, the following commented out line of code was used here // // FileOutputFormat.setOutputName(taskContext, baseFileName); // // we can't do that because this method has package visibility but we can do something // even worse and inline that code // taskContext.getConfiguration().set(BASE_OUTPUT_NAME, baseFileName); try { writer = ((OutputFormat) ReflectionUtils.newInstance( taskContext.getOutputFormatClass(), taskContext.getConfiguration())) .getRecordWriter(taskContext); } catch (ClassNotFoundException e) { throw new IOException(e); } // if counters are enabled, wrap the writer with context // to increment counters if (countersEnabled) { writer = new RecordWriterWithCounter(writer, baseFileName, context); } // add the record-writer to the cache recordWriters.put(baseFileName, writer); } return writer; } // Create a taskAttemptContext for the named output with // output format and output key/value types put in the context private TaskAttemptContext getContext(String nameOutput) throws IOException { TaskAttemptContext taskContext = taskContexts.get(nameOutput); if (taskContext != null) { return taskContext; } taskContext = _getContext(context,nameOutput); taskContexts.put(nameOutput, taskContext); return taskContext; } static TaskAttemptContext _getContext(TaskAttemptContext context,String nameOutput) throws IOException { TaskAttemptContext taskContext; // The following trick leverages the instantiation of a record writer via // the job thus supporting arbitrary output formats; it also bypasses // the lack of the set method we want on Job here. Configuration clonedConfiguration=new Configuration(context.getConfiguration()); clonedConfiguration.set("mapred.output.dir",getHdfsPath(context, nameOutput)); Job job = new Job(clonedConfiguration); job.setOutputFormatClass(getNamedOutputFormatClass(context, nameOutput)); job.setOutputKeyClass(getNamedOutputKeyClass(context, nameOutput)); job.setOutputValueClass(getNamedOutputValueClass(context, nameOutput)); taskContext = new TaskAttemptContext(job.getConfiguration(), context .getTaskAttemptID()); return taskContext; } /** * Closes all the opened outputs. * * This should be called from cleanup method of map/reduce task. * If overridden subclasses must invoke super.close() at the * end of their close() * */ @SuppressWarnings("unchecked") public void close() throws IOException, InterruptedException { for (RecordWriter writer : recordWriters.values()) { writer.close(context); } } }





© 2015 - 2024 Weber Informatics LLC | Privacy Policy