org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJob.java Source code

Java tutorial

Introduction

Here is the source code for org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJob.java

Source

/**
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

package org.apache.hadoop.mapreduce.lib.aggregate;

import java.io.IOException;
import java.util.ArrayList;

import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.classification.InterfaceStability;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.InputFormat;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.MRJobConfig;
import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
import org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat;
import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;
import org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob;
import org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl;
import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
import org.apache.hadoop.util.GenericOptionsParser;

/**
 * This is the main class for creating a map/reduce job using Aggregate
 * framework. The Aggregate is a specialization of map/reduce framework,
 * specializing for performing various simple aggregations.
 * 
 * Generally speaking, in order to implement an application using Map/Reduce
 * model, the developer is to implement Map and Reduce functions (and possibly
 * combine function). However, a lot of applications related to counting and
 * statistics computing have very similar characteristics. Aggregate abstracts
 * out the general patterns of these functions and implementing those patterns.
 * In particular, the package provides generic mapper/redducer/combiner 
 * classes, and a set of built-in value aggregators, and a generic utility 
 * class that helps user create map/reduce jobs using the generic class. 
 * The built-in aggregators include:
 * 
 * sum over numeric values count the number of distinct values compute the
 * histogram of values compute the minimum, maximum, media,average, standard
 * deviation of numeric values
 * 
 * The developer using Aggregate will need only to provide a plugin class
 * conforming to the following interface:
 * 
 * public interface ValueAggregatorDescriptor { public ArrayList<Entry>
 * generateKeyValPairs(Object key, Object value); public void
 * configure(Configuration conf); }
 * 
 * The package also provides a base class, ValueAggregatorBaseDescriptor,
 * implementing the above interface. The user can extend the base class and
 * implement generateKeyValPairs accordingly.
 * 
 * The primary work of generateKeyValPairs is to emit one or more key/value
 * pairs based on the input key/value pair. The key in an output key/value pair
 * encode two pieces of information: aggregation type and aggregation id. The
 * value will be aggregated onto the aggregation id according the aggregation
 * type.
 * 
 * This class offers a function to generate a map/reduce job using Aggregate
 * framework. The function takes the following parameters: input directory spec
 * input format (text or sequence file) output directory a file specifying the
 * user plugin class
 * 
 */
@InterfaceAudience.Public
@InterfaceStability.Stable
public class ValueAggregatorJob {

    public static JobControl createValueAggregatorJobs(String args[],
            Class<? extends ValueAggregatorDescriptor>[] descriptors) throws IOException {

        JobControl theControl = new JobControl("ValueAggregatorJobs");
        ArrayList<ControlledJob> dependingJobs = new ArrayList<ControlledJob>();
        Configuration conf = new Configuration();
        if (descriptors != null) {
            conf = setAggregatorDescriptors(descriptors);
        }
        Job job = createValueAggregatorJob(conf, args);
        ControlledJob cjob = new ControlledJob(job, dependingJobs);
        theControl.addJob(cjob);
        return theControl;
    }

    public static JobControl createValueAggregatorJobs(String args[]) throws IOException {
        return createValueAggregatorJobs(args, null);
    }

    /**
     * Create an Aggregate based map/reduce job.
     * 
     * @param conf The configuration for job
     * @param args the arguments used for job creation. Generic hadoop
     * arguments are accepted.
     * @return a Job object ready for submission.
     * 
     * @throws IOException
     * @see GenericOptionsParser
     */
    public static Job createValueAggregatorJob(Configuration conf, String args[]) throws IOException {

        GenericOptionsParser genericParser = new GenericOptionsParser(conf, args);
        args = genericParser.getRemainingArgs();

        if (args.length < 2) {
            System.out.println(
                    "usage: inputDirs outDir " + "[numOfReducer [textinputformat|seq [specfile [jobName]]]]");
            GenericOptionsParser.printGenericCommandUsage(System.out);
            System.exit(2);
        }
        String inputDir = args[0];
        String outputDir = args[1];
        int numOfReducers = 1;
        if (args.length > 2) {
            numOfReducers = Integer.parseInt(args[2]);
        }

        Class<? extends InputFormat> theInputFormat = null;
        if (args.length > 3 && args[3].compareToIgnoreCase("textinputformat") == 0) {
            theInputFormat = TextInputFormat.class;
        } else {
            theInputFormat = SequenceFileInputFormat.class;
        }

        Path specFile = null;

        if (args.length > 4) {
            specFile = new Path(args[4]);
        }

        String jobName = "";

        if (args.length > 5) {
            jobName = args[5];
        }

        if (specFile != null) {
            conf.addResource(specFile);
        }
        String userJarFile = conf.get(ValueAggregatorJobBase.USER_JAR);
        if (userJarFile != null) {
            conf.set(MRJobConfig.JAR, userJarFile);
        }

        Job theJob = Job.getInstance(conf);
        if (userJarFile == null) {
            theJob.setJarByClass(ValueAggregator.class);
        }
        theJob.setJobName("ValueAggregatorJob: " + jobName);

        FileInputFormat.addInputPaths(theJob, inputDir);

        theJob.setInputFormatClass(theInputFormat);

        theJob.setMapperClass(ValueAggregatorMapper.class);
        FileOutputFormat.setOutputPath(theJob, new Path(outputDir));
        theJob.setOutputFormatClass(TextOutputFormat.class);
        theJob.setMapOutputKeyClass(Text.class);
        theJob.setMapOutputValueClass(Text.class);
        theJob.setOutputKeyClass(Text.class);
        theJob.setOutputValueClass(Text.class);
        theJob.setReducerClass(ValueAggregatorReducer.class);
        theJob.setCombinerClass(ValueAggregatorCombiner.class);
        theJob.setNumReduceTasks(numOfReducers);
        return theJob;
    }

    public static Job createValueAggregatorJob(String args[],
            Class<? extends ValueAggregatorDescriptor>[] descriptors) throws IOException {
        return createValueAggregatorJob(setAggregatorDescriptors(descriptors), args);
    }

    public static Configuration setAggregatorDescriptors(Class<? extends ValueAggregatorDescriptor>[] descriptors) {
        Configuration conf = new Configuration();
        conf.setInt(ValueAggregatorJobBase.DESCRIPTOR_NUM, descriptors.length);
        //specify the aggregator descriptors
        for (int i = 0; i < descriptors.length; i++) {
            conf.set(ValueAggregatorJobBase.DESCRIPTOR + i, "UserDefined," + descriptors[i].getName());
        }
        return conf;
    }

    /**
     * create and run an Aggregate based map/reduce job.
     * 
     * @param args the arguments used for job creation
     * @throws IOException
     */
    public static void main(String args[]) throws IOException, InterruptedException, ClassNotFoundException {
        Job job = ValueAggregatorJob.createValueAggregatorJob(new Configuration(), args);
        int ret = job.waitForCompletion(true) ? 0 : 1;
        System.exit(ret);
    }
}