Example usage for org.apache.hadoop.mapreduce OutputFormat subclass-usage

List of usage examples for org.apache.hadoop.mapreduce OutputFormat subclass-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapreduce OutputFormat subclass-usage.

Usage

From source file org.apache.cassandra.hadoop2.BulkOutputFormat.java

public class BulkOutputFormat extends OutputFormat<ByteBuffer, List<Mutation>>
        implements org.apache.hadoop.mapred.OutputFormat<ByteBuffer, List<Mutation>> {

    @Override
    public void checkOutputSpecs(JobContext context) {
        checkOutputSpecs(context.getConfiguration());

From source file org.apache.crunch.impl.mr.run.CrunchOutputFormat.java

public class CrunchOutputFormat<K, V> extends OutputFormat<K, V> {
    @Override
    public RecordWriter<K, V> getRecordWriter(TaskAttemptContext taskAttemptContext)
            throws IOException, InterruptedException {
        return new RecordWriter<K, V>() {
            @Override

From source file org.apache.crunch.io.hbase.TableOutputFormat.java

class TableOutputFormat<K> extends OutputFormat<K, Writable> {

    private final Log LOG = LogFactory.getLog(TableOutputFormat.class);

    /** Job parameter that specifies the output table. */
    public static final String OUTPUT_TABLE = "hbase.mapred.outputtable";

From source file org.apache.fluo.mapreduce.FluoOutputFormat.java

/**
 * This output format enables the execution of load transaction against a Fluo table using map
 * reduce.
 */
public class FluoOutputFormat extends OutputFormat<Loader, NullWritable> {

From source file org.apache.giraph.bsp.BspOutputFormat.java

/**
 * This is for internal use only.  Allows the vertex output format routines
 * to be called as if a normal Hadoop job.
 */
public class BspOutputFormat extends OutputFormat<Text, Text> {
    /** Class logger */

From source file org.apache.gora.mapreduce.GoraOutputFormat.java

/**
 * {@link OutputFormat} for Hadoop jobs that want to store the job outputs 
 * to a Gora store. 
 * <p>
 * Hadoop jobs can be either configured through static 
 * <code>setOutput()</code> methods, or if the job is not map-only from {@link GoraReducer}.

From source file org.apache.gora.pig.mapreduce.PigGoraOutputFormat.java

public class PigGoraOutputFormat<K, T extends Persistent> extends OutputFormat<K, T> implements Configurable {

    public static final Logger LOG = LoggerFactory.getLogger(PigGoraOutputFormat.class);

    protected Configuration conf;

From source file org.apache.hcatalog.mapreduce.HCatBaseOutputFormat.java

public abstract class HCatBaseOutputFormat extends OutputFormat<WritableComparable<?>, HCatRecord> {

    //  static final private Log LOG = LogFactory.getLog(HCatBaseOutputFormat.class);

    /**
     * @see org.apache.hcatalog.mapreduce.HCatBaseOutputFormat#getTableSchema(org.apache.hadoop.conf.Configuration)

From source file org.apache.hcatalog.mapreduce.MultiOutputFormat.java

/**
 * The MultiOutputFormat class simplifies writing output data to multiple
 * outputs.
 * <p>
 * Multiple output formats can be defined each with its own
 * <code>OutputFormat</code> class, own key class and own value class. Any

From source file org.apache.hcatalog.mapreduce.OutputFormatContainer.java

/**
 *  This container class is used to wrap OutputFormat implementations and augment them with
 *  behavior necessary to work with HCatalog (ie metastore updates, hcatalog delegation tokens, etc).
 *  Containers are also used to provide storage specific implementations of some HCatalog features (ie dynamic partitioning).
 *  Hence users wishing to create storage specific implementations of HCatalog features should implement this class and override
 *  HCatStorageHandler.getOutputFormatContainer(OutputFormat outputFormat) to return the implementation.