Example usage for org.apache.hadoop.mapreduce.lib.output FileOutputFormat subclass-usage

List of usage examples for org.apache.hadoop.mapreduce.lib.output FileOutputFormat subclass-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapreduce.lib.output FileOutputFormat subclass-usage.

Usage

From source file com.ci.backports.avro.mapreduce.AvroOutputFormat.java

/**
 * A {@link org.apache.hadoop.mapreduce.OutputFormat} for Avro container files.
 */
public class AvroOutputFormat<T> extends FileOutputFormat<AvroWrapper<T>, NullWritable> {

    /** Enable output compression using the deflate codec and specify its level.*/

From source file com.ci.backports.hadoop.hbase.ZHFileOutputFormat.java

/**
 * Writes HFiles. Passed KeyValues must arrive in order.
 * Currently, can only write files to a single column family at a
 * time.  Multiple column families requires coordinating keys cross family.
 * Writes current time as the sequence id for the file. Sets the major compacted
 * attribute on created hfiles. Calling write(null,null) will forceably roll

From source file com.citic.zxyjs.zwlscx.mapreduce.lib.input.HFileOutputFormatBase.java

/**
 * Writes HFiles. Passed KeyValues must arrive in order. Writes current time as
 * the sequence id for the file. Sets the major compacted attribute on created
 * hfiles. Calling write(null,null) will forceably roll all HFiles being
 * written.
 * <p>

From source file com.cloudera.crunch.type.avro.AvroOutputFormat.java

/** An {@link org.apache.hadoop.mapreduce.OutputFormat} for Avro data files. */
public class AvroOutputFormat<T> extends FileOutputFormat<AvroWrapper<T>, NullWritable> {

    @Override
    public RecordWriter<AvroWrapper<T>, NullWritable> getRecordWriter(TaskAttemptContext context)
            throws IOException, InterruptedException {

From source file com.cloudera.recordservice.examples.terasort.TeraOutputFormat.java

/**
 * An output format that writes the key and value appended together.
 */
public class TeraOutputFormat extends FileOutputFormat<Text, Text> {
    static final String FINAL_SYNC_ATTRIBUTE = "mapreduce.terasort.final.sync";
    private OutputCommitter committer = null;

From source file com.cloudera.sqoop.mapreduce.AvroOutputFormat.java

/** An {@link org.apache.hadoop.mapred.OutputFormat} for Avro data files. */
public class AvroOutputFormat<T> extends FileOutputFormat<AvroWrapper<T>, NullWritable> {

    @Override
    public RecordWriter<AvroWrapper<T>, NullWritable> getRecordWriter(TaskAttemptContext context)
            throws IOException, InterruptedException {

From source file com.cloudera.sqoop.mapreduce.RawKeyTextOutputFormat.java

/** An {@link OutputFormat} that writes plain text files.
 * Only writes the key. Does not write any delimiter/newline after the key.
 */
public class RawKeyTextOutputFormat<K, V> extends FileOutputFormat<K, V> {

    protected static class RawKeyRecordWriter<K, V> extends RecordWriter<K, V> {

From source file com.datasalt.pangool.solr.TupleSolrOutputFormat.java

/**
 * Instantiable OutputFormat that can be used in Pangool for indexing {@link ITuple} in SOLR. It behaves similar to
 * SOLR-1301's SolrOutputFormat with the difference that configuration is passed via instance (constructor params). This
 * allows us to easily have multiple {@link TupleSolrOutputFormat} in the same Pangool Job. Also, it is much easier to
 * configure: it just needs to be instantiated (no need to call multiple static methods to configure it). Everything
 * will be configured underneath.

From source file com.datasalt.pangool.tuplemr.avro.AvroOutputFormat.java

/**
 * This is the Pangool's version of {@link org.apache.avro.mapred.AvroOutputFormat}.
 * It implements the new Hadoop's api in package {@link org.apache.hadoop.mapreduce.lib.output}
 * Any {@link AvroOutputFormat} instance is stateful and is not configured 
 * via {@link Configuration}. Instead, it uses Java-serialization to store its state in 
 * a Distributed Cache file.  

From source file com.datasalt.pangool.tuplemr.mapred.lib.output.ProxyOutputFormat.java

/**
 * This special implementation of {@link FileOutputFormat} is used as a proxy for being able to support any type of
 * OutputFormat at the same time that we support Multiple Output Formats (also with any type of OutputFormat).
 * <p>
 * The idea is to use the "_temporary" folder for storing everything (including multiple sub/folders) so that we have
 * the full control over the commit / fail process.