List of usage examples for org.apache.hadoop.mapreduce.lib.output FileOutputFormat subclass-usage
From source file com.ci.backports.avro.mapreduce.AvroOutputFormat.java
/** * A {@link org.apache.hadoop.mapreduce.OutputFormat} for Avro container files. */ public class AvroOutputFormat<T> extends FileOutputFormat<AvroWrapper<T>, NullWritable> { /** Enable output compression using the deflate codec and specify its level.*/
From source file com.ci.backports.hadoop.hbase.ZHFileOutputFormat.java
/**
* Writes HFiles. Passed KeyValues must arrive in order.
* Currently, can only write files to a single column family at a
* time. Multiple column families requires coordinating keys cross family.
* Writes current time as the sequence id for the file. Sets the major compacted
* attribute on created hfiles. Calling write(null,null) will forceably roll
From source file com.citic.zxyjs.zwlscx.mapreduce.lib.input.HFileOutputFormatBase.java
/**
* Writes HFiles. Passed KeyValues must arrive in order. Writes current time as
* the sequence id for the file. Sets the major compacted attribute on created
* hfiles. Calling write(null,null) will forceably roll all HFiles being
* written.
* <p>
From source file com.cloudera.crunch.type.avro.AvroOutputFormat.java
/** An {@link org.apache.hadoop.mapreduce.OutputFormat} for Avro data files. */ public class AvroOutputFormat<T> extends FileOutputFormat<AvroWrapper<T>, NullWritable> { @Override public RecordWriter<AvroWrapper<T>, NullWritable> getRecordWriter(TaskAttemptContext context) throws IOException, InterruptedException {
From source file com.cloudera.recordservice.examples.terasort.TeraOutputFormat.java
/** * An output format that writes the key and value appended together. */ public class TeraOutputFormat extends FileOutputFormat<Text, Text> { static final String FINAL_SYNC_ATTRIBUTE = "mapreduce.terasort.final.sync"; private OutputCommitter committer = null;
From source file com.cloudera.sqoop.mapreduce.AvroOutputFormat.java
/** An {@link org.apache.hadoop.mapred.OutputFormat} for Avro data files. */ public class AvroOutputFormat<T> extends FileOutputFormat<AvroWrapper<T>, NullWritable> { @Override public RecordWriter<AvroWrapper<T>, NullWritable> getRecordWriter(TaskAttemptContext context) throws IOException, InterruptedException {
From source file com.cloudera.sqoop.mapreduce.RawKeyTextOutputFormat.java
/** An {@link OutputFormat} that writes plain text files. * Only writes the key. Does not write any delimiter/newline after the key. */ public class RawKeyTextOutputFormat<K, V> extends FileOutputFormat<K, V> { protected static class RawKeyRecordWriter<K, V> extends RecordWriter<K, V> {
From source file com.datasalt.pangool.solr.TupleSolrOutputFormat.java
/**
* Instantiable OutputFormat that can be used in Pangool for indexing {@link ITuple} in SOLR. It behaves similar to
* SOLR-1301's SolrOutputFormat with the difference that configuration is passed via instance (constructor params). This
* allows us to easily have multiple {@link TupleSolrOutputFormat} in the same Pangool Job. Also, it is much easier to
* configure: it just needs to be instantiated (no need to call multiple static methods to configure it). Everything
* will be configured underneath.
From source file com.datasalt.pangool.tuplemr.avro.AvroOutputFormat.java
/**
* This is the Pangool's version of {@link org.apache.avro.mapred.AvroOutputFormat}.
* It implements the new Hadoop's api in package {@link org.apache.hadoop.mapreduce.lib.output}
* Any {@link AvroOutputFormat} instance is stateful and is not configured
* via {@link Configuration}. Instead, it uses Java-serialization to store its state in
* a Distributed Cache file.
From source file com.datasalt.pangool.tuplemr.mapred.lib.output.ProxyOutputFormat.java
/**
* This special implementation of {@link FileOutputFormat} is used as a proxy for being able to support any type of
* OutputFormat at the same time that we support Multiple Output Formats (also with any type of OutputFormat).
* <p>
* The idea is to use the "_temporary" folder for storing everything (including multiple sub/folders) so that we have
* the full control over the commit / fail process.