List of usage examples for org.apache.hadoop.mapred FileOutputFormat subclass-usage
From source file HiveKeyIgnoringBAMOutputFormat.java
public class HiveKeyIgnoringBAMOutputFormat extends FileOutputFormat<Writable, SAMRecordWritable> implements HiveOutputFormat<Writable, SAMRecordWritable> { private KeyIgnoringBAMOutputFormat<Writable> wrappedOutputFormat = new KeyIgnoringBAMOutputFormat<Writable>(); private void setSAMHeaderFrom(JobConf job) throws IOException { if (wrappedOutputFormat.getSAMHeader() != null)
From source file cn.spark.Case.MyMultipleOutputFormat.java
/**
* This abstract class extends the FileOutputFormat, allowing to write the
* output data to different output files. There are three basic use cases for
* this class.
*
* Case one: This class is used for a map reduce job with at least one reducer.
From source file com.blackberry.logdriver.mapred.BinaryOutputFormat.java
public class BinaryOutputFormat extends FileOutputFormat<BytesWritable, NullWritable> { @Override public RecordWriter<BytesWritable, NullWritable> getRecordWriter(FileSystem ignored, JobConf job, String name, Progressable progress) throws IOException { return new BinaryRecordWriter(job);
From source file com.blackberry.logdriver.mapred.boom.ReBoomOutputFormat.java
public class ReBoomOutputFormat extends FileOutputFormat<LogLineData, Text> { @Override public RecordWriter<LogLineData, Text> getRecordWriter(FileSystem ignored, JobConf job, String name, Progressable progress) throws IOException { return new ReBoomRecordWriter(this, job);
From source file com.blm.orc.OrcOutputFormat.java
/** * A Hive OutputFormat for ORC files. */ public class OrcOutputFormat extends FileOutputFormat<NullWritable, OrcSerdeRow> implements AcidOutputFormat<NullWritable, OrcSerdeRow> {
From source file com.cloudera.science.avro.streaming.AvroAsJSONOutputFormat.java
/** * Some changes introduced in the original code by Mateusz Kobos */ public class AvroAsJSONOutputFormat extends FileOutputFormat<Text, Text> { public static final String SCHEMA_LITERAL = "output.schema.literal"; public static final String SCHEMA_URL = "output.schema.url";
From source file com.datascience.hadoop.CsvOutputFormat.java
/** * CSV output format. * * @author <a href="http://github.com/kuujo">Jordan Halterman</a> */ public class CsvOutputFormat extends FileOutputFormat<LongWritable, ListWritable<Text>> {
From source file com.digitalpebble.behemoth.solr.LucidWorksOutputFormat.java
public class LucidWorksOutputFormat extends FileOutputFormat<Text, BehemothDocument> { public RecordWriter<Text, BehemothDocument> getRecordWriter(FileSystem ignored, JobConf job, String name, Progressable progress) throws IOException { final LucidWorksWriter writer = new LucidWorksWriter(progress);
From source file com.digitalpebble.behemoth.solr.SOLROutputFormat.java
public class SOLROutputFormat extends FileOutputFormat<Text, BehemothDocument> { public RecordWriter<Text, BehemothDocument> getRecordWriter(FileSystem ignored, JobConf job, String name, Progressable progress) throws IOException { final SOLRWriter writer = new SOLRWriter(progress);
From source file com.dumbomr.mapred.TypedBytesTableOutputFormat.java
/**
* This is a modified version of
* org.apache.hadoop.hbase.mapred.TableOutputFormat. Converts Map/Reduce
* typedbytes output and write it to an HBase table
*/