Example usage for org.apache.hadoop.mapred FileOutputFormat subclass-usage

List of usage examples for org.apache.hadoop.mapred FileOutputFormat subclass-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapred FileOutputFormat subclass-usage.

Usage

From source file HiveKeyIgnoringBAMOutputFormat.java

public class HiveKeyIgnoringBAMOutputFormat extends FileOutputFormat<Writable, SAMRecordWritable>
        implements HiveOutputFormat<Writable, SAMRecordWritable> {
    private KeyIgnoringBAMOutputFormat<Writable> wrappedOutputFormat = new KeyIgnoringBAMOutputFormat<Writable>();

    private void setSAMHeaderFrom(JobConf job) throws IOException {
        if (wrappedOutputFormat.getSAMHeader() != null)

From source file cn.spark.Case.MyMultipleOutputFormat.java

/**
 * This abstract class extends the FileOutputFormat, allowing to write the
 * output data to different output files. There are three basic use cases for
 * this class.
 * 
 * Case one: This class is used for a map reduce job with at least one reducer.

From source file com.blackberry.logdriver.mapred.BinaryOutputFormat.java

public class BinaryOutputFormat extends FileOutputFormat<BytesWritable, NullWritable> {

    @Override
    public RecordWriter<BytesWritable, NullWritable> getRecordWriter(FileSystem ignored, JobConf job, String name,
            Progressable progress) throws IOException {
        return new BinaryRecordWriter(job);

From source file com.blackberry.logdriver.mapred.boom.ReBoomOutputFormat.java

public class ReBoomOutputFormat extends FileOutputFormat<LogLineData, Text> {

    @Override
    public RecordWriter<LogLineData, Text> getRecordWriter(FileSystem ignored, JobConf job, String name,
            Progressable progress) throws IOException {
        return new ReBoomRecordWriter(this, job);

From source file com.blm.orc.OrcOutputFormat.java

/**
 * A Hive OutputFormat for ORC files.
 */
public class OrcOutputFormat extends FileOutputFormat<NullWritable, OrcSerdeRow>
        implements AcidOutputFormat<NullWritable, OrcSerdeRow> {

From source file com.cloudera.science.avro.streaming.AvroAsJSONOutputFormat.java

/**
 * Some changes introduced in the original code by Mateusz Kobos
 */
public class AvroAsJSONOutputFormat extends FileOutputFormat<Text, Text> {
    public static final String SCHEMA_LITERAL = "output.schema.literal";
    public static final String SCHEMA_URL = "output.schema.url";

From source file com.datascience.hadoop.CsvOutputFormat.java

/**
 * CSV output format.
 *
 * @author <a href="http://github.com/kuujo">Jordan Halterman</a>
 */
public class CsvOutputFormat extends FileOutputFormat<LongWritable, ListWritable<Text>> {

From source file com.digitalpebble.behemoth.solr.LucidWorksOutputFormat.java

public class LucidWorksOutputFormat extends FileOutputFormat<Text, BehemothDocument> {

    public RecordWriter<Text, BehemothDocument> getRecordWriter(FileSystem ignored, JobConf job, String name,
            Progressable progress) throws IOException {

        final LucidWorksWriter writer = new LucidWorksWriter(progress);

From source file com.digitalpebble.behemoth.solr.SOLROutputFormat.java

public class SOLROutputFormat extends FileOutputFormat<Text, BehemothDocument> {

    public RecordWriter<Text, BehemothDocument> getRecordWriter(FileSystem ignored, JobConf job, String name,
            Progressable progress) throws IOException {

        final SOLRWriter writer = new SOLRWriter(progress);

From source file com.dumbomr.mapred.TypedBytesTableOutputFormat.java

/**
 * This is a modified version of
 * org.apache.hadoop.hbase.mapred.TableOutputFormat. Converts Map/Reduce
 * typedbytes output and write it to an HBase table
 */