Example usage for org.apache.hadoop.mapreduce.lib.output FileOutputFormat subclass-usage

List of usage examples for org.apache.hadoop.mapreduce.lib.output FileOutputFormat subclass-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapreduce.lib.output FileOutputFormat subclass-usage.

Usage

From source file org.bgi.flexlab.gaea.tools.mapreduce.vcf.sort.VCFSort.java

final class SortOutputFormat<K> extends FileOutputFormat<K, VariantContextWritable> {
    public static final String INPUT_PATH_PROP = "hadoopbam.vcfsort.inpath";

    private KeyIgnoringVCFOutputFormat<K> baseOF;

    private void initBaseOF(Configuration conf) {

From source file org.bgi.flexlab.gaea.tools.mapreduce.vcfqualitycontrol.VCFQualityControl.java

final class VariantRecalibrationOutputFormat<K> extends FileOutputFormat<K, VariantContextWritable> {

    private KeyIgnoringVCFOutputFormat<K> baseOF;

    private void initBaseOF(Configuration conf) {
        if (baseOF == null)

From source file org.broadinstitute.sting.gatk.hadoop.hbamsrc.BAMOutputFormat.java

/**
 * Currently this only locks down the value type of the
 * {@link org.apache.hadoop.mapreduce.OutputFormat}: contains no functionality.
 */
public abstract class BAMOutputFormat<K> extends FileOutputFormat<K, SAMRecordWritable> {
}

From source file org.cloudgraph.mapreduce.GraphXmlOutputFormat.java

/**
 * An XML oriented output format based on <code>FileOutputFormat</code>
 * 
 * @see org.cloudgraph.mapreduce.GraphXmlRecordReader
 * @see org.cloudgraph.mapreduce.GraphWritable
 * 

From source file org.diqube.hadoop.DiqubeOutputFormat.java

/**
 * {@link OutputFormat} writing .diqube files which can be easily imported to diqube-server.
 * 
 * The result is the same as executing diqube-tool "transpose" on other input files.
 *
 * @author Bastian Gloeckle

From source file org.gbif.ocurrence.index.solr.SolrOutputFormat.java

public class SolrOutputFormat<K, V> extends FileOutputFormat<K, V> {

    private static final Log LOG = LogFactory.getLog(SolrOutputFormat.class);

    /**
     * The parameter used to pass the solr config zip file information. This will

From source file org.godhuli.rhipe.RHMapFileOutputFormat.java

public class RHMapFileOutputFormat extends FileOutputFormat<RHBytesWritable, RHBytesWritable> {
    private final static HashPartitioner<RHBytesWritable, RHBytesWritable> hp = new HashPartitioner<RHBytesWritable, RHBytesWritable>();

    public static HashPartitioner<RHBytesWritable, RHBytesWritable> getHP() {
        return hp;
    }

From source file org.godhuli.rhipe.RHSequenceAsTextOutputFormat.java

/**
 * An {@link OutputFormat} that writes {@link SequenceFile}s.
 */
public class RHSequenceAsTextOutputFormat extends FileOutputFormat<RHBytesWritable, RHBytesWritable> {
    class ElementWriter extends RecordWriter<RHBytesWritable, RHBytesWritable> {
        final Text a;

From source file org.godhuli.rhipe.RXBinaryOutputFormat.java

public class RXBinaryOutputFormat extends FileOutputFormat<RHBytesWritable, RHBytesWritable> {

    public RecordWriter<RHBytesWritable, RHBytesWritable> getRecordWriter(TaskAttemptContext context)
            throws IOException, InterruptedException {
        Configuration conf = context.getConfiguration();
        Path file = getDefaultWorkFile(context, "");

From source file org.godhuli.rhipe.RXTextOutputFormat.java

public class RXTextOutputFormat extends FileOutputFormat<RHBytesWritable, RHBytesWritable> {

    protected static class RXTextRecordWriter extends RecordWriter<RHBytesWritable, RHBytesWritable> {
        private byte[] newLine = "\r\n".getBytes();
        private static byte[] keyvaluesep = " ".getBytes();
        private static final String utf8 = "UTF-8";