List of usage examples for org.apache.hadoop.mapreduce.lib.output FileOutputFormat subclass-usage
From source file org.bgi.flexlab.gaea.tools.mapreduce.vcf.sort.VCFSort.java
final class SortOutputFormat<K> extends FileOutputFormat<K, VariantContextWritable> { public static final String INPUT_PATH_PROP = "hadoopbam.vcfsort.inpath"; private KeyIgnoringVCFOutputFormat<K> baseOF; private void initBaseOF(Configuration conf) {
From source file org.bgi.flexlab.gaea.tools.mapreduce.vcfqualitycontrol.VCFQualityControl.java
final class VariantRecalibrationOutputFormat<K> extends FileOutputFormat<K, VariantContextWritable> { private KeyIgnoringVCFOutputFormat<K> baseOF; private void initBaseOF(Configuration conf) { if (baseOF == null)
From source file org.broadinstitute.sting.gatk.hadoop.hbamsrc.BAMOutputFormat.java
/** * Currently this only locks down the value type of the * {@link org.apache.hadoop.mapreduce.OutputFormat}: contains no functionality. */ public abstract class BAMOutputFormat<K> extends FileOutputFormat<K, SAMRecordWritable> { }
From source file org.cloudgraph.mapreduce.GraphXmlOutputFormat.java
/**
* An XML oriented output format based on <code>FileOutputFormat</code>
*
* @see org.cloudgraph.mapreduce.GraphXmlRecordReader
* @see org.cloudgraph.mapreduce.GraphWritable
*
From source file org.diqube.hadoop.DiqubeOutputFormat.java
/**
* {@link OutputFormat} writing .diqube files which can be easily imported to diqube-server.
*
* The result is the same as executing diqube-tool "transpose" on other input files.
*
* @author Bastian Gloeckle
From source file org.gbif.ocurrence.index.solr.SolrOutputFormat.java
public class SolrOutputFormat<K, V> extends FileOutputFormat<K, V> { private static final Log LOG = LogFactory.getLog(SolrOutputFormat.class); /** * The parameter used to pass the solr config zip file information. This will
From source file org.godhuli.rhipe.RHMapFileOutputFormat.java
public class RHMapFileOutputFormat extends FileOutputFormat<RHBytesWritable, RHBytesWritable> { private final static HashPartitioner<RHBytesWritable, RHBytesWritable> hp = new HashPartitioner<RHBytesWritable, RHBytesWritable>(); public static HashPartitioner<RHBytesWritable, RHBytesWritable> getHP() { return hp; }
From source file org.godhuli.rhipe.RHSequenceAsTextOutputFormat.java
/** * An {@link OutputFormat} that writes {@link SequenceFile}s. */ public class RHSequenceAsTextOutputFormat extends FileOutputFormat<RHBytesWritable, RHBytesWritable> { class ElementWriter extends RecordWriter<RHBytesWritable, RHBytesWritable> { final Text a;
From source file org.godhuli.rhipe.RXBinaryOutputFormat.java
public class RXBinaryOutputFormat extends FileOutputFormat<RHBytesWritable, RHBytesWritable> { public RecordWriter<RHBytesWritable, RHBytesWritable> getRecordWriter(TaskAttemptContext context) throws IOException, InterruptedException { Configuration conf = context.getConfiguration(); Path file = getDefaultWorkFile(context, "");
From source file org.godhuli.rhipe.RXTextOutputFormat.java
public class RXTextOutputFormat extends FileOutputFormat<RHBytesWritable, RHBytesWritable> { protected static class RXTextRecordWriter extends RecordWriter<RHBytesWritable, RHBytesWritable> { private byte[] newLine = "\r\n".getBytes(); private static byte[] keyvaluesep = " ".getBytes(); private static final String utf8 = "UTF-8";