List of usage examples for org.apache.hadoop.mapreduce.lib.output FileOutputFormat subclass-usage
From source file org.mrgeo.format.TsvOutputFormat.java
public class TsvOutputFormat extends FileOutputFormat<LongWritable, Geometry> implements Serializable { private static final long serialVersionUID = 1L; @Override public RecordWriter<LongWritable, Geometry> getRecordWriter(TaskAttemptContext context) throws IOException, InterruptedException {
From source file org.mrgeo.hadoop.multipleoutputs.MultipleOutputFormat.java
public class MultipleOutputFormat<K, V> extends FileOutputFormat<K, V> { private static final String base = MultipleOutputFormat.class.getSimpleName(); private static final String ROOT_OUTPUT_FORMAT = base + ".rootOutputFormat"; private FileOutputFormat<K, V> _innerFormat;
From source file org.mrgeo.hdfs.output.image.HdfsMrsPyramidOutputFormat.java
public class HdfsMrsPyramidOutputFormat extends FileOutputFormat<WritableComparable<?>, Writable> { private static final Logger log = LoggerFactory.getLogger(HdfsMrsPyramidOutputFormat.class); @Override public RecordWriter<WritableComparable<?>, Writable> getRecordWriter(TaskAttemptContext context) throws IOException {
From source file org.mrgeo.hdfs.output.MapFileOutputFormat.java
/** * An {@link org.apache.hadoop.mapreduce.OutputFormat} that writes * {@link MapFile}s. */ public class MapFileOutputFormat extends FileOutputFormat<WritableComparable<?>, Writable> {
From source file org.mrgeo.hdfs.vector.CsvOutputFormat.java
/** * Reads tab separated values as geometries. * <p> * http://code.flickr.com/blog/2009/05/21/flickr-shapefiles-public-dataset-10/ */ public class CsvOutputFormat extends FileOutputFormat<FeatureIdWritable, Geometry> implements Serializable {
From source file org.msgpack.hadoop.mapreduce.output.MessagePackOutputFormat.java
public class MessagePackOutputFormat extends FileOutputFormat<NullWritable, MessagePackWritable> { public RecordWriter<NullWritable, MessagePackWritable> getRecordWriter(TaskAttemptContext job) throws IOException, InterruptedException { Configuration conf = job.getConfiguration(); Path file = getDefaultWorkFile(job, ""); FileSystem fs = file.getFileSystem(conf);
From source file org.ojai.json.mapreduce.JSONFileOutputFormat.java
public class JSONFileOutputFormat extends FileOutputFormat<LongWritable, Document> { @Override public RecordWriter<LongWritable, Document> getRecordWriter(TaskAttemptContext job) throws IOException, InterruptedException {
From source file org.schedoscope.export.ftp.outputformat.FtpUploadOutputFormat.java
/** * The FtpUpload output format is responsible to set up the record writers and * (s)ftp connection settings. */ public class FtpUploadOutputFormat<K, V> extends FileOutputFormat<K, V> {
From source file org.seqdoop.hadoop_bam.AnySAMOutputFormat.java
/** An abstract {@link org.apache.hadoop.mapreduce.OutputFormat} for SAM and * BAM files. Only locks down the value type and stores the output format * requested. */ public abstract class AnySAMOutputFormat<K> extends FileOutputFormat<K, SAMRecordWritable> { /** A string property defining the output format to use. The value is read
From source file org.seqdoop.hadoop_bam.BAMOutputFormat.java
/** Currently this only locks down the value type of the {@link * org.apache.hadoop.mapreduce.OutputFormat}: contains no functionality. */ public abstract class BAMOutputFormat<K> extends FileOutputFormat<K, SAMRecordWritable> { }