List of usage examples for org.apache.hadoop.mapreduce.lib.output FileOutputFormat subclass-usage
From source file info.halo9pan.word2vec.hadoop.terasort.TeraOutputFormat.java
/** * An output format that writes the key and value appended together. */ public class TeraOutputFormat extends FileOutputFormat<Text, Text> { static final String FINAL_SYNC_ATTRIBUTE = "mapreduce.terasort.final.sync"; private OutputCommitter committer = null;
From source file input_format.MultiHFileOutputFormat.java
public class MultiHFileOutputFormat extends FileOutputFormat<ImmutableBytesWritable, KeyValue> { static final String COMPRESSION_CONF_KEY = "hbase.hfileoutputformat.families.compression"; Map<ImmutableBytesWritable, RecordWriter<ImmutableBytesWritable, KeyValue>> writers = new HashMap<ImmutableBytesWritable, RecordWriter<ImmutableBytesWritable, KeyValue>>(); /* Data structure to hold a Writer and amount of data written on it. */
From source file io.amient.kafka.hadoop.io.MultiOutputFormat.java
public class MultiOutputFormat extends FileOutputFormat<MsgMetadataWritable, BytesWritable> { final private static Logger log = LoggerFactory.getLogger(MultiOutputFormat.class); private static final String CONFIG_PATH_FORMAT = "multioutput.path.format";
From source file io.fluo.mapreduce.FluoFileOutputFormat.java
/**
* This class wraps the {@link AccumuloFileOutputFormat} and converts to Fluo's data format. You can use the static methods on
* {@link AccumuloFileOutputFormat} to configure the output file.
*
* The intended use of this output format is for seeding an initialized Fluo table on which no transactions have executed.
*
From source file it.crs4.seal.demux.DemuxOutputFormat.java
public class DemuxOutputFormat extends FileOutputFormat<Text, SequencedFragment> { protected static class DemuxMultiFileLineRecordWriter extends RecordWriter<Text, SequencedFragment> implements Configurable { protected static final String DEFAULT_OUTPUT_FORMAT = "qseq"; protected HashMap<Text, RecordWriter<Text, SequencedFragment>> outputs; protected FileSystem fs;
From source file it.crs4.seal.prq.PrqOutputFormat.java
/** * Output format for the prq format. * */ public class PrqOutputFormat extends FileOutputFormat<Text, ReadPair> { public static class PrqRecordWriter extends RecordWriter<Text, ReadPair> {
From source file kogiri.common.hadoop.io.format.map.BloomMapFileOutputFormat.java
public class BloomMapFileOutputFormat extends FileOutputFormat<WritableComparable<?>, Writable> { @Override public RecordWriter<WritableComparable<?>, Writable> getRecordWriter(TaskAttemptContext context) throws IOException { Configuration conf = context.getConfiguration();
From source file mapreduce.bbs.GBKOutputFormat.java
/** An {@link OutputFormat} that writes plain text files. */ @InterfaceAudience.Public @InterfaceStability.Stable public class GBKOutputFormat<K, V> extends FileOutputFormat<K, V> { public static String SEPERATOR = "mapreduce.output.textoutputformat.separator";
From source file matrixFormat.MatrixOutputFormat.java
/** An {@link OutputFormat} that writes plain text files. */ public class MatrixOutputFormat<K, V> extends FileOutputFormat<K, V> { private static final Log LOG = LogFactory.getLog(MatrixOutputFormat.class); protected static class MatrixRecordWriter<K, V> extends RecordWriter<K, V> { private static final String utf8 = "UTF-8";
From source file microbench.terasort.TeraOutputFormat.java
/** * An output format that writes the key and value appended together. */ public class TeraOutputFormat extends FileOutputFormat<Text, Text> { static final String FINAL_SYNC_ATTRIBUTE = "mapreduce.eastcircle.terasort.final.sync"; private OutputCommitter committer = null;