Example usage for org.apache.hadoop.mapreduce.lib.output FileOutputFormat subclass-usage

List of usage examples for org.apache.hadoop.mapreduce.lib.output FileOutputFormat subclass-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapreduce.lib.output FileOutputFormat subclass-usage.

Usage

From source file info.halo9pan.word2vec.hadoop.terasort.TeraOutputFormat.java

/**
 * An output format that writes the key and value appended together.
 */
public class TeraOutputFormat extends FileOutputFormat<Text, Text> {
    static final String FINAL_SYNC_ATTRIBUTE = "mapreduce.terasort.final.sync";
    private OutputCommitter committer = null;

From source file input_format.MultiHFileOutputFormat.java

public class MultiHFileOutputFormat extends FileOutputFormat<ImmutableBytesWritable, KeyValue> {
    static final String COMPRESSION_CONF_KEY = "hbase.hfileoutputformat.families.compression";

    Map<ImmutableBytesWritable, RecordWriter<ImmutableBytesWritable, KeyValue>> writers = new HashMap<ImmutableBytesWritable, RecordWriter<ImmutableBytesWritable, KeyValue>>();

    /* Data structure to hold a Writer and amount of data written on it. */

From source file io.amient.kafka.hadoop.io.MultiOutputFormat.java

public class MultiOutputFormat extends FileOutputFormat<MsgMetadataWritable, BytesWritable> {

    final private static Logger log = LoggerFactory.getLogger(MultiOutputFormat.class);

    private static final String CONFIG_PATH_FORMAT = "multioutput.path.format";

From source file io.fluo.mapreduce.FluoFileOutputFormat.java

/**
 * This class wraps the {@link AccumuloFileOutputFormat} and converts to Fluo's data format. You can use the static methods on
 * {@link AccumuloFileOutputFormat} to configure the output file.
 * 
 * The intended use of this output format is for seeding an initialized Fluo table on which no transactions have executed.
 * 

From source file it.crs4.seal.demux.DemuxOutputFormat.java

public class DemuxOutputFormat extends FileOutputFormat<Text, SequencedFragment> {
    protected static class DemuxMultiFileLineRecordWriter extends RecordWriter<Text, SequencedFragment>
            implements Configurable {
        protected static final String DEFAULT_OUTPUT_FORMAT = "qseq";
        protected HashMap<Text, RecordWriter<Text, SequencedFragment>> outputs;
        protected FileSystem fs;

From source file it.crs4.seal.prq.PrqOutputFormat.java

/**
 * Output format for the prq format.
 *
 */
public class PrqOutputFormat extends FileOutputFormat<Text, ReadPair> {
    public static class PrqRecordWriter extends RecordWriter<Text, ReadPair> {

From source file kogiri.common.hadoop.io.format.map.BloomMapFileOutputFormat.java

public class BloomMapFileOutputFormat extends FileOutputFormat<WritableComparable<?>, Writable> {

    @Override
    public RecordWriter<WritableComparable<?>, Writable> getRecordWriter(TaskAttemptContext context)
            throws IOException {
        Configuration conf = context.getConfiguration();

From source file mapreduce.bbs.GBKOutputFormat.java

/** An {@link OutputFormat} that writes plain text files. */
@InterfaceAudience.Public
@InterfaceStability.Stable
public class GBKOutputFormat<K, V> extends FileOutputFormat<K, V> {
    public static String SEPERATOR = "mapreduce.output.textoutputformat.separator";

From source file matrixFormat.MatrixOutputFormat.java

/** An {@link OutputFormat} that writes plain text files. */
public class MatrixOutputFormat<K, V> extends FileOutputFormat<K, V> {
    private static final Log LOG = LogFactory.getLog(MatrixOutputFormat.class);

    protected static class MatrixRecordWriter<K, V> extends RecordWriter<K, V> {
        private static final String utf8 = "UTF-8";

From source file microbench.terasort.TeraOutputFormat.java

/**
 * An output format that writes the key and value appended together.
 */
public class TeraOutputFormat extends FileOutputFormat<Text, Text> {
    static final String FINAL_SYNC_ATTRIBUTE = "mapreduce.eastcircle.terasort.final.sync";
    private OutputCommitter committer = null;