Example usage for org.apache.hadoop.mapreduce RecordWriter subclass-usage

List of usage examples for org.apache.hadoop.mapreduce RecordWriter subclass-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapreduce RecordWriter subclass-usage.

Usage

From source file org.apache.cassandra.hadoop.AbstractColumnFamilyRecordWriter.java

/**
 * The <code>ColumnFamilyRecordWriter</code> maps the output &lt;key, value&gt;
 * pairs to a Cassandra column family. In particular, it applies all mutations
 * in the value, which it associates with the key, and in turn the responsible
 * endpoint.
 *

From source file org.apache.cassandra.hadoop.BulkRecordWriter.java

@Deprecated
public final class BulkRecordWriter extends RecordWriter<ByteBuffer, List<Mutation>>
        implements org.apache.hadoop.mapred.RecordWriter<ByteBuffer, List<Mutation>> {
    public final static String OUTPUT_LOCATION = "mapreduce.output.bulkoutputformat.localdir";
    public final static String BUFFER_SIZE_IN_MB = "mapreduce.output.bulkoutputformat.buffersize";
    public final static String STREAM_THROTTLE_MBITS = "mapreduce.output.bulkoutputformat.streamthrottlembits";

From source file org.apache.cassandra.hadoop.ColumnFamilyRecordWriter.java

/**
 * The <code>ColumnFamilyRecordWriter</code> maps the output &lt;key, value&gt;
 * pairs to a Cassandra column family. In particular, it applies all mutations
 * in the value, which it associates with the key, and in turn the responsible
 * endpoint.
 *

From source file org.apache.cassandra.hadoop.cql3.CqlBulkRecordWriter.java

/**
 * The <code>CqlBulkRecordWriter</code> maps the output &lt;key, value&gt;
 * pairs to a Cassandra column family. In particular, it applies the binded variables
 * in the value to the prepared statement, which it associates with the key, and in 
 * turn the responsible endpoint.
 *

From source file org.apache.cassandra.hadoop.cql3.CqlRecordWriter.java

/**
 * The <code>CqlRecordWriter</code> maps the output &lt;key, value&gt;
 * pairs to a Cassandra table. In particular, it applies the binded variables
 * in the value to the prepared statement, which it associates with the key, and in 
 * turn the responsible endpoint.
 *

From source file org.apache.cassandra.hadoop2.AbstractColumnFamilyRecordWriter.java

/**
 * The <code>ColumnFamilyRecordWriter</code> maps the output &lt;key, value&gt;
 * pairs to a Cassandra column family. In particular, it applies all mutations
 * in the value, which it associates with the key, and in turn the responsible
 * endpoint.
 *

From source file org.apache.cassandra.hadoop2.BulkRecordWriter.java

final class BulkRecordWriter extends RecordWriter<ByteBuffer, List<Mutation>>
        implements org.apache.hadoop.mapred.RecordWriter<ByteBuffer, List<Mutation>> {

    private final static String OUTPUT_LOCATION = "mapreduce.output.bulkoutputformat.localdir";
    private final static String BUFFER_SIZE_IN_MB = "mapreduce.output.bulkoutputformat.buffersize";
    private final static String STREAM_THROTTLE_MBITS = "mapreduce.output.bulkoutputformat.streamthrottlembits";

From source file org.apache.cassandra.io.sstable.CrunchBulkRecordWriter.java

/**
 * This is an almost-copy of {@link org.apache.cassandra.hadoop.BulkRecordWriter}.
 * <p>
 * We had to re-implement this class because of https://issues.apache.org/jira/browse/CASSANDRA-8367
 * </p>
 */

From source file org.apache.crunch.io.avro.trevni.TrevniRecordWriter.java

/**
 *
 */
public class TrevniRecordWriter<T> extends RecordWriter<AvroKey<T>, NullWritable> {

    /** trevni file extension */

From source file org.apache.giraph.bsp.BspRecordWriter.java

/**
 * Used by {@link BspOutputFormat} since some versions of Hadoop
 * require that a RecordWriter is returned from getRecordWriter.
 * Does nothing, except insures that write is never called.
 */
public class BspRecordWriter extends RecordWriter<Text, Text> {