Example usage for org.apache.hadoop.mapreduce RecordWriter subclass-usage

List of usage examples for org.apache.hadoop.mapreduce RecordWriter subclass-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapreduce RecordWriter subclass-usage.

Usage

From source file alluxio.hadoop.mapreduce.KeyValueRecordWriter.java

/**
 * A {@link RecordWriter} to write key-value pairs into a temporary key-value store.
 */
@ThreadSafe
final class KeyValueRecordWriter extends RecordWriter<BytesWritable, BytesWritable> {
    private final KeyValueStoreWriter mWriter;

From source file andromache.hadoop.CassandraRecordWriter.java

/**
 * The <code>ColumnFamilyRecordWriter</code> maps the output &lt;key, value&gt;
 * pairs to a Cassandra column family. In particular, it applies all mutations
 * in the value, which it associates with the key, and in turn the responsible
 * endpoint.
 * <p/>

From source file co.cask.cdap.internal.app.runtime.batch.dataset.DataSetRecordWriter.java

final class DataSetRecordWriter<KEY, VALUE> extends RecordWriter<KEY, VALUE> {
    private static final Logger LOG = LoggerFactory.getLogger(DataSetRecordWriter.class);

    private final BatchWritable<KEY, VALUE> batchWritable;
    private final BasicMapReduceContext mrContext;
    private final MapReduceContextProvider mrContextProvider;

From source file co.cask.cdap.internal.app.runtime.batch.dataset.partitioned.DynamicPartitionerWriterWrapper.java

/**
 * A RecordWriter that allows writing dynamically to multiple partitions of a PartitionedFileSet.
 */
abstract class DynamicPartitionerWriterWrapper<K, V> extends RecordWriter<K, V> {

    private TaskAttemptContext job;

From source file co.cask.cdap.internal.app.runtime.spark.dataset.DatasetRecordWriter.java

final class DatasetRecordWriter<KEY, VALUE> extends RecordWriter<KEY, VALUE> {

    private static final Logger LOG = LoggerFactory.getLogger(DatasetRecordWriter.class);

    //TODO: Needs support for metrics

From source file com.aerospike.hadoop.mapreduce.AerospikeRecordWriter.java

public abstract class AerospikeRecordWriter<KK, VV> extends RecordWriter<KK, VV>
        implements org.apache.hadoop.mapred.RecordWriter<KK, VV> {

    private static final Log log = LogFactory.getLog(AerospikeRecordWriter.class);

    private static final int NO_TASK_ID = -1;

From source file com.aerospike.hadoop.mapreduce.AsyncRecordWriter.java

public abstract class AsyncRecordWriter<KK, VV> extends RecordWriter<KK, VV>
        implements org.apache.hadoop.mapred.RecordWriter<KK, VV> {

    private static final Log log = LogFactory.getLog(AsyncRecordWriter.class);

    private static final int NO_TASK_ID = -1;

From source file com.aliyun.openservices.tablestore.hadoop.TableStoreRecordWriter.java

public class TableStoreRecordWriter extends RecordWriter<Writable, BatchWriteWritable> {
    private static final Logger logger = LoggerFactory.getLogger(TableStoreRecordWriter.class);

    private SyncClientInterface ots;
    private String outputTable;
    private long rowCounter = 0;

From source file com.asakusafw.runtime.mapreduce.simple.ShuffleWriter.java

/**
 * A {@link RecordWriter} for shuffle input.
 * @param <K> the key type
 * @param <V> the value type
 * @since 0.7.1
 */

From source file com.basho.riak.hadoop.RiakRecordWriter.java

/**
 * Writes reducer results to Riak
 * 
 * @author russell
 * @param <V>
 *