List of usage examples for org.apache.hadoop.mapreduce RecordWriter subclass-usage
From source file alluxio.hadoop.mapreduce.KeyValueRecordWriter.java
/** * A {@link RecordWriter} to write key-value pairs into a temporary key-value store. */ @ThreadSafe final class KeyValueRecordWriter extends RecordWriter<BytesWritable, BytesWritable> { private final KeyValueStoreWriter mWriter;
From source file andromache.hadoop.CassandraRecordWriter.java
/**
* The <code>ColumnFamilyRecordWriter</code> maps the output <key, value>
* pairs to a Cassandra column family. In particular, it applies all mutations
* in the value, which it associates with the key, and in turn the responsible
* endpoint.
* <p/>
From source file co.cask.cdap.internal.app.runtime.batch.dataset.DataSetRecordWriter.java
final class DataSetRecordWriter<KEY, VALUE> extends RecordWriter<KEY, VALUE> { private static final Logger LOG = LoggerFactory.getLogger(DataSetRecordWriter.class); private final BatchWritable<KEY, VALUE> batchWritable; private final BasicMapReduceContext mrContext; private final MapReduceContextProvider mrContextProvider;
From source file co.cask.cdap.internal.app.runtime.batch.dataset.partitioned.DynamicPartitionerWriterWrapper.java
/** * A RecordWriter that allows writing dynamically to multiple partitions of a PartitionedFileSet. */ abstract class DynamicPartitionerWriterWrapper<K, V> extends RecordWriter<K, V> { private TaskAttemptContext job;
From source file co.cask.cdap.internal.app.runtime.spark.dataset.DatasetRecordWriter.java
final class DatasetRecordWriter<KEY, VALUE> extends RecordWriter<KEY, VALUE> { private static final Logger LOG = LoggerFactory.getLogger(DatasetRecordWriter.class); //TODO: Needs support for metrics
From source file com.aerospike.hadoop.mapreduce.AerospikeRecordWriter.java
public abstract class AerospikeRecordWriter<KK, VV> extends RecordWriter<KK, VV> implements org.apache.hadoop.mapred.RecordWriter<KK, VV> { private static final Log log = LogFactory.getLog(AerospikeRecordWriter.class); private static final int NO_TASK_ID = -1;
From source file com.aerospike.hadoop.mapreduce.AsyncRecordWriter.java
public abstract class AsyncRecordWriter<KK, VV> extends RecordWriter<KK, VV> implements org.apache.hadoop.mapred.RecordWriter<KK, VV> { private static final Log log = LogFactory.getLog(AsyncRecordWriter.class); private static final int NO_TASK_ID = -1;
From source file com.aliyun.openservices.tablestore.hadoop.TableStoreRecordWriter.java
public class TableStoreRecordWriter extends RecordWriter<Writable, BatchWriteWritable> { private static final Logger logger = LoggerFactory.getLogger(TableStoreRecordWriter.class); private SyncClientInterface ots; private String outputTable; private long rowCounter = 0;
From source file com.asakusafw.runtime.mapreduce.simple.ShuffleWriter.java
/**
* A {@link RecordWriter} for shuffle input.
* @param <K> the key type
* @param <V> the value type
* @since 0.7.1
*/
From source file com.basho.riak.hadoop.RiakRecordWriter.java
/**
* Writes reducer results to Riak
*
* @author russell
* @param <V>
*