List of usage examples for org.apache.hadoop.mapreduce RecordWriter subclass-usage
From source file org.apache.cassandra.hadoop.AbstractColumnFamilyRecordWriter.java
/**
* The <code>ColumnFamilyRecordWriter</code> maps the output <key, value>
* pairs to a Cassandra column family. In particular, it applies all mutations
* in the value, which it associates with the key, and in turn the responsible
* endpoint.
*
From source file org.apache.cassandra.hadoop.BulkRecordWriter.java
@Deprecated public final class BulkRecordWriter extends RecordWriter<ByteBuffer, List<Mutation>> implements org.apache.hadoop.mapred.RecordWriter<ByteBuffer, List<Mutation>> { public final static String OUTPUT_LOCATION = "mapreduce.output.bulkoutputformat.localdir"; public final static String BUFFER_SIZE_IN_MB = "mapreduce.output.bulkoutputformat.buffersize"; public final static String STREAM_THROTTLE_MBITS = "mapreduce.output.bulkoutputformat.streamthrottlembits";
From source file org.apache.cassandra.hadoop.ColumnFamilyRecordWriter.java
/**
* The <code>ColumnFamilyRecordWriter</code> maps the output <key, value>
* pairs to a Cassandra column family. In particular, it applies all mutations
* in the value, which it associates with the key, and in turn the responsible
* endpoint.
*
From source file org.apache.cassandra.hadoop.cql3.CqlBulkRecordWriter.java
/**
* The <code>CqlBulkRecordWriter</code> maps the output <key, value>
* pairs to a Cassandra column family. In particular, it applies the binded variables
* in the value to the prepared statement, which it associates with the key, and in
* turn the responsible endpoint.
*
From source file org.apache.cassandra.hadoop.cql3.CqlRecordWriter.java
/**
* The <code>CqlRecordWriter</code> maps the output <key, value>
* pairs to a Cassandra table. In particular, it applies the binded variables
* in the value to the prepared statement, which it associates with the key, and in
* turn the responsible endpoint.
*
From source file org.apache.cassandra.hadoop2.AbstractColumnFamilyRecordWriter.java
/**
* The <code>ColumnFamilyRecordWriter</code> maps the output <key, value>
* pairs to a Cassandra column family. In particular, it applies all mutations
* in the value, which it associates with the key, and in turn the responsible
* endpoint.
*
From source file org.apache.cassandra.hadoop2.BulkRecordWriter.java
final class BulkRecordWriter extends RecordWriter<ByteBuffer, List<Mutation>> implements org.apache.hadoop.mapred.RecordWriter<ByteBuffer, List<Mutation>> { private final static String OUTPUT_LOCATION = "mapreduce.output.bulkoutputformat.localdir"; private final static String BUFFER_SIZE_IN_MB = "mapreduce.output.bulkoutputformat.buffersize"; private final static String STREAM_THROTTLE_MBITS = "mapreduce.output.bulkoutputformat.streamthrottlembits";
From source file org.apache.cassandra.io.sstable.CrunchBulkRecordWriter.java
/**
* This is an almost-copy of {@link org.apache.cassandra.hadoop.BulkRecordWriter}.
* <p>
* We had to re-implement this class because of https://issues.apache.org/jira/browse/CASSANDRA-8367
* </p>
*/
From source file org.apache.crunch.io.avro.trevni.TrevniRecordWriter.java
/** * */ public class TrevniRecordWriter<T> extends RecordWriter<AvroKey<T>, NullWritable> { /** trevni file extension */
From source file org.apache.giraph.bsp.BspRecordWriter.java
/** * Used by {@link BspOutputFormat} since some versions of Hadoop * require that a RecordWriter is returned from getRecordWriter. * Does nothing, except insures that write is never called. */ public class BspRecordWriter extends RecordWriter<Text, Text> {