Example usage for org.apache.hadoop.mapred RecordWriter interface-usage

List of usage examples for org.apache.hadoop.mapred RecordWriter interface-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapred RecordWriter interface-usage.

Usage

From source file com.aerospike.hadoop.mapreduce.AerospikeRecordWriter.java

public abstract class AerospikeRecordWriter<KK, VV> extends RecordWriter<KK, VV>
        implements org.apache.hadoop.mapred.RecordWriter<KK, VV> {

    private static final Log log = LogFactory.getLog(AerospikeRecordWriter.class);

    private static final int NO_TASK_ID = -1;

From source file com.aerospike.hadoop.mapreduce.AsyncRecordWriter.java

public abstract class AsyncRecordWriter<KK, VV> extends RecordWriter<KK, VV>
        implements org.apache.hadoop.mapred.RecordWriter<KK, VV> {

    private static final Log log = LogFactory.getLog(AsyncRecordWriter.class);

    private static final int NO_TASK_ID = -1;

From source file com.bah.culvert.hive.CulvertRecordWriter.java

/**
 * Hive record writer that outputs data to a culvert table.
 */
public class CulvertRecordWriter
        implements RecordWriter, org.apache.hadoop.mapred.RecordWriter<BytesWritable, Put> {

From source file com.blackberry.logdriver.mapred.BinaryRecordWriter.java

public class BinaryRecordWriter implements RecordWriter<BytesWritable, NullWritable> {
    private static final Logger LOG = LoggerFactory.getLogger(BinaryRecordWriter.class);

    private FSDataOutputStream out;

    public BinaryRecordWriter(JobConf job) {

From source file com.blackberry.logdriver.mapred.boom.ReBoomRecordWriter.java

public class ReBoomRecordWriter implements RecordWriter<LogLineData, Text> {

    private ReBoomWriter writer;

    public ReBoomRecordWriter(ReBoomOutputFormat reBoomOutputFormat, JobConf job) throws IOException {
        String taskid = job.get("mapred.task.id");

From source file com.buzzinate.dm.cassandra.ColumnFamilyRecordWriter.java

/**
 * The <code>ColumnFamilyRecordWriter</code> maps the output &lt;key, value&gt;
 * pairs to a Cassandra column family. In particular, it applies all mutations
 * in the value, which it associates with the key, and in turn the responsible
 * endpoint.
 *

From source file com.cloudera.science.avro.streaming.AvroAsJSONRecordWriter.java

public class AvroAsJSONRecordWriter implements RecordWriter<Text, Text> {

    private final DataFileWriter<GenericRecord> writer;
    private final JsonConverter converter;
    private final boolean readKey;

From source file com.datascience.hadoop.CsvRecordWriter.java

/**
 * CSV record writer.
 * <p>
 * The CSV record writer handles writing {@link com.datascience.hadoop.ListWritable} instances out to CSV files on behalf
 * of {@link com.datascience.hadoop.CsvOutputFormat}. Internally, it uses a {@link org.apache.commons.csv.CSVPrinter}
 * configured by its parent {@link org.apache.hadoop.mapred.OutputFormat} to write CSV records out to the provided

From source file com.google.mr4c.hadoop.MR4CRecordWriter.java

public class MR4CRecordWriter implements RecordWriter<Text, Text> {

    protected Logger m_log = MR4CLogging.getLogger(MR4CRecordWriter.class);

    private AlgoRunner m_algoRunner;
    private DatasetSerializer m_serializer;

From source file com.hotels.corc.mapred.CorcRecordWriter.java

/**
 * A wrapper for {@link OrcRecordWriter} exposing {@link Corc} in place of {@link OrcStruct}.
 */
class CorcRecordWriter implements RecordWriter<NullWritable, Corc> {
    @SuppressWarnings("rawtypes")
    private final RecordWriter writer; // OrcSerdeRow is package private so have to use rawtypes