List of usage examples for org.apache.hadoop.mapred RecordWriter interface-usage
From source file com.aerospike.hadoop.mapreduce.AerospikeRecordWriter.java
public abstract class AerospikeRecordWriter<KK, VV> extends RecordWriter<KK, VV> implements org.apache.hadoop.mapred.RecordWriter<KK, VV> { private static final Log log = LogFactory.getLog(AerospikeRecordWriter.class); private static final int NO_TASK_ID = -1;
From source file com.aerospike.hadoop.mapreduce.AsyncRecordWriter.java
public abstract class AsyncRecordWriter<KK, VV> extends RecordWriter<KK, VV> implements org.apache.hadoop.mapred.RecordWriter<KK, VV> { private static final Log log = LogFactory.getLog(AsyncRecordWriter.class); private static final int NO_TASK_ID = -1;
From source file com.bah.culvert.hive.CulvertRecordWriter.java
/** * Hive record writer that outputs data to a culvert table. */ public class CulvertRecordWriter implements RecordWriter, org.apache.hadoop.mapred.RecordWriter<BytesWritable, Put> {
From source file com.blackberry.logdriver.mapred.BinaryRecordWriter.java
public class BinaryRecordWriter implements RecordWriter<BytesWritable, NullWritable> { private static final Logger LOG = LoggerFactory.getLogger(BinaryRecordWriter.class); private FSDataOutputStream out; public BinaryRecordWriter(JobConf job) {
From source file com.blackberry.logdriver.mapred.boom.ReBoomRecordWriter.java
public class ReBoomRecordWriter implements RecordWriter<LogLineData, Text> { private ReBoomWriter writer; public ReBoomRecordWriter(ReBoomOutputFormat reBoomOutputFormat, JobConf job) throws IOException { String taskid = job.get("mapred.task.id");
From source file com.buzzinate.dm.cassandra.ColumnFamilyRecordWriter.java
/**
* The <code>ColumnFamilyRecordWriter</code> maps the output <key, value>
* pairs to a Cassandra column family. In particular, it applies all mutations
* in the value, which it associates with the key, and in turn the responsible
* endpoint.
*
From source file com.cloudera.science.avro.streaming.AvroAsJSONRecordWriter.java
public class AvroAsJSONRecordWriter implements RecordWriter<Text, Text> { private final DataFileWriter<GenericRecord> writer; private final JsonConverter converter; private final boolean readKey;
From source file com.datascience.hadoop.CsvRecordWriter.java
/**
* CSV record writer.
* <p>
* The CSV record writer handles writing {@link com.datascience.hadoop.ListWritable} instances out to CSV files on behalf
* of {@link com.datascience.hadoop.CsvOutputFormat}. Internally, it uses a {@link org.apache.commons.csv.CSVPrinter}
* configured by its parent {@link org.apache.hadoop.mapred.OutputFormat} to write CSV records out to the provided
From source file com.google.mr4c.hadoop.MR4CRecordWriter.java
public class MR4CRecordWriter implements RecordWriter<Text, Text> { protected Logger m_log = MR4CLogging.getLogger(MR4CRecordWriter.class); private AlgoRunner m_algoRunner; private DatasetSerializer m_serializer;
From source file com.hotels.corc.mapred.CorcRecordWriter.java
/** * A wrapper for {@link OrcRecordWriter} exposing {@link Corc} in place of {@link OrcStruct}. */ class CorcRecordWriter implements RecordWriter<NullWritable, Corc> { @SuppressWarnings("rawtypes") private final RecordWriter writer; // OrcSerdeRow is package private so have to use rawtypes