List of usage examples for org.apache.hadoop.mapreduce RecordReader subclass-usage
From source file com.ambiata.ivory.operation.hadoop.DelegatingRecordReader.java
/**
* This is a delegating RecordReader, which delegates the functionality to the
* underlying record reader in {@link TaggedInputSplit}
*/
@InterfaceAudience.Private
@InterfaceStability.Unstable
From source file com.asakusafw.bridge.hadoop.ModelInputRecordReader.java
/** * An implementation of {@link RecordReader} for reading objects from {@link ModelInput}. * @param <T> the target data model type */ public final class ModelInputRecordReader<T> extends RecordReader<NullWritable, Object> {
From source file com.asakusafw.runtime.stage.input.StageInputRecordReader.java
/** * An implementation of Hadoop {@link RecordReader} for handling multiple Map operations. * @since 0.1.0 * @version 0.2.6 */ @SuppressWarnings("rawtypes")
From source file com.ashishpaliwal.hadoop.utils.inputformat.CsvRecordReader.java
/** * CSV record reader. This input format is valid only for scenarios where CSV input is enclosed * within double quotes. This reader takes to skip new line and comma char in input if they are enclosed * within double quotes */ public class CsvRecordReader extends RecordReader<LongWritable, Text> {
From source file com.awcoleman.BouncyCastleGenericCDRHadoop.RawFileRecordReader.java
/**
*
* Input Format for "Simple Generic CDR"
* Reads in entire ASN.1 file as key.
*
* @author awcoleman
From source file com.awcoleman.BouncyCastleGenericCDRHadoopWithWritable.RawFileRecordReader.java
/**
*
* Input Format for "Simple Generic CDR" with the base data container as Writable
* Reads in entire ASN.1 file as key.
*
* Outputs Key of startDate and Value of Writable CallDetailRecord
From source file com.awcoleman.ExampleJobSummaryLogWithOutput.BinRecRecordReader.java
/** * @author awcoleman * @version 20150624 * license: Apache License 2.0; http://www.apache.org/licenses/LICENSE-2.0 */ public class BinRecRecordReader extends RecordReader<LongWritable, AvroValue<BinRecForPartitions>> {
From source file com.basho.riak.hadoop.RiakRecordReader.java
/**
* Wrapper around a {@link RiakClient} for reading values from Riak
*
* @author russell
*
*/
From source file com.baynote.hadoop.DelegatingRecordReader.java
/**
* This class is taken directly from the Hadoop source code so that we can leverage {@link TaggedInputSplit} which is
* package-private in the Hadoop library.
*
* @see org.apache.hadoop.mapreduce.lib.input.DelegatingRecordReader
* @see com.baynote.kafka.hadoop.MultipleKafkaInputFormat
From source file com.baynote.kafka.hadoop.KafkaRecordReader.java
/**
* A record reader that reads a subsection, [{@link #getStart()}, {@link #getEnd()}), of a Kafka queue
* {@link com.baynote.kafka.Partition}.
*
* <p>
* Thanks to <a href="https://github.com/miniway">Dongmin Yu</a> for providing the inspiration for this code.