Example usage for org.apache.hadoop.mapreduce RecordReader subclass-usage

List of usage examples for org.apache.hadoop.mapreduce RecordReader subclass-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapreduce RecordReader subclass-usage.

Usage

From source file com.cloudera.crunch.impl.mr.run.CrunchRecordReader.java

class CrunchRecordReader<K, V> extends RecordReader<K, V> {

    private final RecordReader<K, V> delegate;

    public CrunchRecordReader(InputSplit inputSplit, TaskAttemptContext context)
            throws IOException, InterruptedException {

From source file com.cloudera.crunch.type.avro.AvroRecordReader.java

/** An {@link RecordReader} for Avro data files. */
public class AvroRecordReader<T> extends RecordReader<AvroWrapper<T>, NullWritable> {

    private FileReader<T> reader;
    private long start;
    private long end;

From source file com.cloudera.fts.spark.format.RawFileRecordReader.java

public class RawFileRecordReader extends RecordReader<Text, BytesWritable> {
    private static final int BUFFER_SIZE = 8192;
    private Text key;
    private BytesWritable value;
    private FSDataInputStream fileIn = null;
    private boolean finished;

From source file com.cloudera.integration.oracle.goldengate.ldv.mapreduce.lib.input.LengthDelimitedRecordReader.java

/**
 *
 * @author jcustenborder
 */
public class LengthDelimitedRecordReader extends RecordReader<LongWritable, LengthDelimitedWritable> {

From source file com.cloudera.recordservice.hcatalog.mapreduce.HCatRecordReader.java

/**
 * The HCat wrapper for the underlying RecordReader this ensures that the initialize on
 * the underlying record reader is done with the underlying split not with HCatSplit.
 */
class HCatRecordReader extends RecordReader<WritableComparable, RecordServiceRecord> {
    private static final Logger LOG = LoggerFactory.getLogger(HCatRecordReader.class);

From source file com.cloudera.sa.ExcelRecordReader.java

/**
 * @author Sameer Abhyankar
 */
public class ExcelRecordReader extends RecordReader<NullWritable, TextArrayWritable> {

    private TextArrayWritable value = new TextArrayWritable();

From source file com.conductor.hadoop.DelegatingRecordReader.java

/**
 * This class is taken directly from the Hadoop source code so that we can leverage {@link TaggedInputSplit} which is
 * package-private in the Hadoop library.
 * 
 * @see org.apache.hadoop.mapreduce.lib.input.DelegatingRecordReader
 * @see com.conductor.kafka.hadoop.MultipleKafkaInputFormat

From source file com.conductor.kafka.hadoop.KafkaRecordReader.java

/**
 * A record reader that reads a subsection, [{@link #getStart()}, {@link #getEnd()}), of a Kafka queue
 * {@link com.conductor.kafka.Partition}.
 * 
 * <p/>
 * Thanks to <a href="https://github.com/miniway">Dongmin Yu</a> for providing the inspiration for this code.

From source file com.cotdp.hadoop.BrotliFileRecordReader.java

public class BrotliFileRecordReader extends RecordReader<Text, Text> {
    /** InputStream used to read the ZIP file from the FileSystem */
    private FSDataInputStream fsin;

    /** Brotli file parser/decompresser */
    private Process decompressor;

From source file com.cotdp.hadoop.ZipFileRecordReader.java

/**
 * This RecordReader implementation extracts individual files from a ZIP
 * file and hands them over to the Mapper. The "key" is the decompressed
 * file name, the "value" is the file contents.
 */
public class ZipFileRecordReader extends RecordReader<Text, BytesWritable> {