List of usage examples for org.apache.hadoop.mapreduce RecordReader subclass-usage
From source file com.cloudera.crunch.impl.mr.run.CrunchRecordReader.java
class CrunchRecordReader<K, V> extends RecordReader<K, V> { private final RecordReader<K, V> delegate; public CrunchRecordReader(InputSplit inputSplit, TaskAttemptContext context) throws IOException, InterruptedException {
From source file com.cloudera.crunch.type.avro.AvroRecordReader.java
/** An {@link RecordReader} for Avro data files. */ public class AvroRecordReader<T> extends RecordReader<AvroWrapper<T>, NullWritable> { private FileReader<T> reader; private long start; private long end;
From source file com.cloudera.fts.spark.format.RawFileRecordReader.java
public class RawFileRecordReader extends RecordReader<Text, BytesWritable> { private static final int BUFFER_SIZE = 8192; private Text key; private BytesWritable value; private FSDataInputStream fileIn = null; private boolean finished;
From source file com.cloudera.integration.oracle.goldengate.ldv.mapreduce.lib.input.LengthDelimitedRecordReader.java
/** * * @author jcustenborder */ public class LengthDelimitedRecordReader extends RecordReader<LongWritable, LengthDelimitedWritable> {
From source file com.cloudera.recordservice.hcatalog.mapreduce.HCatRecordReader.java
/** * The HCat wrapper for the underlying RecordReader this ensures that the initialize on * the underlying record reader is done with the underlying split not with HCatSplit. */ class HCatRecordReader extends RecordReader<WritableComparable, RecordServiceRecord> { private static final Logger LOG = LoggerFactory.getLogger(HCatRecordReader.class);
From source file com.cloudera.sa.ExcelRecordReader.java
/** * @author Sameer Abhyankar */ public class ExcelRecordReader extends RecordReader<NullWritable, TextArrayWritable> { private TextArrayWritable value = new TextArrayWritable();
From source file com.conductor.hadoop.DelegatingRecordReader.java
/**
* This class is taken directly from the Hadoop source code so that we can leverage {@link TaggedInputSplit} which is
* package-private in the Hadoop library.
*
* @see org.apache.hadoop.mapreduce.lib.input.DelegatingRecordReader
* @see com.conductor.kafka.hadoop.MultipleKafkaInputFormat
From source file com.conductor.kafka.hadoop.KafkaRecordReader.java
/**
* A record reader that reads a subsection, [{@link #getStart()}, {@link #getEnd()}), of a Kafka queue
* {@link com.conductor.kafka.Partition}.
*
* <p/>
* Thanks to <a href="https://github.com/miniway">Dongmin Yu</a> for providing the inspiration for this code.
From source file com.cotdp.hadoop.BrotliFileRecordReader.java
public class BrotliFileRecordReader extends RecordReader<Text, Text> { /** InputStream used to read the ZIP file from the FileSystem */ private FSDataInputStream fsin; /** Brotli file parser/decompresser */ private Process decompressor;
From source file com.cotdp.hadoop.ZipFileRecordReader.java
/** * This RecordReader implementation extracts individual files from a ZIP * file and hands them over to the Mapper. The "key" is the decompressed * file name, the "value" is the file contents. */ public class ZipFileRecordReader extends RecordReader<Text, BytesWritable> {