Example usage for org.apache.hadoop.mapreduce RecordReader subclass-usage

List of usage examples for org.apache.hadoop.mapreduce RecordReader subclass-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapreduce RecordReader subclass-usage.

Usage

From source file org.zuinnote.hadoop.office.format.mapreduce.AbstractSpreadSheetDocumentRecordReader.java

/**
* This abstract reader implements some generic functionality (e.g. buffers etc.) for reading tables from various formats
*
*
*/

From source file parquet.hadoop.ParquetRecordReader.java

/**
 * Reads the records from a block of a Parquet file
 *
 * @see ParquetInputFormat
 *
 * @author Julien Le Dem

From source file reuter.searcher.ZipFileRecordReader.java

/**
 * This RecordReader implementation extracts individual files from a ZIP
 * file and hands them over to the Mapper. The "key" is the decompressed
 * file name, the "value" is the file contents.
 */
public class ZipFileRecordReader extends RecordReader<Text, BytesWritable> {

From source file reuter.util.ZipFileRecordReader.java

/**
 * This RecordReader implementation extracts individual files from a ZIP
 * file and hands them over to the Mapper. The "key" is the decompressed
 * file name, the "value" is the file contents.
 */
public class ZipFileRecordReader extends RecordReader<Text, BytesWritable> {

From source file simsql.runtime.CombineInputFileDeserializer.java

public class CombineInputFileDeserializer extends RecordReader<Nothing, Record> {

    // the delegate
    private InputFileDeserializer reader;

    // our input

From source file simsql.runtime.InputFileDeserializer.java

public class InputFileDeserializer extends RecordReader<Nothing, Record> {

    // records where we are currently in the file that we are reading from, as well as the
    // location where we are supposed to start reading, as well as the last byte assigned to us
    private long start;
    private long pos;

From source file trec.MyLineRecordReader.java

/**
 * Treats keys as offset in file and value as line. 
 */
public class MyLineRecordReader extends RecordReader<LongWritable, Text> {
    private static final Log LOG = LogFactory.getLog(LineRecordReader.class);

From source file trec.PRRecordReader.java

/**
 * Treats keys as offset in file and value as line. 
 */
public class PRRecordReader extends RecordReader<LongWritable, Text> {
    private static final Log LOG = LogFactory.getLog(PRRecordReader.class);

From source file trec.TrecRecordReader.java

/**
 * Treats keys as offset in file and value as line.
 */
public class TrecRecordReader extends RecordReader<LongWritable, Text> {

    private CompressionCodecFactory compressionCodecs = null;

From source file uk.bl.wa.hadoop.mapreduce.lib.DereferencingArchiveToCDXRecordReader.java

public class DereferencingArchiveToCDXRecordReader<Key extends WritableComparable<?>, Value extends Writable>
        extends RecordReader<Text, Text> {

    public static final String CDX_09 = " CDX N b a m s k r V g";
    public static final String CDX_10 = " CDX A b a m s k r M V g";
    public static final String CDX_11 = " CDX N b a m s k r M S V g";