Example usage for org.apache.hadoop.mapreduce RecordReader subclass-usage

List of usage examples for org.apache.hadoop.mapreduce RecordReader subclass-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapreduce RecordReader subclass-usage.

Usage

From source file fi.tkk.ics.hadoop.bam.VCFRecordReader.java

/** The key is the bitwise OR of the chromosome index in the upper 32 bits
 * and the 0-based leftmost coordinate in the lower.
 *
 * The chromosome index is based on the ordering of the contig lines in the VCF
 * header. If a chromosome name that cannot be found in the contig lines is
 * used, that name is instead hashed to form the upper part of the key.

From source file fire.util.fileformats.combineimagefileinputformat.CombineFileImageRecordReader.java

/**
 * RecordReader is responsible from extracting records from a chunk
 * of the CombineFileSplit.
 */
public class CombineFileImageRecordReader extends RecordReader<Text, BytesWritable> {

From source file fire.util.fileformats.combinetextfileinputformat.CombineFileLineRecordReader.java

/**
 * RecordReader is responsible from extracting records from a chunk
 * of the CombineFileSplit.
 */
public class CombineFileLineRecordReader extends RecordReader<WordOffset, Text> {

From source file fire.util.fileformats.pdf.PdfRecordReader.java

public class PdfRecordReader extends RecordReader<Text, Text> {

    private FileSplit fileSplit;
    private Configuration conf;
    private Text key = null;
    private Text value = null;

From source file fire.util.fileformats.tika.TikaRecordReader.java

public class TikaRecordReader extends RecordReader<Text, Text> {

    private CombineFileSplit split;
    private FileSystem fs;
    private Text key, value;
    private Path[] paths;

From source file format.OverlapLengthRecordReader.java

/**
 * A reader to read fixed length records from a split.  Record offset is
 * returned as key and the record as bytes is returned in value.
 */
@InterfaceAudience.Private
@InterfaceStability.Evolving

From source file format.OverlapRecordReader.java

/**
 * Reads line from a 4mc compressed text file.
 * Treats keys as offset in file and value as line.
 */
public class OverlapRecordReader extends RecordReader<LongWritable, Text> {

From source file fr.ens.biologie.genomique.eoulsan.bio.io.hadoop.FastqLineRecordReader.java

/**
 * Treats keys as offset in file and value as line.
 */
public class FastqLineRecordReader extends RecordReader<LongWritable, Text> {

    public static final String MAX_LINE_LENGTH = "mapreduce.input.linerecordreader.line.maxlength";

From source file fr.ens.biologie.genomique.eoulsan.bio.io.hadoop.FastqRecordReader.java

/**
 * This class define a RecordReader for FASTQ files for the Hadoop MapReduce
 * framework.
 * @since 1.0
 * @author Laurent Jourdren
 */

From source file gov.jgi.meta.hadoop.input.FastaBlockRecordReader.java

/**
 * Treats keys as offset in file and value as line.
 */
public class FastaBlockRecordReader extends RecordReader<Text, Map<String, String>> {
    private static final Log LOG = LogFactory.getLog(FastaBlockRecordReader.class);