Example usage for org.apache.hadoop.mapreduce RecordReader subclass-usage

List of usage examples for org.apache.hadoop.mapreduce RecordReader subclass-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapreduce RecordReader subclass-usage.

Usage

From source file com.linkedin.cubert.io.virtual.VirtualRecordReader.java

/**
 * Record reader for the virtual storage.
 * The purpose of this data structure is to support a predecided number of mappers
 *
 * For the purposes of progress reporting. The nextKeyValue() returns true always! This implies that the map input
 * records counter will essentially report a garbage value. However, the number of bytes read is 0.

From source file com.linkedin.cubert.pig.piggybank.storage.avro.PigAvroRecordReader.java

/**
 * This is an implementation of record reader which reads in avro data and
 * convert them into <NullWritable, Writable> pairs.
 */
public class PigAvroRecordReader extends RecordReader<NullWritable, Writable> {

From source file com.mapred.EmplRecordReader.java

public class EmplRecordReader extends RecordReader<LongWritable, Text> {

    LineRecordReader in;
    LongWritable key;
    Text txt;
    Long position;

From source file com.marcolotz.lung.io.inputFormat.MultipleFilesRecordReader.java

/**
 * The reader that {@link MultipleFilesInputFormat} uses.
 * @author Marco Aurelio Lotz
 *
 */
public class MultipleFilesRecordReader extends RecordReader<NullWritable, BytesWritable> {

From source file com.marcolotz.lung.io.inputFormat.WholeFileRecordReader.java

/**
 * Writes the whole file as a value for a Null key. This implementation is
 * described in the Tom White Hadoop book.
 * 
 * This WholeFileRecordReader may the compatible with the
 * {@link MultipleFilesInputFormat}, but no checks were performed so far.

From source file com.marklogic.contentpump.ImportRecordReader.java

/**
 * Abstract class of RecorderReader for import.
 * @author ali
 *
 * @param <VALUEIN>
 */

From source file com.marklogic.mapreduce.examples.ContentLoader.java

class FileReader extends RecordReader<Text, Text> {

    private Text key = new Text();
    private Text value = new Text();
    private long bytesRead;
    private long bytesTotal;

From source file com.marklogic.mapreduce.examples.LinkCountHDFS.java

class LinkRecordReader extends RecordReader<IntWritable, Text> {
    private static final String PATH_EXPRESSION = "//wp:a[@title and @href and not (starts-with(@href, '#') or "
            + "starts-with(@href, 'http://') or starts-with(@href, 'File:') "
            + " or starts-with(@href, 'Image:'))]/@title";
    private IntWritable key;
    private Text value;

From source file com.marklogic.mapreduce.examples.WikiLoader.java

class WikiReader extends RecordReader<Text, Text> {

    static final int BUFFER_SIZE = 65536;
    static final int READ_AHEAD_SIZE = 2048;
    static final String BEGIN_PAGE_TAG = "<page>";
    static final String END_PAGE_TAG = "</page>";

From source file com.marklogic.mapreduce.examples.ZipContentLoader.java

class ZipContentReader extends RecordReader<Text, Text> {

    private Text key = new Text();
    private Text value = new Text();
    private ZipInputStream zipIn;
    private byte[] buf = new byte[65536];