Example usage for org.apache.hadoop.mapreduce RecordReader subclass-usage

List of usage examples for org.apache.hadoop.mapreduce RecordReader subclass-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapreduce RecordReader subclass-usage.

Usage

From source file org.apache.jena.grande.mapreduce.io.TripleRecordReader.java

public class TripleRecordReader extends RecordReader<LongWritable, TripleWritable> {

    private static final Log LOG = LogFactory.getLog(TripleRecordReader.class);
    public static final String MAX_LINE_LENGTH = "mapreduce.input.linerecordreader.line.maxlength";

    private LongWritable key = null;

From source file org.apache.jena.hadoop.rdf.io.input.readers.AbstractBlockBasedNodeTupleReader.java

/**
 * An abstract implementation for a record reader that reads records from blocks
 * of files, this is a hybrid between {@link AbstractLineBasedNodeTupleReader}
 * and {@link AbstractWholeFileNodeTupleReader} in that it can only be used by
 * formats which can be split by lines but reduces the overhead by parsing the
 * split as a whole rather than as individual lines.

From source file org.apache.jena.hadoop.rdf.io.input.readers.AbstractLineBasedNodeTupleReader.java

/**
 * An abstract implementation of a record reader that reads records from line
 * based tuple formats. This only supports reading from file splits currently.
 * <p>
 * The keys produced are the position of the line in the file and the values
 * will be node tuples

From source file org.apache.jena.hadoop.rdf.io.input.readers.AbstractRdfReader.java

/**
 * An abstract record reader for arbitrary RDF which provides support for
 * selecting the actual record reader to use based on detecting the RDF language
 * from the file name
 * 
 * @param <TValue>

From source file org.apache.jena.hadoop.rdf.io.input.readers.AbstractWholeFileNodeTupleReader.java

/**
 * An abstract implementation for a record reader that reads records from whole
 * files i.e. the whole file must be kept together to allow tuples to be
 * successfully read. This only supports reading from file splits currently.
 * <p>
 * The keys produced are the approximate position in the file at which a tuple

From source file org.apache.jena.hadoop.rdf.io.input.readers.TriplesToQuadsReader.java

/**
 * A record reader that converts triples into quads by wrapping a
 * {@code RecordReader<LongWritable, TripleWritable>} implementation
 * 
 * 
 * 

From source file org.apache.jena.tdbloader4.io.QuadRecordReader.java

public class QuadRecordReader extends RecordReader<LongWritable, QuadWritable> {

    private static final Log LOG = LogFactory.getLog(QuadRecordReader.class);
    public static final String MAX_LINE_LENGTH = "mapreduce.input.linerecordreader.line.maxlength";

    private LongWritable key = null;

From source file org.apache.kylin.source.kafka.hadoop.KafkaInputRecordReader.java

/**
 * Convert Kafka topic to Hadoop InputFormat
 * Modified from the kafka-hadoop-loader in https://github.com/amient/kafka-hadoop-loader
 */
public class KafkaInputRecordReader extends RecordReader<LongWritable, BytesWritable> {

From source file org.apache.mahout.hadoop.mapreduce.lib.AvroRecordReader.java

/**
 * Reads Avro data from an Avro file.
 * Records are returned in the key portion.
 * The value represents the byte offset of the current object in the file.
 */
public class AvroRecordReader<K> extends RecordReader<K, LongWritable> {

From source file org.apache.mahout.text.LuceneSegmentRecordReader.java

/**
 * {@link RecordReader} implementation for Lucene segments. Each {@link InputSplit} contains a separate Lucene segment.
 * Emits records consisting of a {@link Text} document ID and a null key.
 */
public class LuceneSegmentRecordReader extends RecordReader<Text, NullWritable> {