List of usage examples for org.apache.hadoop.mapreduce RecordReader subclass-usage
From source file org.apache.jena.grande.mapreduce.io.TripleRecordReader.java
public class TripleRecordReader extends RecordReader<LongWritable, TripleWritable> { private static final Log LOG = LogFactory.getLog(TripleRecordReader.class); public static final String MAX_LINE_LENGTH = "mapreduce.input.linerecordreader.line.maxlength"; private LongWritable key = null;
From source file org.apache.jena.hadoop.rdf.io.input.readers.AbstractBlockBasedNodeTupleReader.java
/**
* An abstract implementation for a record reader that reads records from blocks
* of files, this is a hybrid between {@link AbstractLineBasedNodeTupleReader}
* and {@link AbstractWholeFileNodeTupleReader} in that it can only be used by
* formats which can be split by lines but reduces the overhead by parsing the
* split as a whole rather than as individual lines.
From source file org.apache.jena.hadoop.rdf.io.input.readers.AbstractLineBasedNodeTupleReader.java
/**
* An abstract implementation of a record reader that reads records from line
* based tuple formats. This only supports reading from file splits currently.
* <p>
* The keys produced are the position of the line in the file and the values
* will be node tuples
From source file org.apache.jena.hadoop.rdf.io.input.readers.AbstractRdfReader.java
/**
* An abstract record reader for arbitrary RDF which provides support for
* selecting the actual record reader to use based on detecting the RDF language
* from the file name
*
* @param <TValue>
From source file org.apache.jena.hadoop.rdf.io.input.readers.AbstractWholeFileNodeTupleReader.java
/**
* An abstract implementation for a record reader that reads records from whole
* files i.e. the whole file must be kept together to allow tuples to be
* successfully read. This only supports reading from file splits currently.
* <p>
* The keys produced are the approximate position in the file at which a tuple
From source file org.apache.jena.hadoop.rdf.io.input.readers.TriplesToQuadsReader.java
/**
* A record reader that converts triples into quads by wrapping a
* {@code RecordReader<LongWritable, TripleWritable>} implementation
*
*
*
From source file org.apache.jena.tdbloader4.io.QuadRecordReader.java
public class QuadRecordReader extends RecordReader<LongWritable, QuadWritable> { private static final Log LOG = LogFactory.getLog(QuadRecordReader.class); public static final String MAX_LINE_LENGTH = "mapreduce.input.linerecordreader.line.maxlength"; private LongWritable key = null;
From source file org.apache.kylin.source.kafka.hadoop.KafkaInputRecordReader.java
/** * Convert Kafka topic to Hadoop InputFormat * Modified from the kafka-hadoop-loader in https://github.com/amient/kafka-hadoop-loader */ public class KafkaInputRecordReader extends RecordReader<LongWritable, BytesWritable> {
From source file org.apache.mahout.hadoop.mapreduce.lib.AvroRecordReader.java
/** * Reads Avro data from an Avro file. * Records are returned in the key portion. * The value represents the byte offset of the current object in the file. */ public class AvroRecordReader<K> extends RecordReader<K, LongWritable> {
From source file org.apache.mahout.text.LuceneSegmentRecordReader.java
/** * {@link RecordReader} implementation for Lucene segments. Each {@link InputSplit} contains a separate Lucene segment. * Emits records consisting of a {@link Text} document ID and a null key. */ public class LuceneSegmentRecordReader extends RecordReader<Text, NullWritable> {