List of usage examples for org.apache.hadoop.mapreduce RecordReader subclass-usage
From source file com.yahoo.druid.hadoop.DruidRecordReader.java
public class DruidRecordReader extends RecordReader<DateTime, Map<String, Object>> { private static final Logger logger = LoggerFactory.getLogger(DruidRecordReader.class); private IngestSegmentFirehose rowYielder;
From source file com.zjy.mongo.input.BSONFileRecordReader.java
/**
* <p>
* Copyright (c) 2008 - 2013 10gen, Inc. (http://10gen.com)
* </p>
* <p>
* Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may
From source file com.zjy.mongo.input.MongoRecordReader.java
public class MongoRecordReader extends RecordReader<Object, BSONObject> { public MongoRecordReader(final MongoInputSplit split) { this.split = split; cursor = split.getCursor(); }
From source file cosmos.mapred.LongLineRecordReader.java
/** * A copy of {@link LineRecordReader} which does not discard lines longer than "mapred.linerecordreader.maxlength". Instead, it returns them, leaving it to the * mapper to decide what to do with it. It also does not treat '\r' (CR) characters as new lines -- it uses {@link LfLineReader} instead of {@link LineReader} * to read lines. */ public class LongLineRecordReader extends RecordReader<LongWritable, Text> {
From source file crunch.MaxTemperature.java
class WholeFileRecordReader extends RecordReader<NullWritable, BytesWritable> { private FileSplit fileSplit; private Configuration conf; private BytesWritable value = new BytesWritable(); private boolean processed = false;
From source file cs480a2.yqiu.recSystem.mapreduce.input.SingleBookReader.java
/**
* Created by Qiu on 3/18/15.
* This record reader takes a single book as input.
* Output key: Text ---> " Title / Maximum Word Count "
* Output value: TextArray ---> [ "Word A / Word A Count", ... ]
*/
From source file de.gesundkrank.wikipedia.hadoop.inputformat.WikiInputRecordReader.java
/** * Jan Graegger<jan.grassegger@uni-weimar.de> */ public class WikiInputRecordReader extends RecordReader<LongWritable, WikiRevisionWritable> { private static final Logger LOGGER = Logger.getLogger(WikiInputRecordReader.class);
From source file de.l3s.common.hadoop.WholeFileRecordReader.java
class WholeFileRecordReader extends RecordReader<LongWritable, Text> { private FileSplit fileSplit; private Configuration conf; private boolean processed = false;
From source file de.l3s.streamcorpus.terrier.ThriftFileCollectionRecordReader.java
/**
* RecordReader that gets a file split containing urls of streamcorpus files (rather
* than the files themselves) and emit the filename together with the stream items.
*
* @author tuan
*/
From source file de.rwhq.hdfs.index.LineRecordReader.java
/**
* Minor modified version of the original {@code LineRecordReader}.
* Some fields had to be made protected for enabling the IndexedRecordReader to alter them.
*
*
* Treats keys as offset in file and value as line.