Example usage for org.apache.hadoop.mapreduce RecordReader subclass-usage

List of usage examples for org.apache.hadoop.mapreduce RecordReader subclass-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapreduce RecordReader subclass-usage.

Usage

From source file com.yahoo.druid.hadoop.DruidRecordReader.java

public class DruidRecordReader extends RecordReader<DateTime, Map<String, Object>> {

    private static final Logger logger = LoggerFactory.getLogger(DruidRecordReader.class);

    private IngestSegmentFirehose rowYielder;

From source file com.zjy.mongo.input.BSONFileRecordReader.java

/**
 * <p>
 * Copyright (c) 2008 - 2013 10gen, Inc. (http://10gen.com)
 * </p>
 * <p>
 * Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may

From source file com.zjy.mongo.input.MongoRecordReader.java

public class MongoRecordReader extends RecordReader<Object, BSONObject> {

    public MongoRecordReader(final MongoInputSplit split) {
        this.split = split;
        cursor = split.getCursor();
    }

From source file cosmos.mapred.LongLineRecordReader.java

/**
 * A copy of {@link LineRecordReader} which does not discard lines longer than "mapred.linerecordreader.maxlength". Instead, it returns them, leaving it to the
 * mapper to decide what to do with it. It also does not treat '\r' (CR) characters as new lines -- it uses {@link LfLineReader} instead of {@link LineReader}
 * to read lines.
 */
public class LongLineRecordReader extends RecordReader<LongWritable, Text> {

From source file crunch.MaxTemperature.java

    class WholeFileRecordReader extends RecordReader<NullWritable, BytesWritable> {

        private FileSplit fileSplit;
        private Configuration conf;
        private BytesWritable value = new BytesWritable();
        private boolean processed = false;

From source file cs480a2.yqiu.recSystem.mapreduce.input.SingleBookReader.java

/**
 * Created by Qiu on 3/18/15.
 * This record reader takes a single book as input.
 * Output key: Text ---> " Title / Maximum Word Count "
 * Output value: TextArray ---> [ "Word A / Word A Count", ... ]
 */

From source file de.gesundkrank.wikipedia.hadoop.inputformat.WikiInputRecordReader.java

/**
 * Jan Graegger<jan.grassegger@uni-weimar.de>
 */
public class WikiInputRecordReader extends RecordReader<LongWritable, WikiRevisionWritable> {
    private static final Logger LOGGER = Logger.getLogger(WikiInputRecordReader.class);

From source file de.l3s.common.hadoop.WholeFileRecordReader.java

class WholeFileRecordReader extends RecordReader<LongWritable, Text> {

    private FileSplit fileSplit;
    private Configuration conf;
    private boolean processed = false;

From source file de.l3s.streamcorpus.terrier.ThriftFileCollectionRecordReader.java

/**
 * RecordReader that gets a file split containing urls of streamcorpus files (rather
 * than the files themselves) and emit the filename together with the stream items. 
 * 
 * @author tuan
 */

From source file de.rwhq.hdfs.index.LineRecordReader.java

/**
 * Minor modified version of the original {@code LineRecordReader}.
 * Some fields had to be made protected for enabling the IndexedRecordReader to alter them.
 *
 *
 * Treats keys as offset in file and value as line.