Example usage for org.apache.hadoop.mapreduce RecordReader subclass-usage

List of usage examples for org.apache.hadoop.mapreduce RecordReader subclass-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapreduce RecordReader subclass-usage.

Usage

From source file ca.uwaterloo.iss4e.hadoop.io.CartesianRecordReader.java

/**
 *  Copyright (c) 2014 Xiufeng Liu ( xiufeng.liu@uwaterloo.ca )
 *
 *  This file is free software: you may copy, redistribute and/or modify it
 *  under the terms of the GNU General Public License version 2
 *  as published by the Free Software Foundation.

From source file chaohBIM.ZipFileRecordReader.java

/**
 * This RecordReader implementation extracts individual files from a ZIP
 * file and hands them over to the Mapper. The "key" is the decompressed
 * file name, the "value" is the file contents.
 */
public class ZipFileRecordReader extends RecordReader<Text, BytesWritable> {

From source file chaohParse.ZipFileRecordReader.java

/**
 * This RecordReader implementation extracts individual files from a ZIP
 * file and hands them over to the Mapper. The "key" is the decompressed
 * file name, the "value" is the file contents.
 */
public class ZipFileRecordReader extends RecordReader<Text, BytesWritable> {

From source file cn.uc.hadoop.mapreduce.lib.input.FileNameLineRecordReader.java

/**
 * ??FileNameTextinputformat
 * ??key,?value
 * @author qiujw
 */
public class FileNameLineRecordReader extends RecordReader<Text, Text> {

From source file cn.uc.hadoop.mapreduce.lib.input.FilePathLineRecordReader.java

/**
 * ??FilePathTextinputformat
 * key,value
 * @author qiujw
 */
public class FilePathLineRecordReader extends RecordReader<Text, Text> {

From source file co.cask.cdap.data.stream.StreamRecordReader.java

/**
 * A {@link RecordReader} for reading stream events.
 *
 * @param <K> Key type read by this record reader.
 * @param <V> Value type read by this record reader.
 */

From source file co.cask.cdap.internal.app.runtime.batch.dataset.DataSetRecordReader.java

final class DataSetRecordReader<KEY, VALUE> extends RecordReader<KEY, VALUE> {
    private static final Logger LOG = LoggerFactory.getLogger(DataSetRecordReader.class);
    private final SplitReader<KEY, VALUE> splitReader;
    private final MapReduceContextProvider mrContextProvider;
    private final BasicMapReduceContext context;

From source file co.cask.cdap.internal.app.runtime.batch.dataset.input.DelegatingRecordReader.java

/**
 * This is a delegating RecordReader, which delegates the functionality to the
 * underlying record reader in {@link TaggedInputSplit}
 *
 * @param <K> Type of key
 * @param <V> Type of value

From source file co.cask.cdap.internal.app.runtime.spark.dataset.DatasetRecordReader.java

final class DatasetRecordReader<KEY, VALUE> extends RecordReader<KEY, VALUE> {

    private final SplitReader<KEY, VALUE> splitReader;

    public DatasetRecordReader(final SplitReader<KEY, VALUE> splitReader) {
        this.splitReader = splitReader;

From source file co.cask.hydrator.plugin.batch.CopybookRecordReader.java

/**
 * Record Reader for CopybookReader plugin.
 * <p>
 * This will return the field name and value using the binary data and copybook contents, to be used as the
 * transform method input.
 */