List of usage examples for org.apache.hadoop.mapreduce RecordReader subclass-usage
From source file ca.uwaterloo.iss4e.hadoop.io.CartesianRecordReader.java
/**
* Copyright (c) 2014 Xiufeng Liu ( xiufeng.liu@uwaterloo.ca )
*
* This file is free software: you may copy, redistribute and/or modify it
* under the terms of the GNU General Public License version 2
* as published by the Free Software Foundation.
From source file chaohBIM.ZipFileRecordReader.java
/** * This RecordReader implementation extracts individual files from a ZIP * file and hands them over to the Mapper. The "key" is the decompressed * file name, the "value" is the file contents. */ public class ZipFileRecordReader extends RecordReader<Text, BytesWritable> {
From source file chaohParse.ZipFileRecordReader.java
/** * This RecordReader implementation extracts individual files from a ZIP * file and hands them over to the Mapper. The "key" is the decompressed * file name, the "value" is the file contents. */ public class ZipFileRecordReader extends RecordReader<Text, BytesWritable> {
From source file cn.uc.hadoop.mapreduce.lib.input.FileNameLineRecordReader.java
/** * ??FileNameTextinputformat * ??key,?value * @author qiujw */ public class FileNameLineRecordReader extends RecordReader<Text, Text> {
From source file cn.uc.hadoop.mapreduce.lib.input.FilePathLineRecordReader.java
/** * ??FilePathTextinputformat * key,value * @author qiujw */ public class FilePathLineRecordReader extends RecordReader<Text, Text> {
From source file co.cask.cdap.data.stream.StreamRecordReader.java
/**
* A {@link RecordReader} for reading stream events.
*
* @param <K> Key type read by this record reader.
* @param <V> Value type read by this record reader.
*/
From source file co.cask.cdap.internal.app.runtime.batch.dataset.DataSetRecordReader.java
final class DataSetRecordReader<KEY, VALUE> extends RecordReader<KEY, VALUE> { private static final Logger LOG = LoggerFactory.getLogger(DataSetRecordReader.class); private final SplitReader<KEY, VALUE> splitReader; private final MapReduceContextProvider mrContextProvider; private final BasicMapReduceContext context;
From source file co.cask.cdap.internal.app.runtime.batch.dataset.input.DelegatingRecordReader.java
/**
* This is a delegating RecordReader, which delegates the functionality to the
* underlying record reader in {@link TaggedInputSplit}
*
* @param <K> Type of key
* @param <V> Type of value
From source file co.cask.cdap.internal.app.runtime.spark.dataset.DatasetRecordReader.java
final class DatasetRecordReader<KEY, VALUE> extends RecordReader<KEY, VALUE> { private final SplitReader<KEY, VALUE> splitReader; public DatasetRecordReader(final SplitReader<KEY, VALUE> splitReader) { this.splitReader = splitReader;
From source file co.cask.hydrator.plugin.batch.CopybookRecordReader.java
/**
* Record Reader for CopybookReader plugin.
* <p>
* This will return the field name and value using the binary data and copybook contents, to be used as the
* transform method input.
*/