Example usage for org.apache.hadoop.mapred FileInputFormat subclass-usage

List of usage examples for org.apache.hadoop.mapred FileInputFormat subclass-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapred FileInputFormat subclass-usage.

Usage

From source file WholeSplitInputFormat.java

public class WholeSplitInputFormat extends FileInputFormat<NullWritable, TextArrayWritable> {
    @Override
    protected boolean isSplitable(FileSystem fs, Path filename) {
        return true;
    }

From source file PageInputFormat.java

/** An {@link InputFormat} for plain text files.  Files are broken into lines.
 * Either linefeed or carriage-return are used to signal end of line.  Keys are
 * the position in the file, and values are the line of text.. */
public class PageInputFormat extends FileInputFormat<Text, Text> {
    public RecordReader<Text, Text> getRecordReader(InputSplit arg0, JobConf arg1, Reporter arg2)
            throws IOException {

From source file DeprecatedBAMInputFormat.java

public class DeprecatedBAMInputFormat extends FileInputFormat<LongWritable, SAMRecordWritable> {
    @Override
    public RecordReader<LongWritable, SAMRecordWritable> getRecordReader(InputSplit split, JobConf job,
            Reporter reporter) throws IOException {
        return new DeprecatedBAMRecordReader(split, job, reporter);
    }

From source file DeprecatedBAMBaseInputFormat.java

public class DeprecatedBAMBaseInputFormat extends FileInputFormat<LongWritable, SAMBaseRecord> {
    private final DeprecatedBAMInputFormat bamInputFormat = new DeprecatedBAMInputFormat();

    @Override
    public RecordReader<LongWritable, SAMBaseRecord> getRecordReader(InputSplit split, JobConf job,
            Reporter reporter) throws IOException {

From source file NewPageInputFormat.java

/** An {@link InputFormat} for plain text files.  Files are broken into lines.
 * Either linefeed or carriage-return are used to signal end of line.  Keys are
 * the position in the file, and values are the line of text.. */
public class NewPageInputFormat extends FileInputFormat<Text, Text> {
    public RecordReader<Text, Text> getRecordReader(InputSplit arg0, JobConf arg1, Reporter arg2)
            throws IOException {

From source file Job1RecordReader.java

public class Job1InputFormat extends FileInputFormat<Text, Text> {
    public RecordReader<Text, Text> getRecordReader(InputSplit arg0, JobConf arg1, Reporter arg2)
            throws IOException {
        return new Job1RecordReader(arg1, (FileSplit) arg0);
    }
}

From source file ca.sparkera.adapters.mapred.MainframeVBInputFormat.java

/**
 * MainframeVBInputFormat is an input format used to read input files which
 * contain binary data in which record starts with RDW (record length) of 4 bytes. 
 */
@InterfaceAudience.Public
@InterfaceStability.Stable

From source file cascading.tap.hadoop.ZipInputFormat.java

/**
 * Class ZipInputFormat is an {@link InputFormat} for zip files. Each file within a zip file is broken
 * into lines. Either line-feed or carriage-return are used to signal end of
 * line. Keys are the position in the file, and values are the line of text.
 * <p/>
 * If the underlying {@link FileSystem} is HDFS or FILE, each {@link ZipEntry} is returned

From source file co.nubetech.hiho.mapred.input.FileStreamInputFormat.java

/**
 * Updated version of {@link co.nubetech.hiho.mapreduce.lib.input.FileStreamInputFormat}.
 *
 * This class returns file streams as records which can be used directly for
 * load data infile operations of databases The key is the filename
 *

From source file com.alexholmes.hadooputils.sort.SortInputFormatOld.java

/**
 * The {@link InputFormat} used for reading the source files. The key is the sort key, and the value
 * is the entire sort line.
 */
public class SortInputFormatOld extends FileInputFormat<Text, Text> implements JobConfigurable {