List of usage examples for org.apache.hadoop.mapred FileInputFormat subclass-usage
From source file WholeSplitInputFormat.java
public class WholeSplitInputFormat extends FileInputFormat<NullWritable, TextArrayWritable> { @Override protected boolean isSplitable(FileSystem fs, Path filename) { return true; }
From source file PageInputFormat.java
/** An {@link InputFormat} for plain text files. Files are broken into lines. * Either linefeed or carriage-return are used to signal end of line. Keys are * the position in the file, and values are the line of text.. */ public class PageInputFormat extends FileInputFormat<Text, Text> { public RecordReader<Text, Text> getRecordReader(InputSplit arg0, JobConf arg1, Reporter arg2) throws IOException {
From source file DeprecatedBAMInputFormat.java
public class DeprecatedBAMInputFormat extends FileInputFormat<LongWritable, SAMRecordWritable> { @Override public RecordReader<LongWritable, SAMRecordWritable> getRecordReader(InputSplit split, JobConf job, Reporter reporter) throws IOException { return new DeprecatedBAMRecordReader(split, job, reporter); }
From source file DeprecatedBAMBaseInputFormat.java
public class DeprecatedBAMBaseInputFormat extends FileInputFormat<LongWritable, SAMBaseRecord> { private final DeprecatedBAMInputFormat bamInputFormat = new DeprecatedBAMInputFormat(); @Override public RecordReader<LongWritable, SAMBaseRecord> getRecordReader(InputSplit split, JobConf job, Reporter reporter) throws IOException {
From source file NewPageInputFormat.java
/** An {@link InputFormat} for plain text files. Files are broken into lines. * Either linefeed or carriage-return are used to signal end of line. Keys are * the position in the file, and values are the line of text.. */ public class NewPageInputFormat extends FileInputFormat<Text, Text> { public RecordReader<Text, Text> getRecordReader(InputSplit arg0, JobConf arg1, Reporter arg2) throws IOException {
From source file Job1RecordReader.java
public class Job1InputFormat extends FileInputFormat<Text, Text> { public RecordReader<Text, Text> getRecordReader(InputSplit arg0, JobConf arg1, Reporter arg2) throws IOException { return new Job1RecordReader(arg1, (FileSplit) arg0); } }
From source file ca.sparkera.adapters.mapred.MainframeVBInputFormat.java
/**
* MainframeVBInputFormat is an input format used to read input files which
* contain binary data in which record starts with RDW (record length) of 4 bytes.
*/
@InterfaceAudience.Public
@InterfaceStability.Stable
From source file cascading.tap.hadoop.ZipInputFormat.java
/**
* Class ZipInputFormat is an {@link InputFormat} for zip files. Each file within a zip file is broken
* into lines. Either line-feed or carriage-return are used to signal end of
* line. Keys are the position in the file, and values are the line of text.
* <p/>
* If the underlying {@link FileSystem} is HDFS or FILE, each {@link ZipEntry} is returned
From source file co.nubetech.hiho.mapred.input.FileStreamInputFormat.java
/**
* Updated version of {@link co.nubetech.hiho.mapreduce.lib.input.FileStreamInputFormat}.
*
* This class returns file streams as records which can be used directly for
* load data infile operations of databases The key is the filename
*
From source file com.alexholmes.hadooputils.sort.SortInputFormatOld.java
/** * The {@link InputFormat} used for reading the source files. The key is the sort key, and the value * is the entire sort line. */ public class SortInputFormatOld extends FileInputFormat<Text, Text> implements JobConfigurable {