List of usage examples for org.apache.hadoop.mapred JobConfigurable interface-usage
From source file org.apache.nutchbase.util.hbase.TableMapReduce.java
@SuppressWarnings("unchecked") public abstract class TableMapReduce<K extends WritableComparable, V extends Writable> extends Configured implements TableMap<K, V>, TableReduce<K, V>, Closeable, JobConfigurable { /** * Use this before submitting a TableMapReduce job. It will
From source file org.apache.orc.mapred.OrcKey.java
/**
* This type provides a wrapper for OrcStruct so that it can be sent through
* the MapReduce shuffle as a key.
*
* The user should set the JobConf with orc.mapred.key.type with the type
* string of the type.
From source file org.apache.orc.mapred.OrcValue.java
/**
* This type provides a wrapper for OrcStruct so that it can be sent through
* the MapReduce shuffle as a value.
*
* The user should set the JobConf with orc.mapred.value.type with the type
* string of the type.
From source file org.apache.pig.backend.hadoop.executionengine.mapreduceExec.PigInputFormat.java
public class PigInputFormat implements InputFormat<Text, Tuple>, JobConfigurable { @SuppressWarnings("unchecked") public InputSplit[] getSplits(JobConf job, int numSplits) throws IOException { boolean isSplittable = job.getBoolean("pig.input.splittable", true); ArrayList<FileSpec> inputs = (ArrayList<FileSpec>) ObjectSerializer.deserialize(job.get("pig.inputs"));
From source file org.cloudata.core.parallel.hadoop.AbstractTabletInputFormat.java
/** * An InputFormat for Table. Table is splited by Tablet. Each InputSplit represents single Tablet.<BR> * Key is Row.Key, and value is Row(if RowScan is true) or ScanCell(if RowScan is false). * */ public abstract class AbstractTabletInputFormat implements InputFormat, JobConfigurable {
From source file org.cloudata.util.matrix.MatrixInputFormat.java
public class MatrixInputFormat implements InputFormat, JobConfigurable { static final Logger LOG = Logger.getLogger(MatrixInputFormat.class.getName()); public static final String MATRIX_INPUT_TABLE = "cloudata.matrix.input.table"; public static final String MATRIX_INPUT_COLUMN = "cloudata.matrix.input.column"; public static final String MATRIX_TARGET_TABLE = "cloudata.matrix.target.table"; public static final String MATRIX_TARGET_COLUMN = "cloudata.matrix.target.column";
From source file org.commoncrawl.hadoop.io.ARCInputFormat.java
/** * A map reduce input format for gzipped ARC files. * * @author Albert Chern (Netseer Corp.) */ public class ARCInputFormat implements InputFormat, JobConfigurable {
From source file org.commoncrawl.hadoop.io.ARCSplitCalculator.java
/**
* Calculates splits based on the desired number of files per split and the
* desired size of each split.
*
* <p>
* Concrete implementations should override {@link #getARCResources}.
From source file org.commoncrawl.hadoop.io.deprecated.ARCInputFormat.java
/** * A map reduce input format for gzipped ARC files. * * @author Albert Chern (Netseer Corp.) */ public class ARCInputFormat implements InputFormat, JobConfigurable {
From source file org.commoncrawl.hadoop.io.deprecated.ARCSplitCalculator.java
/**
* Calculates splits based on the desired number of files per split and the
* desired size of each split.
*
* <p>
* Concrete implementations should override {@link #getARCResources}.