Example usage for org.apache.hadoop.mapred JobConfigurable interface-usage

List of usage examples for org.apache.hadoop.mapred JobConfigurable interface-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapred JobConfigurable interface-usage.

Usage

From source file org.apache.nutchbase.util.hbase.TableMapReduce.java

@SuppressWarnings("unchecked")
public abstract class TableMapReduce<K extends WritableComparable, V extends Writable> extends Configured
        implements TableMap<K, V>, TableReduce<K, V>, Closeable, JobConfigurable {

    /**
     * Use this before submitting a TableMapReduce job. It will

From source file org.apache.orc.mapred.OrcKey.java

/**
 * This type provides a wrapper for OrcStruct so that it can be sent through
 * the MapReduce shuffle as a key.
 *
 * The user should set the JobConf with orc.mapred.key.type with the type
 * string of the type.

From source file org.apache.orc.mapred.OrcValue.java

/**
 * This type provides a wrapper for OrcStruct so that it can be sent through
 * the MapReduce shuffle as a value.
 *
 * The user should set the JobConf with orc.mapred.value.type with the type
 * string of the type.

From source file org.apache.pig.backend.hadoop.executionengine.mapreduceExec.PigInputFormat.java

public class PigInputFormat implements InputFormat<Text, Tuple>, JobConfigurable {

    @SuppressWarnings("unchecked")
    public InputSplit[] getSplits(JobConf job, int numSplits) throws IOException {
        boolean isSplittable = job.getBoolean("pig.input.splittable", true);
        ArrayList<FileSpec> inputs = (ArrayList<FileSpec>) ObjectSerializer.deserialize(job.get("pig.inputs"));

From source file org.cloudata.core.parallel.hadoop.AbstractTabletInputFormat.java

/**
 * An InputFormat for Table. Table is splited by Tablet. Each InputSplit represents single Tablet.<BR>
 * Key is Row.Key, and value is Row(if RowScan is true) or ScanCell(if RowScan is false).   
 *
 */
public abstract class AbstractTabletInputFormat implements InputFormat, JobConfigurable {

From source file org.cloudata.util.matrix.MatrixInputFormat.java

public class MatrixInputFormat implements InputFormat, JobConfigurable {
    static final Logger LOG = Logger.getLogger(MatrixInputFormat.class.getName());
    public static final String MATRIX_INPUT_TABLE = "cloudata.matrix.input.table";
    public static final String MATRIX_INPUT_COLUMN = "cloudata.matrix.input.column";
    public static final String MATRIX_TARGET_TABLE = "cloudata.matrix.target.table";
    public static final String MATRIX_TARGET_COLUMN = "cloudata.matrix.target.column";

From source file org.commoncrawl.hadoop.io.ARCInputFormat.java

/**
 * A map reduce input format for gzipped ARC files.
 * 
 * @author Albert Chern (Netseer Corp.)
 */
public class ARCInputFormat implements InputFormat, JobConfigurable {

From source file org.commoncrawl.hadoop.io.ARCSplitCalculator.java

/**
 * Calculates splits based on the desired number of files per split and the
 * desired size of each split.
 * 
 * <p>
 * Concrete implementations should override {@link #getARCResources}.

From source file org.commoncrawl.hadoop.io.deprecated.ARCInputFormat.java

/**
 * A map reduce input format for gzipped ARC files.
 * 
 * @author Albert Chern (Netseer Corp.)
 */
public class ARCInputFormat implements InputFormat, JobConfigurable {

From source file org.commoncrawl.hadoop.io.deprecated.ARCSplitCalculator.java

/**
 * Calculates splits based on the desired number of files per split and the
 * desired size of each split.
 * 
 * <p>
 * Concrete implementations should override {@link #getARCResources}.