Example usage for org.apache.hadoop.mapreduce TaskType MAP

List of usage examples for org.apache.hadoop.mapreduce TaskType MAP

Introduction

In this page you can find the example usage for org.apache.hadoop.mapreduce TaskType MAP.

Prototype

TaskType MAP

To view the source code for org.apache.hadoop.mapreduce TaskType MAP.

Click Source Link

Usage

From source file:org.apache.pig.tools.pigstats.mapreduce.MRJobStats.java

License:Apache License

void addMapReduceStatistics(Job job) {
    TaskReport[] maps = null;/*w w w.ja  va 2s  .co  m*/
    try {
        maps = HadoopShims.getTaskReports(job, TaskType.MAP);
    } catch (IOException e) {
        LOG.warn("Failed to get map task report", e);
    }
    TaskReport[] reduces = null;
    try {
        reduces = HadoopShims.getTaskReports(job, TaskType.REDUCE);
    } catch (IOException e) {
        LOG.warn("Failed to get reduce task report", e);
    }
    addMapReduceStatistics(maps, reduces);
}

From source file:org.apache.tajo.storage.hbase.HFileAppender.java

License:Apache License

@Override
public void init() throws IOException {
    super.init();

    Configuration taskConf = new Configuration();
    Path stagingResultDir = new Path(stagingDir, TajoConstants.RESULT_DIR_NAME);
    taskConf.set(FileOutputFormat.OUTDIR, stagingResultDir.toString());

    ExecutionBlockId ebId = taskAttemptId.getTaskId().getExecutionBlockId();
    writerContext = new TaskAttemptContextImpl(taskConf, new TaskAttemptID(ebId.getQueryId().toString(),
            ebId.getId(), TaskType.MAP, taskAttemptId.getTaskId().getId(), taskAttemptId.getId()));

    HFileOutputFormat2 hFileOutputFormat2 = new HFileOutputFormat2();
    try {// ww w . j av  a 2s . c o m
        writer = hFileOutputFormat2.getRecordWriter(writerContext);

        committer = new FileOutputCommitter(FileOutputFormat.getOutputPath(writerContext), writerContext);
        workingFilePath = committer.getWorkPath();
    } catch (InterruptedException e) {
        throw new IOException(e.getMessage(), e);
    }

    LOG.info("Created hbase file writer: " + workingFilePath);
}

From source file:org.apache.tez.mapreduce.combine.MRCombiner.java

License:Apache License

public MRCombiner(TaskContext taskContext) throws IOException {
    this.conf = TezUtils.createConfFromUserPayload(taskContext.getUserPayload());

    assert (taskContext instanceof InputContext || taskContext instanceof OutputContext);
    if (taskContext instanceof OutputContext) {
        this.keyClass = ConfigUtils.getIntermediateOutputKeyClass(conf);
        this.valClass = ConfigUtils.getIntermediateOutputValueClass(conf);
        this.comparator = ConfigUtils.getIntermediateOutputKeyComparator(conf);
        this.reporter = new MRTaskReporter((OutputContext) taskContext);
    } else {/*from  w ww . jav a 2 s .  co m*/
        this.keyClass = ConfigUtils.getIntermediateInputKeyClass(conf);
        this.valClass = ConfigUtils.getIntermediateInputValueClass(conf);
        this.comparator = ConfigUtils.getIntermediateInputKeyComparator(conf);
        this.reporter = new MRTaskReporter((InputContext) taskContext);
    }

    this.useNewApi = ConfigUtils.useNewApi(conf);

    combineInputKeyCounter = taskContext.getCounters().findCounter(TaskCounter.COMBINE_INPUT_RECORDS);
    combineInputValueCounter = taskContext.getCounters().findCounter(TaskCounter.COMBINE_OUTPUT_RECORDS);

    boolean isMap = conf.getBoolean(MRConfig.IS_MAP_PROCESSOR, false);
    this.mrTaskAttemptID = new TaskAttemptID(
            new TaskID(String.valueOf(taskContext.getApplicationId().getClusterTimestamp()),
                    taskContext.getApplicationId().getId(), isMap ? TaskType.MAP : TaskType.REDUCE,
                    taskContext.getTaskIndex()),
            taskContext.getTaskAttemptNumber());

    LOG.info("Using combineKeyClass: " + keyClass + ", combineValueClass: " + valClass + ", combineComparator: "
            + comparator + ", useNewApi: " + useNewApi);
}

From source file:org.apache.tez.mapreduce.committer.MROutputCommitter.java

License:Apache License

@SuppressWarnings("rawtypes")
private org.apache.hadoop.mapreduce.OutputCommitter getOutputCommitter(OutputCommitterContext context) {

    org.apache.hadoop.mapreduce.OutputCommitter committer = null;
    newApiCommitter = false;//from   w w  w  .  jav  a2  s  .  co  m
    if (jobConf.getBoolean("mapred.reducer.new-api", false)
            || jobConf.getBoolean("mapred.mapper.new-api", false)) {
        newApiCommitter = true;
        LOG.info("Using mapred newApiCommitter.");
    }

    if (newApiCommitter) {
        TaskAttemptID taskAttemptID = new TaskAttemptID(
                Long.toString(context.getApplicationId().getClusterTimestamp()),
                context.getApplicationId().getId(),
                ((jobConf.getBoolean(MRConfig.IS_MAP_PROCESSOR, false) ? TaskType.MAP : TaskType.REDUCE)), 0,
                context.getDAGAttemptNumber());

        TaskAttemptContext taskContext = new TaskAttemptContextImpl(jobConf, taskAttemptID);
        try {
            OutputFormat outputFormat = ReflectionUtils.newInstance(taskContext.getOutputFormatClass(),
                    jobConf);
            committer = outputFormat.getOutputCommitter(taskContext);
        } catch (Exception e) {
            throw new TezUncheckedException(e);
        }
    } else {
        committer = ReflectionUtils.newInstance(jobConf.getClass("mapred.output.committer.class",
                FileOutputCommitter.class, org.apache.hadoop.mapred.OutputCommitter.class), jobConf);
    }
    LOG.info("OutputCommitter for outputName=" + context.getOutputName() + ", vertexName="
            + context.getVertexName() + ", outputCommitterClass=" + committer.getClass().getName());
    return committer;
}

From source file:org.apache.tez.mapreduce.committer.MROutputCommitter.java

License:Apache License

@Override
public void recoverTask(int taskIndex, int attemptId) throws IOException {
    if (!initialized) {
        throw new RuntimeException("Committer not initialized");
    }/* w  ww .  j ava2 s.  co  m*/
    TaskAttemptID taskAttemptID = new TaskAttemptID(
            Long.toString(getContext().getApplicationId().getClusterTimestamp())
                    + String.valueOf(getContext().getVertexIndex()),
            getContext().getApplicationId().getId(),
            ((jobConf.getBoolean(MRConfig.IS_MAP_PROCESSOR, false) ? TaskType.MAP : TaskType.REDUCE)),
            taskIndex, attemptId);
    TaskAttemptContext taskContext = new TaskAttemptContextImpl(jobConf, taskAttemptID);
    committer.recoverTask(taskContext);
}

From source file:org.apache.tez.mapreduce.hadoop.IDConverter.java

License:Apache License

public static TaskID toMRTaskId(TezTaskID taskid) {
    return new TaskID(toMRJobId(taskid.getVertexID().getDAGId()),
            taskid.getVertexID().getId() == 0 ? TaskType.MAP : TaskType.REDUCE, taskid.getId());
}

From source file:org.apache.tez.mapreduce.hadoop.IDConverter.java

License:Apache License

public static TezTaskID fromMRTaskId(org.apache.hadoop.mapreduce.TaskID taskid) {
    return TezTaskID.getInstance(TezVertexID.getInstance(fromMRJobId(taskid.getJobID()),
            (taskid.getTaskType() == TaskType.MAP ? 0 : 1)), taskid.getId());
}

From source file:org.apache.tez.mapreduce.hadoop.mapreduce.TaskAttemptContextImpl.java

License:Apache License

public static org.apache.hadoop.mapred.TaskAttemptID createMockTaskAttemptID(long clusterId, int vertexIndex,
        int appId, int taskIndex, int taskAttemptNumber, boolean isMap) {
    return new org.apache.hadoop.mapred.TaskAttemptID(
            new org.apache.hadoop.mapred.TaskID(String.valueOf(clusterId) + String.valueOf(vertexIndex), appId,
                    isMap ? TaskType.MAP : TaskType.REDUCE, taskIndex),
            taskAttemptNumber);//  ww  w  . ja  v  a  2  s. c o m
}

From source file:org.apache.tez.mapreduce.hadoop.mapreduce.TaskAttemptContextImpl.java

License:Apache License

public static org.apache.hadoop.mapred.TaskAttemptID createMockTaskAttemptIDFromTezTaskAttemptId(
        TezTaskAttemptID tezTaId, boolean isMap) {
    TezVertexID vId = tezTaId.getTaskID().getVertexID();
    ApplicationId appId = vId.getDAGId().getApplicationId();
    return new org.apache.hadoop.mapred.TaskAttemptID(new org.apache.hadoop.mapred.TaskID(
            String.valueOf(appId.getClusterTimestamp()) + String.valueOf(vId.getId()), appId.getId(),
            isMap ? TaskType.MAP : TaskType.REDUCE, tezTaId.getTaskID().getId()), tezTaId.getId());
}

From source file:org.apache.tez.mapreduce.hadoop.mapreduce.TaskAttemptContextImpl.java

License:Apache License

public static org.apache.hadoop.mapred.TaskID createMockTaskAttemptIDFromTezTaskId(TezTaskID tezTaId,
        boolean isMap) {
    TezVertexID vId = tezTaId.getVertexID();
    ApplicationId appId = vId.getDAGId().getApplicationId();
    return new org.apache.hadoop.mapred.TaskID(
            String.valueOf(appId.getClusterTimestamp()) + String.valueOf(vId.getId()), appId.getId(),
            isMap ? TaskType.MAP : TaskType.REDUCE, tezTaId.getId());
}