Example usage for org.apache.hadoop.mapreduce Mapper subclass-usage

List of usage examples for org.apache.hadoop.mapreduce Mapper subclass-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapreduce Mapper subclass-usage.

Usage

From source file com.avira.couchdoop.demo.ImportMapper.java

/**
 * Mapper which receives JSON session documents from Couchbase and reformats
 * them as delimited HDFS text files in order to be used by a recommending
 * system.
 */
public class ImportMapper extends Mapper<Text, ViewRow, Text, Text> {

From source file com.avira.couchdoop.exp.CsvToCouchbaseMapper.java

/**
 * This mapper maps key-value pairs read from TSV files as documents in Couchbase by using keys as IDs and values as
 * documents.
 */
public class CsvToCouchbaseMapper extends Mapper<LongWritable, Text, String, CouchbaseAction> {

From source file com.avira.couchdoop.imp.CouchbaseViewToFileMapper.java

/**
 * Mapper class which maps each document retrieved from a Couchbase view key to text key-values.
 */
public class CouchbaseViewToFileMapper extends Mapper<Text, ViewRow, Text, Text> {

    @Override

From source file com.avira.couchdoop.imp.CouchbaseViewToHBaseMapper.java

/**
 * Mapper class which maps each document retrieved from a Couchbase view key to text key-values.
 */
public class CouchbaseViewToHBaseMapper extends Mapper<Text, ViewRow, ImmutableBytesWritable, Put> {

    private String columnFamily;

From source file com.avira.couchdoop.update.CouchbaseUpdateMapper.java

/**
 * This Mapper class is meant to update Couchbase documents, by using data from a configurable InputFormat.
 *
 * <p>The Couchbase keys should be read from the configured InputFormat. Any other data read from that source
 * can be combined with current document values from Couchbase in order to compute the values.</p>
 *

From source file com.baidu.cloud.bmr.mapreduce.AccessLogAnalyzerMapper.java

/**
 * Mapper class for AccessLogAnalyzer.
 *
 */
public class AccessLogAnalyzerMapper extends Mapper<Object, Text, Text, IntWritable> {

From source file com.baidubce.bmr.mapreduce.AccessLogAnalyzerMapper.java

/**
 * Mapper class for AccessLogAnalyzer.
 *
 */
public class AccessLogAnalyzerMapper extends Mapper<Object, Text, Text, IntWritable> {

From source file com.bark.hadoop.lab3.AdjMapper.java

public class AdjMapper extends Mapper<LongWritable, Text, Text, Text> {

    @Override
    protected void map(LongWritable key, Text value, Context context) throws IOException, InterruptedException {
        /**
         * Split up the line into (key,value) pair - separated by tab

From source file com.bark.hadoop.lab3.PageCountMapper.java

public class PageCountMapper extends Mapper<LongWritable, Text, Text, IntWritable> {

    @Override
    protected void map(LongWritable key, Text value, Context context) throws IOException, InterruptedException {
        context.write(new Text("N"), new IntWritable(1));
    }

From source file com.bark.hadoop.lab3.PageRankMapper.java

public class PageRankMapper extends Mapper<LongWritable, Text, Text, Text> {

    @Override
    protected void map(LongWritable key, Text value, Context context) throws IOException, InterruptedException {
        /**
         * We have N, we don't need to read two files For first iteration: Read