List of usage examples for org.apache.hadoop.mapreduce Mapper subclass-usage
From source file com.avira.couchdoop.demo.ImportMapper.java
/** * Mapper which receives JSON session documents from Couchbase and reformats * them as delimited HDFS text files in order to be used by a recommending * system. */ public class ImportMapper extends Mapper<Text, ViewRow, Text, Text> {
From source file com.avira.couchdoop.exp.CsvToCouchbaseMapper.java
/** * This mapper maps key-value pairs read from TSV files as documents in Couchbase by using keys as IDs and values as * documents. */ public class CsvToCouchbaseMapper extends Mapper<LongWritable, Text, String, CouchbaseAction> {
From source file com.avira.couchdoop.imp.CouchbaseViewToFileMapper.java
/** * Mapper class which maps each document retrieved from a Couchbase view key to text key-values. */ public class CouchbaseViewToFileMapper extends Mapper<Text, ViewRow, Text, Text> { @Override
From source file com.avira.couchdoop.imp.CouchbaseViewToHBaseMapper.java
/** * Mapper class which maps each document retrieved from a Couchbase view key to text key-values. */ public class CouchbaseViewToHBaseMapper extends Mapper<Text, ViewRow, ImmutableBytesWritable, Put> { private String columnFamily;
From source file com.avira.couchdoop.update.CouchbaseUpdateMapper.java
/**
* This Mapper class is meant to update Couchbase documents, by using data from a configurable InputFormat.
*
* <p>The Couchbase keys should be read from the configured InputFormat. Any other data read from that source
* can be combined with current document values from Couchbase in order to compute the values.</p>
*
From source file com.baidu.cloud.bmr.mapreduce.AccessLogAnalyzerMapper.java
/** * Mapper class for AccessLogAnalyzer. * */ public class AccessLogAnalyzerMapper extends Mapper<Object, Text, Text, IntWritable> {
From source file com.baidubce.bmr.mapreduce.AccessLogAnalyzerMapper.java
/** * Mapper class for AccessLogAnalyzer. * */ public class AccessLogAnalyzerMapper extends Mapper<Object, Text, Text, IntWritable> {
From source file com.bark.hadoop.lab3.AdjMapper.java
public class AdjMapper extends Mapper<LongWritable, Text, Text, Text> { @Override protected void map(LongWritable key, Text value, Context context) throws IOException, InterruptedException { /** * Split up the line into (key,value) pair - separated by tab
From source file com.bark.hadoop.lab3.PageCountMapper.java
public class PageCountMapper extends Mapper<LongWritable, Text, Text, IntWritable> { @Override protected void map(LongWritable key, Text value, Context context) throws IOException, InterruptedException { context.write(new Text("N"), new IntWritable(1)); }
From source file com.bark.hadoop.lab3.PageRankMapper.java
public class PageRankMapper extends Mapper<LongWritable, Text, Text, Text> { @Override protected void map(LongWritable key, Text value, Context context) throws IOException, InterruptedException { /** * We have N, we don't need to read two files For first iteration: Read