List of usage examples for org.apache.hadoop.mapreduce Reducer subclass-usage
From source file gaffer.accumulo.bulkimport.BulkImportReducer.java
/**
* Reducer for use in bulk import of data into Accumulo. It merges all values associated
* to the key by converting them into {@link SetOfStatistics} and then merges
* those, and then converts them back to an Accumulo value.
*
* It contains an optimisation so that if there is only one value, we simply
From source file gaffer.accumulo.splitpoints.EstimateSplitPointsReducer.java
/** * Reducer class used for estimating the split points to ensure even distribution of * data in Accumulo after initial insert. */ public class EstimateSplitPointsReducer extends Reducer<Key, Value, Key, Value> {
From source file gaffer.accumulostore.operation.hdfs.handler.job.AccumuloKeyValueReducer.java
/**
* Reducer for use in bulk import of data into Accumulo. It merges all values
* associated to the gaffer.accumulostore.key by converting them into
* {@link gaffer.data.element.Properties} and then merges those, and then
* converts them back to an Accumulo value.
* <p>
From source file gaffer.accumulostore.operation.hdfs.handler.job.AddElementsFromHdfsReducer.java
/**
* Reducer for use in bulk import of data into Accumulo. It merges all values
* associated to the gaffer.accumulostore.key by converting them into
* {@link gaffer.data.element.Properties} and then merges those, and then
* converts them back to an Accumulo value.
* <p>
From source file gaffer.accumulostore.operation.hdfs.reducer.AccumuloKeyValueReducer.java
/**
* Reducer for use in bulk import of data into Accumulo. It merges all values
* associated to the gaffer.accumulostore.key by converting them into
* {@link gaffer.data.element.Properties} and then merges those, and then
* converts them back to an Accumulo value.
* <p>
From source file gaffer.analytic.impl.GraphStatisticsReducer.java
/** * Simply merges together all the {@link SetOfStatistics} for the key. */ public class GraphStatisticsReducer extends Reducer<Text, SetOfStatistics, Text, SetOfStatistics> { protected void reduce(Text key, Iterable<SetOfStatistics> values, Context context)
From source file gobblin.compaction.mapreduce.avro.AvroKeyDedupReducer.java
/**
* Reducer class for compaction MR job for Avro data.
*
* If there are multiple values of the same key, it keeps the last value read.
*
* @author Ziyang Liu
From source file gov.jgi.meta.hadoop.reduce.AssembleByGroupKey.java
/** * simple reducer that just outputs the matches grouped by gene */ public class AssembleByGroupKey extends Reducer<Text, Text, Text, Text> { private IntWritable result = new IntWritable();
From source file gov.jgi.meta.hadoop.reduce.IdentityReducerGroupByKey.java
/** * simple reducer that just outputs the matches grouped by gene */ public class IdentityReducerGroupByKey extends Reducer<Text, Text, Text, Text> { public void reduce(Text key, Iterable<Text> values, Context context) throws InterruptedException, IOException { StringBuilder sb = new StringBuilder();
From source file gov.llnl.ontology.mapreduce.stats.WordCountSumReducer.java
/** * @author Keith Stevens */ public class WordCountSumReducer extends Reducer<StringPair, IntWritable, StringPair, IntWritable> { /**