List of usage examples for org.apache.hadoop.mapreduce OutputFormat subclass-usage
From source file hadoop.MysqlDBOutputFormat.java
/**
* A OutputFormat that sends the reduce output to a SQL table.
* <p>
* {@link MysqlDBOutputFormat} accepts <key,value> pairs, where
* key has a type extending DBWritable. Returned {@link RecordWriter}
* writes <b>only the key</b> to the database with a batch SQL query.
From source file hanb.elasticsearch.expert.hadoop.mr.PrintStreamOutputFormat.java
public class PrintStreamOutputFormat extends org.apache.hadoop.mapreduce.OutputFormat implements OutputFormat { private Stream stream; private class PrintStreamRecordWriter extends org.apache.hadoop.mapreduce.RecordWriter implements RecordWriter {
From source file io.fluo.mapreduce.FluoOutputFormat.java
/** * This output format enables the execution of load transaction against a Fluo table using map reduce. */ public class FluoOutputFormat extends OutputFormat<Loader, NullWritable> { private static String PROPS_CONF_KEY = FluoOutputFormat.class.getName() + ".props";
From source file kafka.bridge.hadoop.KafkaOutputFormat.java
public class KafkaOutputFormat<W extends BytesWritable> extends OutputFormat<NullWritable, W> { public static final String KAFKA_URL = "kafka.output.url"; public static final int KAFKA_PRODUCER_CONNECT_TIMEOUT = 30 * 1000; public static final int KAFKA_PRODUCER_RECONNECT_INTERVAL = 1000; public static final int KAFKA_PRODUCER_BUFFER_SIZE = 64 * 1024; public static final int KAFKA_PRODUCER_MAX_MESSAGE_SIZE = 1024 * 1024;
From source file kafka.bridge.hadoop2.KafkaOutputFormat.java
public class KafkaOutputFormat<K, V> extends OutputFormat<K, V> { private Logger log = Logger.getLogger(KafkaOutputFormat.class); public static final String KAFKA_URL = "kafka.output.url"; /** Bytes to buffer before the OutputFormat does a send (i.e., the amortization window): * We set the default to a million bytes so that the server will not reject the batch of messages
From source file link.neolink.datamonster.DBUpdateOutputFormat.java
/**
* A OutputFormat that sends the reduce output to a SQL table.
*
* {@link DBUpdateOutputFormat} accepts <key,value> pairs, where
* key has a type extending DBWritable. Returned {@link RecordWriter}
* writes <b>only the key</b> to the database with a batch SQL query.
From source file ml.shifu.guagua.mapreduce.GuaguaOutputFormat.java
/** * This is for internal use only. In guagua framework, we don't use mapreduce output format, so this is empty * implementation. */ public class GuaguaOutputFormat extends OutputFormat<Text, Text> {
From source file mvm.rya.accumulo.mr.RyaOutputFormat.java
/**
* Hadoop Map/Reduce class to use Rya, the {@link GeoIndexer}, the {@link FreeTextIndexer}, and the {@link TemporalIndexer} as the sink of {@link Statement} data.
* wrapped in an {@link StatementWritable} objects. This {@link OutputFormat} ignores the Keys and only writes the Values to Rya.
*
* The user must specify connection parameters for Rya, {@link GeoIndexer}, {@link FreeTextIndexer}, and {@link TemporalIndexer}.
*/
From source file org.apache.accumulo.core.client.mapreduce.AccumuloOutputFormat.java
/**
* This class allows MapReduce jobs to use Accumulo as the sink for data. This {@link OutputFormat} accepts keys and values of type {@link Text} (for a table
* name) and {@link Mutation} from the Map and Reduce functions.
*
* The user must specify the following via static configurator methods:
*
From source file org.apache.accumulo.examples.wikisearch.output.SortingRFileOutputFormat.java
public class SortingRFileOutputFormat extends OutputFormat<Text, Mutation> { // private static final Logger log = Logger.getLogger(SortingRFileOutputFormat.class); public static final String PATH_NAME = "sortingrfileoutputformat.path"; public static final String MAX_BUFFER_SIZE = "sortingrfileoutputformat.max.buffer.size";