List of usage examples for org.apache.hadoop.mapreduce OutputFormat subclass-usage
From source file org.apache.cassandra.hadoop2.BulkOutputFormat.java
public class BulkOutputFormat extends OutputFormat<ByteBuffer, List<Mutation>> implements org.apache.hadoop.mapred.OutputFormat<ByteBuffer, List<Mutation>> { @Override public void checkOutputSpecs(JobContext context) { checkOutputSpecs(context.getConfiguration());
From source file org.apache.crunch.impl.mr.run.CrunchOutputFormat.java
public class CrunchOutputFormat<K, V> extends OutputFormat<K, V> { @Override public RecordWriter<K, V> getRecordWriter(TaskAttemptContext taskAttemptContext) throws IOException, InterruptedException { return new RecordWriter<K, V>() { @Override
From source file org.apache.crunch.io.hbase.TableOutputFormat.java
class TableOutputFormat<K> extends OutputFormat<K, Writable> { private final Log LOG = LogFactory.getLog(TableOutputFormat.class); /** Job parameter that specifies the output table. */ public static final String OUTPUT_TABLE = "hbase.mapred.outputtable";
From source file org.apache.fluo.mapreduce.FluoOutputFormat.java
/** * This output format enables the execution of load transaction against a Fluo table using map * reduce. */ public class FluoOutputFormat extends OutputFormat<Loader, NullWritable> {
From source file org.apache.giraph.bsp.BspOutputFormat.java
/** * This is for internal use only. Allows the vertex output format routines * to be called as if a normal Hadoop job. */ public class BspOutputFormat extends OutputFormat<Text, Text> { /** Class logger */
From source file org.apache.gora.mapreduce.GoraOutputFormat.java
/**
* {@link OutputFormat} for Hadoop jobs that want to store the job outputs
* to a Gora store.
* <p>
* Hadoop jobs can be either configured through static
* <code>setOutput()</code> methods, or if the job is not map-only from {@link GoraReducer}.
From source file org.apache.gora.pig.mapreduce.PigGoraOutputFormat.java
public class PigGoraOutputFormat<K, T extends Persistent> extends OutputFormat<K, T> implements Configurable { public static final Logger LOG = LoggerFactory.getLogger(PigGoraOutputFormat.class); protected Configuration conf;
From source file org.apache.hcatalog.mapreduce.HCatBaseOutputFormat.java
public abstract class HCatBaseOutputFormat extends OutputFormat<WritableComparable<?>, HCatRecord> { // static final private Log LOG = LogFactory.getLog(HCatBaseOutputFormat.class); /** * @see org.apache.hcatalog.mapreduce.HCatBaseOutputFormat#getTableSchema(org.apache.hadoop.conf.Configuration)
From source file org.apache.hcatalog.mapreduce.MultiOutputFormat.java
/**
* The MultiOutputFormat class simplifies writing output data to multiple
* outputs.
* <p>
* Multiple output formats can be defined each with its own
* <code>OutputFormat</code> class, own key class and own value class. Any
From source file org.apache.hcatalog.mapreduce.OutputFormatContainer.java
/**
* This container class is used to wrap OutputFormat implementations and augment them with
* behavior necessary to work with HCatalog (ie metastore updates, hcatalog delegation tokens, etc).
* Containers are also used to provide storage specific implementations of some HCatalog features (ie dynamic partitioning).
* Hence users wishing to create storage specific implementations of HCatalog features should implement this class and override
* HCatStorageHandler.getOutputFormatContainer(OutputFormat outputFormat) to return the implementation.