Example usage for org.apache.hadoop.mapred OutputFormat interface-usage

List of usage examples for org.apache.hadoop.mapred OutputFormat interface-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapred OutputFormat interface-usage.

Usage

From source file cascading.jdbc.db.DBOutputFormat.java

/**
 * A OutputFormat that sends the reduce output to a SQL table.
 * <p/>
 * {@link DBOutputFormat} accepts &lt;key,value&gt; pairs, where
 * key has a type extending DBWritable. Returned {@link RecordWriter}
 * writes <b>only the key</b> to the database with a batch SQL query.

From source file cn.edu.hfut.dmic.webcollectorcluster.fetcher.FetcherOutputFormat.java

/**
 *
 * @author hu
 */
public class FetcherOutputFormat implements OutputFormat<Text, WebWritable> {

From source file co.cask.cdap.hive.datasets.DatasetOutputFormat.java

/**
 * Map reduce output format to write to datasets that implement {@link RecordWritable}.
 */
public class DatasetOutputFormat implements OutputFormat<Void, Text> {
    private static final Logger LOG = LoggerFactory.getLogger(DatasetOutputFormat.class);

From source file co.cask.cdap.hive.stream.UnsupportedOutputFormat.java

/**
 * Map reduce output format to write to datasets that implement {@link co.cask.cdap.api.data.batch.RecordWritable}.
 */
public class UnsupportedOutputFormat implements OutputFormat<Void, Text> {

    @Override

From source file com.aerospike.hadoop.mapreduce.AerospikeOutputFormat.java

public abstract class AerospikeOutputFormat<KK, VV> extends OutputFormat<KK, VV>
        implements org.apache.hadoop.mapred.OutputFormat<KK, VV> {

    private static final Log log = LogFactory.getLog(AerospikeOutputFormat.class);

    public static class AerospikeOutputCommitter extends OutputCommitter {

From source file com.aliyun.openservices.tablestore.hive.TableStoreOutputFormat.java

public class TableStoreOutputFormat implements OutputFormat<Writable, BatchWriteWritable> {
    private static Logger logger = LoggerFactory.getLogger(TableStoreOutputFormat.class);

    @Override
    public RecordWriter<Writable, BatchWriteWritable> getRecordWriter(FileSystem ignored, JobConf job, String name,
            Progressable progress) throws IOException {

From source file com.bah.culvert.hive.CulvertOutputFormat.java

/**
 * Output format for writing values to a culvert table.
 * <p>
 * This implementation is primarily oriented towards HIVE interaction.
 */
@SuppressWarnings("deprecation")

From source file com.dappervision.hbase.mapred.TypedBytesTableOutputFormat.java

/**
 * TypedBytesTableOutputFormat
 */
public class TypedBytesTableOutputFormat implements OutputFormat<TypedBytesWritable, TypedBytesWritable> {
    private static final Logger LOG = LoggerFactory.getLogger(TypedBytesTableOutputFormat.class);
    public static final String OUTPUT_TABLE = "hbase.mapred.outputtable";

From source file com.gemstone.gemfire.cache.hdfs.internal.hoplog.mapred.GFOutputFormat.java

/**
 * Output format for gemfire. The records provided to writers created by this
 * output format are PUT in a live gemfire cluster.
 * 
 * @author ashvina
 */

From source file com.github.dryangkun.hbase.tidx.hive.HiveHBaseTableOutputFormat.java

/**
 * HiveHBaseTableOutputFormat implements HiveOutputFormat for HBase tables.
 * We also need to implement the @deprecated org.apache.hadoop.mapred.OutFormat<?,?>
 * class to keep it compliant with Hive interfaces.
 */
public class HiveHBaseTableOutputFormat extends TableOutputFormat<ImmutableBytesWritable>