List of usage examples for org.apache.hadoop.mapred OutputFormat interface-usage
From source file cascading.jdbc.db.DBOutputFormat.java
/**
* A OutputFormat that sends the reduce output to a SQL table.
* <p/>
* {@link DBOutputFormat} accepts <key,value> pairs, where
* key has a type extending DBWritable. Returned {@link RecordWriter}
* writes <b>only the key</b> to the database with a batch SQL query.
From source file cn.edu.hfut.dmic.webcollectorcluster.fetcher.FetcherOutputFormat.java
/** * * @author hu */ public class FetcherOutputFormat implements OutputFormat<Text, WebWritable> {
From source file co.cask.cdap.hive.datasets.DatasetOutputFormat.java
/** * Map reduce output format to write to datasets that implement {@link RecordWritable}. */ public class DatasetOutputFormat implements OutputFormat<Void, Text> { private static final Logger LOG = LoggerFactory.getLogger(DatasetOutputFormat.class);
From source file co.cask.cdap.hive.stream.UnsupportedOutputFormat.java
/** * Map reduce output format to write to datasets that implement {@link co.cask.cdap.api.data.batch.RecordWritable}. */ public class UnsupportedOutputFormat implements OutputFormat<Void, Text> { @Override
From source file com.aerospike.hadoop.mapreduce.AerospikeOutputFormat.java
public abstract class AerospikeOutputFormat<KK, VV> extends OutputFormat<KK, VV> implements org.apache.hadoop.mapred.OutputFormat<KK, VV> { private static final Log log = LogFactory.getLog(AerospikeOutputFormat.class); public static class AerospikeOutputCommitter extends OutputCommitter {
From source file com.aliyun.openservices.tablestore.hive.TableStoreOutputFormat.java
public class TableStoreOutputFormat implements OutputFormat<Writable, BatchWriteWritable> { private static Logger logger = LoggerFactory.getLogger(TableStoreOutputFormat.class); @Override public RecordWriter<Writable, BatchWriteWritable> getRecordWriter(FileSystem ignored, JobConf job, String name, Progressable progress) throws IOException {
From source file com.bah.culvert.hive.CulvertOutputFormat.java
/** * Output format for writing values to a culvert table. * <p> * This implementation is primarily oriented towards HIVE interaction. */ @SuppressWarnings("deprecation")
From source file com.dappervision.hbase.mapred.TypedBytesTableOutputFormat.java
/** * TypedBytesTableOutputFormat */ public class TypedBytesTableOutputFormat implements OutputFormat<TypedBytesWritable, TypedBytesWritable> { private static final Logger LOG = LoggerFactory.getLogger(TypedBytesTableOutputFormat.class); public static final String OUTPUT_TABLE = "hbase.mapred.outputtable";
From source file com.gemstone.gemfire.cache.hdfs.internal.hoplog.mapred.GFOutputFormat.java
/**
* Output format for gemfire. The records provided to writers created by this
* output format are PUT in a live gemfire cluster.
*
* @author ashvina
*/
From source file com.github.dryangkun.hbase.tidx.hive.HiveHBaseTableOutputFormat.java
/** * HiveHBaseTableOutputFormat implements HiveOutputFormat for HBase tables. * We also need to implement the @deprecated org.apache.hadoop.mapred.OutFormat<?,?> * class to keep it compliant with Hive interfaces. */ public class HiveHBaseTableOutputFormat extends TableOutputFormat<ImmutableBytesWritable>