List of usage examples for org.apache.hadoop.mapred FileOutputFormat subclass-usage
From source file IndexService.IndexIFormatOutputFormat.java
@SuppressWarnings("deprecation") public class IndexIFormatOutputFormat<K, V> extends FileOutputFormat<IndexKey, IndexValue> { public static final Log LOG = LogFactory.getLog(IndexIFormatOutputFormat.class); @Override public RecordWriter<IndexKey, IndexValue> getRecordWriter(FileSystem ignored, JobConf job, String name,
From source file IndexService.IndexMergeIFormatOutputFormat.java
@SuppressWarnings("deprecation") public class IndexMergeIFormatOutputFormat<K, V> extends FileOutputFormat<IndexKey, IndexValue> { public static final Log LOG = LogFactory.getLog(IndexMergeIFormatOutputFormat.class); @Override public RecordWriter<IndexKey, IndexValue> getRecordWriter(FileSystem ignored, JobConf job, String name,
From source file lennard.MandelbrotOutputFormat.java
/**
* An {@link OutputFormat} that writes plain text files.
*
* @deprecated Use
* {@link org.apache.hadoop.mapreduce.lib.output.TextOutputFormat}
* instead.
From source file net.iponweb.hadoop.streaming.avro.AvroAsJsonOutputFormat.java
/**
* Output format for streaming jobs converting JSON representation into Avro record
* Multiple output schemas supported in combination with ByKeyOutputFormat. To use
* this feature one must set configuration option
* 'iow.streaming.schema.use.prefix' to true, provide output
* schemas as files to the job and name them in the first column of output in following
From source file net.iponweb.hadoop.streaming.io.ByKeyOutputFormat.java
public class ByKeyOutputFormat extends FileOutputFormat<Text, Text> { private static final Log LOG = LogFactory.getLog(net.iponweb.hadoop.streaming.io.ByKeyOutputFormat.class); private OutputFormat<Text, Text> internalOutputFormat; private KeyValueSplitter splitter; private boolean assumeFileNamesSorted; private HashMap<String, String> SupportedOutputFormats = new HashMap<String, String>();
From source file net.iponweb.hadoop.streaming.parquet.ParquetAsTextOutputFormat.java
public class ParquetAsTextOutputFormat extends FileOutputFormat<Text, Text> { private static final Log LOG = LogFactory.getLog(ParquetAsTextOutputFormat.class); protected ParquetOutputFormat<SimpleGroup> realOutputFormat = new ParquetOutputFormat<>(); public static void setWriteSupportClass(Configuration configuration, Class<?> writeSupportClass) {
From source file newprotobuf.mapred.ProtobufOutputFormat.java
public class ProtobufOutputFormat<K, V> extends FileOutputFormat<K, V> { protected static class ProtobufRecordWriter<K, V> implements RecordWriter<K, V> { protected recordio.RecordWriter recordWriter; protected DataOutputStream out;
From source file org.apache.accumulo.core.client.mapred.AccumuloFileOutputFormat.java
/**
* This class allows MapReduce jobs to write output in the Accumulo data file format.<br>
* Care should be taken to write only sorted data (sorted by {@link Key}), as this is an important requirement of Accumulo data files.
*
* <p>
* The output path to be created must be specified via {@link AccumuloFileOutputFormat#setOutputPath(JobConf, Path)}. This is inherited from
From source file org.apache.accumulo.hadoop.mapred.AccumuloFileOutputFormat.java
/** * @see org.apache.accumulo.hadoop.mapreduce.AccumuloFileOutputFormat * * @since 2.0 */ public class AccumuloFileOutputFormat extends FileOutputFormat<Key, Value> {
From source file org.apache.avro.mapred.AvroOutputFormat.java
/**
* An {@link org.apache.hadoop.mapred.OutputFormat} for Avro data files.
* <p/>
* You can specify various options using Job Configuration properties.
* Look at the fields in {@link AvroJob} as well as this class to get
* an overview of the supported options.