Example usage for org.apache.hadoop.mapreduce RecordWriter subclass-usage

List of usage examples for org.apache.hadoop.mapreduce RecordWriter subclass-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapreduce RecordWriter subclass-usage.

Usage

From source file org.apache.gora.mapreduce.GoraRecordWriter.java

/**
 * Hadoop record writer that flushes the Gora datastore regularly.
 *
 */
public class GoraRecordWriter<K, T> extends RecordWriter<K, T> {
    public static final Logger LOG = LoggerFactory.getLogger(GoraRecordWriter.class);

From source file org.apache.hcatalog.mapreduce.RecordWriterContainer.java

/**
 *  This class will contain an implementation of an RecordWriter.
 *  See {@link OutputFormatContainer} for more information about containers.
 */
abstract class RecordWriterContainer extends RecordWriter<WritableComparable<?>, HCatRecord> {

From source file org.apache.hive.hcatalog.mapreduce.RecordWriterContainer.java

/**
 *  This class will contain an implementation of an RecordWriter.
 *  See {@link OutputFormatContainer} for more information about containers.
 */
abstract class RecordWriterContainer extends RecordWriter<WritableComparable<?>, HCatRecord> {

From source file org.apache.jena.grande.mapreduce.io.QuadRecordWriter.java

public class QuadRecordWriter extends RecordWriter<NullWritable, QuadWritable> {

    private static final Logger log = LoggerFactory.getLogger(QuadRecordWriter.class);

    private Writer out;

From source file org.apache.jena.grande.mapreduce.io.TripleRecordWriter.java

public class TripleRecordWriter extends RecordWriter<NullWritable, TripleWritable> {

    @Override
    public void write(NullWritable key, TripleWritable value) throws IOException, InterruptedException {
        // TODO Auto-generated method stub
    }

From source file org.apache.jena.hadoop.rdf.io.output.writers.AbstractBatchedNodeTupleWriter.java

/**
 * An abstract implementation of a record writer that writes records in batches.
 * <p>
 * It is important to note that the writer will write output periodically once
 * sufficient tuples have been gathered. If there is an incomplete batch when
 * the {@link #close(TaskAttemptContext)} method is called then the final batch

From source file org.apache.jena.hadoop.rdf.io.output.writers.AbstractLineBasedNodeTupleWriter.java

/**
 * An abstract implementation of a record writer that writes records to a line
 * based tuple formats.
 * <p>
 * The implementation only writes the value portion of the key value pair since
 * it is the value portion that is used to convey the node tuples

From source file org.apache.jena.hadoop.rdf.io.output.writers.AbstractNodeWriter.java

/**
 * Abstract implementation of a record writer which writes pairs of nodes and
 * arbitrary values to text based files
 * 
 * 
 * 

From source file org.apache.jena.hadoop.rdf.io.output.writers.AbstractStreamRdfNodeTupleWriter.java

public abstract class AbstractStreamRdfNodeTupleWriter<TKey, TTuple, TValue extends AbstractNodeTupleWritable<TTuple>>
        extends RecordWriter<TKey, TValue> {

    private StreamRDF stream;
    private Writer writer;

From source file org.apache.jena.hadoop.rdf.io.output.writers.AbstractWholeFileNodeTupleWriter.java

/**
 * An abstract implementation of a record writer that writes records to whole
 * file formats.
 * <p>
 * It is important to note that the writer does not actually write any output
 * until the {@link #close(TaskAttemptContext)} method is called as it must