Example usage for org.apache.hadoop.io Writable interface-usage

List of usage examples for org.apache.hadoop.io Writable interface-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.io Writable interface-usage.

Usage

From source file co.cask.cdap.hive.stream.StreamInputSplit.java

/**
 * Represents a mapred InputSplit for stream. This class must extends from {@link FileSplit}, since Hive expected so.
 * For Stream, since we have all the knowledge in the storage handler about the stream
 * location (and is dynamic due to advancement in stream generation id due to truncation), there is no point to
 * use LOCATION in the table. However, for external table in Hive, if LOCATION was not set during table creation
 * time, Hive will assume the location to be some internally managed path and will fail the query job if

From source file co.cask.cdap.internal.app.runtime.batch.dataset.DataSetInputSplit.java

/**
 * Class for {@link InputSplit} of a {@link Dataset}.
 *
 * @see InputSplit
 */
public class DataSetInputSplit extends InputSplit implements Writable {

From source file co.cask.cdap.internal.app.runtime.batch.dataset.input.TaggedInputSplit.java

/**
 * An {@link InputSplit} that tags another InputSplit with extra data for use
 * by {@link DelegatingInputFormat}s.
 */
public class TaggedInputSplit extends InputSplit implements Configurable, Writable {

From source file co.cask.cdap.template.etl.common.DBRecord.java

/**
 * Writable class for DB Source/Sink
 *
 * @see org.apache.hadoop.mapreduce.lib.db.DBInputFormat DBInputFormat
 * @see org.apache.hadoop.mapreduce.lib.db.DBOutputFormat DBOutputFormat
 * @see org.apache.hadoop.mapreduce.lib.db.DBWritable DBWritable

From source file co.cask.hydrator.plugin.batchSource.KafkaSplit.java

/**
 * Kafka split
 */
public class KafkaSplit extends InputSplit implements Writable {
    private KafkaRequest request;
    private long length = 0;

From source file co.cask.hydrator.plugin.DBRecord.java

/**
 * Writable class for DB Source/Sink
 *
 * @see org.apache.hadoop.mapreduce.lib.db.DBInputFormat DBInputFormat
 * @see org.apache.hadoop.mapreduce.lib.db.DBOutputFormat DBOutputFormat
 * @see DBWritable DBWritable

From source file co.cask.tephra.persist.TransactionEdit.java

/**
 * Represents a transaction state change in the {@link TransactionLog}.
 */
public class TransactionEdit implements Writable {

    /**

From source file co.nubetech.hiho.merge.HihoValue.java

public class HihoValue<V extends Writable> implements Writable {

    protected BooleanWritable isOld;
    protected V val;
    private Text valClass;

From source file com.aerospike.hadoop.mapreduce.AerospikeRecord.java

public class AerospikeRecord implements Writable {

    public Map<String, Object> bins;
    public List<Map<String, Object>> duplicates;
    public int generation;
    public int expiration;

From source file com.alibaba.wasp.ipc.ProtocolSignature.java

public class ProtocolSignature implements Writable {
    static { // register a ctor
        WritableFactories.setFactory(ProtocolSignature.class, new WritableFactory() {
            public Writable newInstance() {
                return new ProtocolSignature();
            }