Example usage for org.apache.hadoop.mapreduce.lib.db DBWritable interface-usage

List of usage examples for org.apache.hadoop.mapreduce.lib.db DBWritable interface-usage

Introduction

In this page you can find the example usage for org.apache.hadoop.mapreduce.lib.db DBWritable interface-usage.

Usage

From source file co.cask.cdap.template.etl.common.DBRecord.java

/**
 * Writable class for DB Source/Sink
 *
 * @see org.apache.hadoop.mapreduce.lib.db.DBInputFormat DBInputFormat
 * @see org.apache.hadoop.mapreduce.lib.db.DBOutputFormat DBOutputFormat
 * @see org.apache.hadoop.mapreduce.lib.db.DBWritable DBWritable

From source file co.cask.hydrator.plugin.DBRecord.java

/**
 * Writable class for DB Source/Sink
 *
 * @see org.apache.hadoop.mapreduce.lib.db.DBInputFormat DBInputFormat
 * @see org.apache.hadoop.mapreduce.lib.db.DBOutputFormat DBOutputFormat
 * @see DBWritable DBWritable

From source file com.cloudera.sqoop.lib.SqoopRecord.java

/**
 * Interface implemented by the classes generated by sqoop's orm.ClassWriter.
 */
public abstract class SqoopRecord implements Cloneable, DBWritable, FieldMappable, Writable {

    public SqoopRecord() {

From source file com.jbw.recommendsystem.storetodb.UGPBean.java

/**
 *
 * @author alvin
 */
public class UGPBean implements DBWritable, WritableComparable<UGPBean> {

From source file hadoop.DBInputWritable.java

public class DBInputWritable implements Writable, DBWritable {
    private int sqlField;

    public void readFields(DataInput in) throws IOException {
    }

From source file hadoop.DBOutputWritable.java

public class DBOutputWritable implements Writable, DBWritable {
    private int sqlField;

    public DBOutputWritable(int sqlField) {
        this.sqlField = sqlField;
    }

From source file nl.sanoma.hdt.report.generator.DBInputWritable.java

/**
 * Class to help read the values from database.
 *
 * @author Richrd Ern Kiss
 */
public class DBInputWritable implements Writable, DBWritable {

From source file org.apache.beam.sdk.io.hadoop.format.TestRowDBWritable.java

/**
 * A subclass of {@link TestRow} to be used with {@link
 * org.apache.hadoop.mapreduce.lib.db.DBInputFormat}.
 */
@DefaultCoder(AvroCoder.class)
class TestRowDBWritable extends TestRow implements DBWritable, Writable {

From source file org.apache.beam.sdk.io.hadoop.inputformat.TestRowDBWritable.java

/**
 * A subclass of {@link org.apache.beam.sdk.io.common.TestRow} to be used with {@link
 * org.apache.hadoop.mapreduce.lib.db.DBInputFormat}.
 */
@DefaultCoder(AvroCoder.class)
class TestRowDBWritable extends TestRow implements DBWritable, Writable {

From source file org.apache.beam.sdk.io.hadoop.inputformat.unit.tests.inputs.DBInputWritable.java

/**
 * To access the data from RDBMS using {@link DBInputFormat}, you have to create a class to define
 * the data which you are going to read. {@link DBInputWritable} is a class to read data from
 * Postgres DB in test {@link HIFIOWithPostgresIT}. {@link DBInputWritable} holds id and name of the
 * scientist.
 */