List of usage examples for org.apache.hadoop.io Writable interface-usage
From source file com.rockstor.compact.mapreduce.CompactDirInputSplit.java
/** * @author terry * */ public class CompactDirInputSplit extends InputSplit implements Writable { private String taskIdName = null;
From source file com.rw.legion.LegionRecord.java
/** * At its core, this is nothing more than a hash map, which links data keys * (e.g., CSV column headers) to the values associated with those keys (e.g., * data in a CSV row). */ public class LegionRecord implements Writable {
From source file com.scaleoutsoftware.soss.hserver.BucketSplit.java
/** * An input split that corresponds to the contents of the one or several buckets. */ class BucketSplit extends InputSplit implements Writable, Serializable, GridSplit { private static final long serialVersionUID = 1L;
From source file com.scaleoutsoftware.soss.hserver.ImageInputSplit.java
/**
* This input split is part of the {@link GridImage}. It is wrapper over original input split containing
* additional information which allow to read it from the bucket store.
*
* {@link GridImage}
*/
From source file com.scaleoutsoftware.soss.hserver.WritableBucketId.java
/** * This class wraps a BucketId to implement Writable (Hadoop specific code) */ public class WritableBucketId extends BucketId implements Serializable, Writable { WritableBucketId() {
From source file com.scaleunlimited.classify.BaseModel.java
/**
* Classification model, which can be trained using a set of pre-labeled
* T documents, and which can then be used to classify
* a set of unlabeled T documents, outputting the
* classification of each as a {@link DocDatum} document.
*/
From source file com.sematext.hbase.hut.UpdateProcessor.java
/** * Performs records processing. * Implementation SHOULD be stateless and thread-safe */ public abstract class UpdateProcessor implements Writable { public abstract void process(Iterable<Result> records, UpdateProcessingResult processingResult);
From source file com.sensei.indexing.hadoop.keyvalueformat.IntermediateForm.java
/**
* An intermediate form for one or more parsed Lucene documents and/or
* delete terms. It actually uses Lucene file format as the format for
* the intermediate form by using RAM dir files.
*
* Note: If process(*) is ever called, closeWriter() should be called.
From source file com.senseidb.indexing.hadoop.keyvalueformat.IntermediateForm.java
/**
* An intermediate form for one or more parsed Lucene documents and/or
* delete terms. It actually uses Lucene file format as the format for
* the intermediate form by using RAM dir files.
*
* Note: If process(*) is ever called, closeWriter() should be called.
From source file com.soteradefense.dga.hbse.HighBetweennessList.java
/** * Maintains a list of the top N items(an item is defined as an int id, and double value), ranked by a double value. * Designed for use with the giraph Aggregator pattern. */ public class HighBetweennessList implements Writable {