List of usage examples for org.apache.hadoop.mapreduce.lib.output FileOutputCommitter subclass-usage
From source file alluxio.hadoop.mapreduce.KeyValueOutputCommitter.java
/**
* Extension of {@link FileOutputCommitter} where creating, completing, or deleting a {@link
* KeyValueSystem} in different phases of a job's or task's lifecycle is considered.
* <p>
* This committer must be used along with {@link KeyValueOutputFormat} to merge the key-value stores
* created by each Reducer into one key-value store under the MapReduce output directory.
From source file boa.io.BoaOutputCommitter.java
/** * A {@link FileOutputCommitter} that stores the job results into a database. * * @author rdyer */ public class BoaOutputCommitter extends FileOutputCommitter {
From source file co.cask.cdap.internal.app.runtime.batch.dataset.partitioned.DynamicPartitioningOutputCommitter.java
/**
* An OutputCommitter which creates partitions in a configured PartitionedFileSet dataset for all of the partitions
* that were written to by a DynamicPartitioningOutputFormat
* It enables this by having each job write to a job-specific temporary path within that output directory.
* Then, upon commitJob, it moves the files to the final, parent directory if the final output directories do not
* already exist.
From source file com.github.sadikovi.hadoop.riff.RiffOutputCommitter.java
/**
* Riff output committer.
* Provides methods to save metadata for the output folder.
*
* Note: Spark SQL does not use custom output committer when appending data to the existing table,
* in our case it is okay, since we store only static information that does not change with writing
From source file com.inmobi.conduit.distcp.tools.mapred.CopyCommitter.java
public class CopyCommitter extends FileOutputCommitter { private static final Log LOG = LogFactory.getLog(CopyCommitter.class); private final TaskAttemptContext taskAttemptContext; /**
From source file com.practicalHadoop.outputformat.MultpleDirectories.FileOutputCommitter.java
public class FileOutputCommitter extends org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter { private static final Log LOG = LogFactory.getLog(FileOutputCommitter.class); private static Map<String, FileOutputCommitter> committers = new HashMap<String, FileOutputCommitter>(); /**
From source file gobblin.compaction.mapreduce.avro.AvroKeyCompactorOutputCommitter.java
/**
* Class used with {@link MRCompactorAvroKeyDedupJobRunner} to rename files as they
* are being committed. In addition to moving files from their working directory to
* the commit output directory, the files are named to include a timestamp and a
* count of how many records the file contains, in the format
* {recordCount}.{timestamp}.avro.
From source file org.apache.carbondata.hadoop.api.CarbonOutputCommitter.java
/** * Outputcommitter which manages the segments during loading.It commits segment information to the * tablestatus file upon success or fail. */ public class CarbonOutputCommitter extends FileOutputCommitter {
From source file org.apache.falcon.hive.mapreduce.CopyCommitter.java
/** * Copy committer class. */ public class CopyCommitter extends FileOutputCommitter { private static final Logger LOG = LoggerFactory.getLogger(CopyCommitter.class);
From source file org.apache.gobblin.compaction.mapreduce.avro.AvroKeyCompactorOutputCommitter.java
/**
* Class used with {@link MRCompactorAvroKeyDedupJobRunner} to rename files as they
* are being committed. In addition to moving files from their working directory to
* the commit output directory, the files are named to include a timestamp and a
* count of how many records the file contains, in the format
* {recordCount}.{timestamp}.avro.