List of usage examples for org.apache.hadoop.mapreduce JobContext getOutputFormatClass
public Class<? extends OutputFormat<?, ?>> getOutputFormatClass() throws ClassNotFoundException;
From source file:com.scaleoutsoftware.soss.hserver.hadoop.ReducerWrapperMapreduce.java
License:Apache License
public ReducerWrapperMapreduce(HServerInvocationParameters invocationParameters, int hadoopPartition, int appId, int region, boolean sort) throws IOException, ClassNotFoundException, InterruptedException { this.invocationParameters = invocationParameters; Configuration configuration = (Configuration) invocationParameters.getConfiguration(); hadoopVersionSpecificCode = HadoopVersionSpecificCode.getInstance(invocationParameters.getHadoopVersion(), configuration);/* ww w .j a v a2 s . c om*/ JobID jobID = (JobID) invocationParameters.getJobId(); //Setup task ID info TaskAttemptID id = hadoopVersionSpecificCode.createTaskAttemptId(jobID, false, hadoopPartition); JobContext jobContext = hadoopVersionSpecificCode.createJobContext(new JobConf(configuration), jobID); taskContext = hadoopVersionSpecificCode.createTaskAttemptContext(configuration, id); reducer = (org.apache.hadoop.mapreduce.Reducer<INKEY, INVALUE, OUTKEY, OUTVALUE>) ReflectionUtils .newInstance(jobContext.getReducerClass(), configuration); OutputFormat outputFormat = ReflectionUtils.newInstance(jobContext.getOutputFormatClass(), configuration); recordWriter = (org.apache.hadoop.mapreduce.RecordWriter<OUTKEY, OUTVALUE>) outputFormat .getRecordWriter(taskContext); committer = outputFormat.getOutputCommitter(taskContext); committer.setupTask(taskContext); Class<INKEY> keyClass = (Class<INKEY>) jobContext.getMapOutputKeyClass(); WritableSerializerDeserializer<INKEY> firstKeySerializer = new WritableSerializerDeserializer<INKEY>( keyClass, null); WritableSerializerDeserializer<INKEY> secondKeySerializer = new WritableSerializerDeserializer<INKEY>( keyClass, null); Class<INVALUE> valueClass = (Class<INVALUE>) jobContext.getMapOutputValueClass(); WritableSerializerDeserializer<INVALUE> valueSerializer = new WritableSerializerDeserializer<INVALUE>( valueClass, null); DataGridReaderParameters<INKEY, INVALUE> params = new DataGridReaderParameters<INKEY, INVALUE>(region, appId, HServerParameters.getSetting(REDUCE_USEMEMORYMAPPEDFILES, configuration) > 0, firstKeySerializer, valueSerializer, invocationParameters.getSerializationMode(), secondKeySerializer, keyClass, valueClass, sort, HServerParameters.getSetting(REDUCE_CHUNKSTOREADAHEAD, configuration), 1024 * HServerParameters.getSetting(REDUCE_INPUTCHUNKSIZE_KB, configuration), HServerParameters.getSetting(REDUCE_CHUNKREADTIMEOUT, configuration)); DataGridChunkedCollectionReader<INKEY, INVALUE> transport = DataGridChunkedCollectionReader .getGridReader(params); context = hadoopVersionSpecificCode.getReducerContext(configuration, id, committer, recordWriter, transport, null); }
From source file:org.apache.hcatalog.mapreduce.MultiOutputFormat.java
License:Apache License
private static OutputFormat<?, ?> getOutputFormatInstance(JobContext context) { OutputFormat<?, ?> outputFormat; try {// w ww. j a va2s .c o m outputFormat = ReflectionUtils.newInstance(context.getOutputFormatClass(), context.getConfiguration()); } catch (ClassNotFoundException e) { throw new IllegalStateException(e); } return outputFormat; }
From source file:org.apache.ignite.internal.processors.hadoop.impl.v2.HadoopV2Task.java
License:Apache License
/** * Create and configure an OutputFormat instance. * * @param jobCtx Job context./*from www . j a v a 2 s. co m*/ * @return Instance of OutputFormat is specified in job configuration. * @throws ClassNotFoundException If specified class not found. */ protected OutputFormat getOutputFormat(JobContext jobCtx) throws ClassNotFoundException { return ReflectionUtils.newInstance(jobCtx.getOutputFormatClass(), hadoopContext().getConfiguration()); }