List of usage examples for org.apache.hadoop.mapreduce.lib.db DBConfiguration INPUT_CLASS_PROPERTY
String INPUT_CLASS_PROPERTY
To view the source code for org.apache.hadoop.mapreduce.lib.db DBConfiguration INPUT_CLASS_PROPERTY.
Click Source Link
From source file:org.apache.beam.sdk.io.hadoop.format.HadoopFormatIOIT.java
License:Apache License
private static void setupHadoopConfiguration(PostgresIOTestPipelineOptions options) { Configuration conf = new Configuration(); DBConfiguration.configureDB(conf, "org.postgresql.Driver", DatabaseTestHelper.getPostgresDBUrl(options), options.getPostgresUsername(), options.getPostgresPassword()); conf.set(DBConfiguration.INPUT_TABLE_NAME_PROPERTY, tableName); conf.setStrings(DBConfiguration.INPUT_FIELD_NAMES_PROPERTY, "id", "name"); conf.set(DBConfiguration.INPUT_ORDER_BY_PROPERTY, "id ASC"); conf.setClass(DBConfiguration.INPUT_CLASS_PROPERTY, TestRowDBWritable.class, DBWritable.class); conf.setClass("key.class", LongWritable.class, Object.class); conf.setClass("value.class", TestRowDBWritable.class, Object.class); conf.setClass("mapreduce.job.inputformat.class", DBInputFormat.class, InputFormat.class); conf.set(DBConfiguration.OUTPUT_TABLE_NAME_PROPERTY, tableName); conf.set(DBConfiguration.OUTPUT_FIELD_COUNT_PROPERTY, "2"); conf.setStrings(DBConfiguration.OUTPUT_FIELD_NAMES_PROPERTY, "id", "name"); conf.setClass(HadoopFormatIO.OUTPUT_KEY_CLASS, TestRowDBWritable.class, Object.class); conf.setClass(HadoopFormatIO.OUTPUT_VALUE_CLASS, NullWritable.class, Object.class); conf.setClass(HadoopFormatIO.OUTPUT_FORMAT_CLASS_ATTR, DBOutputFormat.class, OutputFormat.class); conf.set(HadoopFormatIO.JOB_ID, String.valueOf(1)); hadoopConfiguration = new SerializableConfiguration(conf); }
From source file:org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIOIT.java
License:Apache License
private static void setupHadoopConfiguration(PostgresIOTestPipelineOptions options) { Configuration conf = new Configuration(); DBConfiguration.configureDB(conf, "org.postgresql.Driver", DatabaseTestHelper.getPostgresDBUrl(options), options.getPostgresUsername(), options.getPostgresPassword()); conf.set(DBConfiguration.INPUT_TABLE_NAME_PROPERTY, tableName); conf.setStrings(DBConfiguration.INPUT_FIELD_NAMES_PROPERTY, "id", "name"); conf.set(DBConfiguration.INPUT_ORDER_BY_PROPERTY, "id ASC"); conf.setClass(DBConfiguration.INPUT_CLASS_PROPERTY, TestRowDBWritable.class, DBWritable.class); conf.setClass("key.class", LongWritable.class, Object.class); conf.setClass("value.class", TestRowDBWritable.class, Object.class); conf.setClass("mapreduce.job.inputformat.class", DBInputFormat.class, InputFormat.class); hadoopConfiguration = new SerializableConfiguration(conf); }
From source file:org.apache.crunch.contrib.io.jdbc.DataBaseSource.java
License:Apache License
private DataBaseSource(Class<T> inputClass, String driverClassName, String url, String username, String password, String selectClause, String countClause) { super(new Path("dbsource"), Writables.writables(inputClass), FormatBundle.forInput(DBInputFormat.class) .set(DBConfiguration.DRIVER_CLASS_PROPERTY, driverClassName).set(DBConfiguration.URL_PROPERTY, url) .set(DBConfiguration.USERNAME_PROPERTY, username).set(DBConfiguration.PASSWORD_PROPERTY, password) .set(DBConfiguration.INPUT_CLASS_PROPERTY, inputClass.getCanonicalName()) .set(DBConfiguration.INPUT_QUERY, selectClause) .set(DBConfiguration.INPUT_COUNT_QUERY, countClause)); }