当前位置: 首页>>代码示例>>Java>>正文


Java HFileOutputFormat.configureIncrementalLoad方法代码示例

本文整理汇总了Java中org.apache.hadoop.hbase.mapreduce.HFileOutputFormat.configureIncrementalLoad方法的典型用法代码示例。如果您正苦于以下问题:Java HFileOutputFormat.configureIncrementalLoad方法的具体用法?Java HFileOutputFormat.configureIncrementalLoad怎么用?Java HFileOutputFormat.configureIncrementalLoad使用的例子?那么, 这里精选的方法代码示例或许可以为您提供帮助。您也可以进一步了解该方法所在org.apache.hadoop.hbase.mapreduce.HFileOutputFormat的用法示例。


在下文中一共展示了HFileOutputFormat.configureIncrementalLoad方法的15个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Java代码示例。

示例1: jobSetup

import org.apache.hadoop.hbase.mapreduce.HFileOutputFormat; //导入方法依赖的package包/类
@Override
protected void jobSetup(Job job) throws IOException, ImportException {
  super.jobSetup(job);

  // we shouldn't have gotten here if bulk load dir is not set
  // so let's throw a ImportException
  if(getContext().getDestination() == null){
    throw new ImportException("Can't run HBaseBulkImportJob without a " +
        "valid destination directory.");
  }

  TableMapReduceUtil.addDependencyJars(job.getConfiguration(), Preconditions.class);
  FileOutputFormat.setOutputPath(job, getContext().getDestination());
  HTable hTable = new HTable(job.getConfiguration(), options.getHBaseTable());
  HFileOutputFormat.configureIncrementalLoad(job, hTable);
}
 
开发者ID:aliyun,项目名称:aliyun-maxcompute-data-collectors,代码行数:17,代码来源:HBaseBulkImportJob.java

示例2: main

import org.apache.hadoop.hbase.mapreduce.HFileOutputFormat; //导入方法依赖的package包/类
public static void main(String[] args) throws Exception {
   Configuration conf = new Configuration();
   conf.set("hbase.table.name", args[2]);    
   
   Job job = new Job(conf, "createipas");
   
   job.setJarByClass(CreateIpAS.class);
   job.setOutputKeyClass(ImmutableBytesWritable.class);
   job.setOutputValueClass(KeyValue.class);
   
   job.setMapperClass(Map.class);
   job.setReducerClass(Reducer.class);
       
   job.setInputFormatClass(TextInputFormat.class);
   job.setOutputFormatClass(HFileOutputFormat.class);
   
   HTable hTable = new HTable(conf, args[2]);
   HFileOutputFormat.configureIncrementalLoad(job, hTable);
   
   FileInputFormat.addInputPath(job, new Path(args[0]));
   FileOutputFormat.setOutputPath(job, new Path(args[1]));
       
   job.waitForCompletion(true);
}
 
开发者ID:dsarlis,项目名称:datix,代码行数:25,代码来源:CreateIpAS.java

示例3: startBulkLoad

import org.apache.hadoop.hbase.mapreduce.HFileOutputFormat; //导入方法依赖的package包/类
@SuppressWarnings("rawtypes")
private Job startBulkLoad(Configuration conf, String inputTable,
		String tableName, Class<? extends TableMapper> clazz, Path outputDir)
		throws Exception {

	// Create our job to bulk load into HBase
	Job job = Job.getInstance(conf, "HBase Bulk Loader");
	job.setJarByClass(getClass());

	// Initialize our mapper by specifying the input table
	TableMapReduceUtil.initTableMapperJob(inputTable, new Scan(), clazz,
			ImmutableBytesWritable.class, KeyValue.class, job);

	HFileOutputFormat.configureIncrementalLoad(job, new HTable(conf,
			tableName));
	HFileOutputFormat.setOutputPath(job, outputDir);

	// launch the job
	job.waitForCompletion(true);
	return job;
}
 
开发者ID:Pivotal-Field-Engineering,项目名称:pmr-common,代码行数:22,代码来源:TwitterTableSplit.java

示例4: main

import org.apache.hadoop.hbase.mapreduce.HFileOutputFormat; //导入方法依赖的package包/类
public static void main(String[] args) throws Exception {
    Configuration conf = new Configuration();
    conf.set("hbase.table.name", TABLE_NAME);

    Job job = new Job(conf);

    job.setJarByClass(Q2Loader.class);
    /* set mapper and reducer keys and values */
    job.setOutputKeyClass(ImmutableBytesWritable.class);
    job.setOutputValueClass(KeyValue.class);
    job.setMapOutputKeyClass(Text.class);
    job.setMapOutputValueClass(Text.class);

    job.setMapperClass(Map.class);
    job.setReducerClass(Reduce.class);

    job.setInputFormatClass(TextInputFormat.class);
    job.setOutputFormatClass(HFileOutputFormat.class);

    /* set the output of the job to be in HFile format */
    HTable hTable = new HTable(conf, TABLE_NAME);
    HFileOutputFormat.configureIncrementalLoad(job, hTable);

    FileInputFormat.addInputPath(job, new Path(args[0]));
    FileOutputFormat.setOutputPath(job, new Path(args[1]));

    job.waitForCompletion(true);

}
 
开发者ID:dsarlis,项目名称:Cloud-Burst,代码行数:30,代码来源:Q2Loader.java

示例5: main

import org.apache.hadoop.hbase.mapreduce.HFileOutputFormat; //导入方法依赖的package包/类
public static void main(String[] args) throws Exception {
    Configuration conf = new Configuration();
    conf.set("hbase.table.name", TABLE_NAME);

    Job job = new Job(conf);

    job.setJarByClass(Q3Loader.class);
    // set mapper and reducer keys and values
    job.setOutputKeyClass(ImmutableBytesWritable.class);
    job.setOutputValueClass(KeyValue.class);
    job.setMapOutputKeyClass(Text.class);
    job.setMapOutputValueClass(Text.class);

    job.setMapperClass(Map.class);
    job.setReducerClass(Reduce.class);

    job.setInputFormatClass(TextInputFormat.class);
    job.setOutputFormatClass(HFileOutputFormat.class);

    // set the output of the job to be in HFile format
    HTable hTable = new HTable(conf, TABLE_NAME);
    HFileOutputFormat.configureIncrementalLoad(job, hTable);

    FileInputFormat.addInputPath(job, new Path(args[0]));
    FileOutputFormat.setOutputPath(job, new Path(args[1]));

    job.waitForCompletion(true);

}
 
开发者ID:dsarlis,项目名称:Cloud-Burst,代码行数:30,代码来源:Q3Loader.java

示例6: main

import org.apache.hadoop.hbase.mapreduce.HFileOutputFormat; //导入方法依赖的package包/类
public static void main(String[] args) throws Exception {
    Configuration conf = new Configuration();
    conf.set("hbase.table.name", TABLE_NAME);

    Job job = new Job(conf);

    job.setJarByClass(Q4Loader.class);
    /* set mapper and reducer keys and values */
    job.setOutputKeyClass(ImmutableBytesWritable.class);
    job.setOutputValueClass(KeyValue.class);
    job.setMapOutputKeyClass(Text.class);
    job.setMapOutputValueClass(Text.class);

    job.setMapperClass(Map.class);
    job.setReducerClass(Reduce.class);

    job.setInputFormatClass(TextInputFormat.class);
    job.setOutputFormatClass(HFileOutputFormat.class);

    /* set the output of the job to be in HFile format */
    HTable hTable = new HTable(conf, TABLE_NAME);
    HFileOutputFormat.configureIncrementalLoad(job, hTable);

    FileInputFormat.addInputPath(job, new Path(args[0]));
    FileOutputFormat.setOutputPath(job, new Path(args[1]));

    job.waitForCompletion(true);
}
 
开发者ID:dsarlis,项目名称:Cloud-Burst,代码行数:29,代码来源:Q4Loader.java

示例7: main

import org.apache.hadoop.hbase.mapreduce.HFileOutputFormat; //导入方法依赖的package包/类
public static void main(String[] args) throws Exception {
        JobContext conf = (JobContext) new Configuration();
//        conf.set("hbase.table.name", TABLE_NAME);

        Job job = new Job(conf);

        job.setJarByClass(HBaseLoader.class);
        job.setOutputKeyClass(ImmutableBytesWritable.class);
        job.setOutputValueClass(KeyValue.class);
        job.setMapOutputKeyClass(Text.class);
        job.setMapOutputValueClass(Text.class);

        job.setMapperClass(Map.class);
        job.setReducerClass(Reduce.class);

        job.setInputFormatClass(TextInputFormat.class);
        job.setOutputFormatClass(HFileOutputFormat.class);

        HTable hTable = new HTable(conf, TABLE_NAME);
        HFileOutputFormat.configureIncrementalLoad(job, hTable);

        FileInputFormat.addInputPath(job, new Path(args[0]));
        FileOutputFormat.setOutputPath(job, new Path(args[1]));

        job.waitForCompletion(true);

    }
 
开发者ID:dsarlis,项目名称:Cloud-Burst,代码行数:28,代码来源:HBaseLoader.java

示例8: run

import org.apache.hadoop.hbase.mapreduce.HFileOutputFormat; //导入方法依赖的package包/类
public int run(String[] args) throws Exception {
    Options options = new Options();

    try {
        options.addOption(OPTION_JOB_NAME);
        options.addOption(OPTION_II_NAME);
        options.addOption(OPTION_INPUT_PATH);
        options.addOption(OPTION_OUTPUT_PATH);
        options.addOption(OPTION_HTABLE_NAME);
        parseOptions(options, args);

        Path output = new Path(getOptionValue(OPTION_OUTPUT_PATH));

        job = Job.getInstance(getConf(), getOptionValue(OPTION_JOB_NAME));

        setJobClasspath(job);

        addInputDirs(getOptionValue(OPTION_INPUT_PATH), job);
        FileOutputFormat.setOutputPath(job, output);

        job.setInputFormatClass(SequenceFileInputFormat.class);
        job.setMapperClass(IICreateHFileMapper.class);
        job.setMapOutputKeyClass(ImmutableBytesWritable.class);
        job.setMapOutputValueClass(KeyValue.class);

        String tableName = getOptionValue(OPTION_HTABLE_NAME);
        HTable htable = new HTable(HBaseConfiguration.create(getConf()), tableName);
        HFileOutputFormat.configureIncrementalLoad(job, htable);

        this.deletePath(job.getConfiguration(), output);

        return waitForCompletion(job);
    } catch (Exception e) {
        printUsage(options);
        throw e;
    }
}
 
开发者ID:KylinOLAP,项目名称:Kylin,代码行数:38,代码来源:IICreateHFileJob.java

示例9: preJobLaunch

import org.apache.hadoop.hbase.mapreduce.HFileOutputFormat; //导入方法依赖的package包/类
@Override
protected void preJobLaunch(CommandLine cmd, Job job) throws Exception {
	job.setJobName("Twitter HBase Bulk Load");
	htable = new HTable(getConf(), cmd.getOptionValue(HTABLE_OPT));

	HFileOutputFormat.configureIncrementalLoad(job, htable);
	HFileOutputFormat.setOutputPath(job, outputDir);
}
 
开发者ID:Pivotal-Field-Engineering,项目名称:pmr-common,代码行数:9,代码来源:TwitterBulkLoad.java

示例10: createSubmittableJob

import org.apache.hadoop.hbase.mapreduce.HFileOutputFormat; //导入方法依赖的package包/类
/**
 * Sets up the actual job.
 * 
 * @param conf
 *            The current configuration.
 * @param args
 *            The command line parameters.
 * @return The newly created job.
 * @throws IOException
 *             When setting up the job fails.
 */
public static Job createSubmittableJob(Configuration conf, String[] args)
		throws IOException {

	String tableName = args[0];
	Path inputDir = new Path(args[1]);
	Job job = new Job(conf, "HBaseToHFileMapReduce");
	job.setJarByClass(HBaseToHFileMapReduce.class);
	FileInputFormat.setInputPaths(job, inputDir);
	job.setInputFormatClass(TextInputFormat.class);
	job.setMapperClass(HourlyImporter.class);

	if (args.length < 3) {
		// ++++ insert into table directly using TableOutputFormat ++++
		TableMapReduceUtil.initTableReducerJob(tableName, null, job);
		job.setNumReduceTasks(0);
	} else {
		// ++++ to generate HFile instead ++++
		HTable table = new HTable(conf, tableName);
		job.setReducerClass(PutSortReducer.class);
		Path outputDir = new Path(args[2]);
		FileOutputFormat.setOutputPath(job, outputDir);
		job.setMapOutputKeyClass(ImmutableBytesWritable.class);
		job.setMapOutputValueClass(Put.class);
		HFileOutputFormat.configureIncrementalLoad(job, table);			
	}		
	
	TableMapReduceUtil.addDependencyJars(job);
	return job;
}
 
开发者ID:willddy,项目名称:bigdata_pattern,代码行数:41,代码来源:HBaseToHFileMapReduce.java

示例11: configureIncrementalLoad

import org.apache.hadoop.hbase.mapreduce.HFileOutputFormat; //导入方法依赖的package包/类
public static void configureIncrementalLoad(Job job, HTable table) throws IOException {
  HFileOutputFormat.configureIncrementalLoad(job, table);
  // Override OutputFormatClass
  job.setOutputFormatClass(IndexHFileOutputFormat.class);
}
 
开发者ID:tenggyut,项目名称:HIndex,代码行数:6,代码来源:IndexHFileOutputFormat.java

示例12: createSubmittableJob

import org.apache.hadoop.hbase.mapreduce.HFileOutputFormat; //导入方法依赖的package包/类
/**
 * Sets up the actual job.
 * 
 * @param conf
 *            The current configuration.
 * @param args
 *            The command line parameters.
 * @return The newly created job.
 * @throws IOException
 *             When setting up the job fails.
 */
public static Job createSubmittableJob(Configuration conf, String[] args)
		throws IOException, ClassNotFoundException {

	// Support non-XML supported characters
	// by re-encoding the passed separator as a Base64 string.
	String actualSeparator = conf.get(SEPARATOR_CONF_KEY);
	if (actualSeparator != null) {
		conf.set(SEPARATOR_CONF_KEY,
				new String(Base64.encodeBytes(actualSeparator.getBytes())));
	}

	// See if a non-default Mapper was set
	String mapperClassName = conf.get(MAPPER_CONF_KEY);
	Class mapperClass = mapperClassName != null ? Class
			.forName(mapperClassName) : DEFAULT_MAPPER;

	String tableName = args[0];
	Path inputDir = new Path(args[1]);
	Job job = new Job(conf, NAME + "_" + tableName);
	job.setJarByClass(mapperClass);
	FileInputFormat.setInputPaths(job, inputDir);

	String inputCodec = conf.get(INPUT_LZO_KEY);
	if (inputCodec == null) {
		FileInputFormat.setMaxInputSplitSize(job, 67108864l); // max split
																// size =
																// 64m
		job.setInputFormatClass(TextInputFormat.class);
	} else {
		if (inputCodec.equalsIgnoreCase("lzo"))
			job.setInputFormatClass(LzoTextInputFormat.class);
		else {
			usage("not supported compression codec!");
			System.exit(-1);
		}
	}

	job.setMapperClass(mapperClass);

	String hfileOutPath = conf.get(BULK_OUTPUT_CONF_KEY);
	if (hfileOutPath != null) {
		HTable table = new HTable(conf, tableName);
		job.setReducerClass(PutSortReducer.class);
		Path outputDir = new Path(hfileOutPath);
		FileOutputFormat.setOutputPath(job, outputDir);
		job.setMapOutputKeyClass(ImmutableBytesWritable.class);
		job.setMapOutputValueClass(Put.class);
		HFileOutputFormat.configureIncrementalLoad(job, table);
	} else {
		// No reducers. Just write straight to table. Call
		// initTableReducerJob
		// to set up the TableOutputFormat.
		TableMapReduceUtil.initTableReducerJob(tableName, null, job);
		job.setNumReduceTasks(0);
	}

	TableMapReduceUtil.addDependencyJars(job);
	TableMapReduceUtil.addDependencyJars(job.getConfiguration(),
			com.google.common.base.Function.class /*
												 * Guava used by TsvParser
												 */);
	return job;
}
 
开发者ID:javachen,项目名称:learning-hadoop,代码行数:75,代码来源:ImportTsv.java

示例13: run

import org.apache.hadoop.hbase.mapreduce.HFileOutputFormat; //导入方法依赖的package包/类
public int run(String[] args) throws Exception {
    Options options = new Options();

    try {
        options.addOption(OPTION_JOB_NAME);
        options.addOption(OPTION_CUBE_NAME);
        options.addOption(OPTION_INPUT_PATH);
        options.addOption(OPTION_OUTPUT_PATH);
        options.addOption(OPTION_HTABLE_NAME);
        parseOptions(options, args);

        Path output = new Path(getOptionValue(OPTION_OUTPUT_PATH));
        String cubeName = getOptionValue(OPTION_CUBE_NAME).toUpperCase();

        CubeManager cubeMgr = CubeManager.getInstance(KylinConfig.getInstanceFromEnv());

        CubeInstance cube = cubeMgr.getCube(cubeName);
        job = Job.getInstance(getConf(), getOptionValue(OPTION_JOB_NAME));

        setJobClasspath(job);

        addInputDirs(getOptionValue(OPTION_INPUT_PATH), job);
        FileOutputFormat.setOutputPath(job, output);

        job.setInputFormatClass(SequenceFileInputFormat.class);
        job.setMapperClass(CubeHFileMapper.class);
        job.setReducerClass(KeyValueSortReducer.class);

        // set job configuration
        job.getConfiguration().set(BatchConstants.CFG_CUBE_NAME, cubeName);
        Configuration conf = HBaseConfiguration.create(getConf());
        // add metadata to distributed cache
        attachKylinPropsAndMetadata(cube, job.getConfiguration());

        String tableName = getOptionValue(OPTION_HTABLE_NAME).toUpperCase();
        HTable htable = new HTable(conf, tableName);

        //Automatic config !
        HFileOutputFormat.configureIncrementalLoad(job, htable);

        // set block replication to 3 for hfiles
        conf.set(DFSConfigKeys.DFS_REPLICATION_KEY, "3");

        this.deletePath(job.getConfiguration(), output);

        return waitForCompletion(job);
    } catch (Exception e) {
        logger.error("error in CubeHFileJob", e);
        printUsage(options);
        throw e;
    }
}
 
开发者ID:KylinOLAP,项目名称:Kylin,代码行数:53,代码来源:CubeHFileJob.java

示例14: bulkLoadNewMacAddresses

import org.apache.hadoop.hbase.mapreduce.HFileOutputFormat; //导入方法依赖的package包/类
private static void bulkLoadNewMacAddresses(
        Configuration conf, String inputPath, String outputPath, String tblName)
        throws Exception {

    // Pass parameters to Mad Reduce
    conf.set("hbase.table.name", tblName);
    conf.set("macs", macAddressesLine);

    // Workaround
    SchemaMetrics.configureGlobally(conf);

    // Load hbase-site.xml
    HBaseConfiguration.addHbaseResources(conf);

    // Create the job
    Job job = new Job(conf, "Load macAddresses in bloomfilters table");

    job.setJarByClass(MapperBulkLoadMacAddresses.class);
    job.setMapperClass(MapperBulkLoadMacAddresses.class);
    job.setMapOutputKeyClass(ImmutableBytesWritable.class);
    job.setMapOutputValueClass(KeyValue.class);

    job.setInputFormatClass(TextInputFormat.class);

    // Get the table
    HTable hTable = new HTable(conf, tblName);

    // Auto configure partitioner and reducer
    HFileOutputFormat.configureIncrementalLoad(job, hTable);

    // Save output path and input path
    FileInputFormat.addInputPath(job, new Path(inputPath));
    FileOutputFormat.setOutputPath(job, new Path(outputPath));

    // Wait for HFiles creations
    job.waitForCompletion(true);

    // Load generated HFiles into table
    LoadIncrementalHFiles loader = new LoadIncrementalHFiles(conf);
    loader.doBulkLoad(new Path(outputPath), hTable);
}
 
开发者ID:dmsl,项目名称:tvm,代码行数:42,代码来源:RadiomapLoader.java

示例15: bulkLoadVectormap

import org.apache.hadoop.hbase.mapreduce.HFileOutputFormat; //导入方法依赖的package包/类
/**
 * Load radiomap in HBase
 *
 * @param conf
 * @throws Exception
 */
private static void bulkLoadVectormap(
        Configuration conf, String inputPath, String outputPath, String tblName)
        throws Exception {

    // Pass parameters to Mad Reduce
    conf.set("hbase.table.name", tblName);
    conf.set("macs", macAddressesLine);

    // Workaround
    SchemaMetrics.configureGlobally(conf);

    // Load hbase-site.xml
    HBaseConfiguration.addHbaseResources(conf);

    // Create the job
    Job job = new Job(conf, "Load radiomap in HBase");

    job.setJarByClass(MapperBulkLoadRadiomap.class);
    job.setMapperClass(MapperBulkLoadRadiomap.class);
    job.setMapOutputKeyClass(ImmutableBytesWritable.class);
    job.setMapOutputValueClass(KeyValue.class);

    job.setInputFormatClass(TextInputFormat.class);

    // Get the table
    HTable hTable = new HTable(conf, tblName);

    // Auto configure partitioner and reducer
    HFileOutputFormat.configureIncrementalLoad(job, hTable);

    // Save output path and input path
    FileInputFormat.addInputPath(job, new Path(inputPath));
    FileOutputFormat.setOutputPath(job, new Path(outputPath));

    // Wait for HFiles creations
    job.waitForCompletion(true);

    // Load generated HFiles into table
    LoadIncrementalHFiles loader = new LoadIncrementalHFiles(conf);
    loader.doBulkLoad(new Path(outputPath), hTable);
}
 
开发者ID:dmsl,项目名称:tvm,代码行数:48,代码来源:RadiomapLoader.java


注:本文中的org.apache.hadoop.hbase.mapreduce.HFileOutputFormat.configureIncrementalLoad方法示例由纯净天空整理自Github/MSDocs等开源代码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。