当前位置: 首页>>代码示例>>Java>>正文


Java LineRecordReader类代码示例

本文整理汇总了Java中org.apache.hadoop.mapreduce.lib.input.LineRecordReader的典型用法代码示例。如果您正苦于以下问题:Java LineRecordReader类的具体用法?Java LineRecordReader怎么用?Java LineRecordReader使用的例子?那么, 这里精选的类代码示例或许可以为您提供帮助。


LineRecordReader类属于org.apache.hadoop.mapreduce.lib.input包,在下文中一共展示了LineRecordReader类的15个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Java代码示例。

示例1: initialize

import org.apache.hadoop.mapreduce.lib.input.LineRecordReader; //导入依赖的package包/类
@Override
public void initialize(InputSplit inputSplit, TaskAttemptContext context) throws IOException
{
  key = new Text();
  value = new MapWritable();
  jsonParser = new JSONParser();

  lineReader = new LineRecordReader();
  lineReader.initialize(inputSplit, context);

  queryString = context.getConfiguration().get("query", "?q=*");

  // Load the data schemas
  FileSystem fs = FileSystem.get(context.getConfiguration());
  try
  {
    SystemConfiguration.setProperty("data.schemas", context.getConfiguration().get("data.schemas"));
    DataSchemaLoader.initialize(true, fs);
  } catch (Exception e)
  {
    e.printStackTrace();
  }
  String dataSchemaName = context.getConfiguration().get("dataSchemaName");
  dataSchema = DataSchemaRegistry.get(dataSchemaName);
}
 
开发者ID:apache,项目名称:incubator-pirk,代码行数:26,代码来源:JSONRecordReader.java

示例2: createChildReader

import org.apache.hadoop.mapreduce.lib.input.LineRecordReader; //导入依赖的package包/类
/**
 * Actually instantiate the user's chosen RecordReader implementation.
 */
@SuppressWarnings("unchecked")
private void createChildReader() throws IOException, InterruptedException {
  LOG.debug("ChildSplit operates on: " + split.getPath(index));

  Configuration conf = context.getConfiguration();

  // Determine the file format we're reading.
  Class rrClass;
  if (ExportJobBase.isSequenceFiles(conf, split.getPath(index))) {
    rrClass = SequenceFileRecordReader.class;
  } else {
    rrClass = LineRecordReader.class;
  }

  // Create the appropriate record reader.
  this.rr = (RecordReader<LongWritable, Object>)
      ReflectionUtils.newInstance(rrClass, conf);
}
 
开发者ID:aliyun,项目名称:aliyun-maxcompute-data-collectors,代码行数:22,代码来源:CombineShimRecordReader.java

示例3: initialize

import org.apache.hadoop.mapreduce.lib.input.LineRecordReader; //导入依赖的package包/类
/**
 * Called once at initialization to initialize the RecordReader.
 *
 * @param genericSplit the split that defines the range of records to read.
 * @param context the information about the task.
 * @throws IOException on IO Error.
 */
@Override
public void initialize(InputSplit genericSplit, TaskAttemptContext context)
    throws IOException {
  if (LOG.isDebugEnabled()) {
    try {
      LOG.debug("initialize('{}', '{}')",
          HadoopToStringUtil.toString(genericSplit), HadoopToStringUtil.toString(context));
    } catch (InterruptedException ie) {
      LOG.debug("InterruptedException during HadoopToStringUtil.toString", ie);
    }
  }
  Preconditions.checkArgument(genericSplit instanceof FileSplit,
      "InputSplit genericSplit should be an instance of FileSplit.");
  // Get FileSplit.
  FileSplit fileSplit = (FileSplit) genericSplit;
  // Create the JsonParser.
  jsonParser = new JsonParser();
  // Initialize the LineRecordReader.
  lineReader = new LineRecordReader();
  lineReader.initialize(fileSplit, context);
}
 
开发者ID:GoogleCloudPlatform,项目名称:bigdata-interop,代码行数:29,代码来源:GsonRecordReader.java

示例4: initialize

import org.apache.hadoop.mapreduce.lib.input.LineRecordReader; //导入依赖的package包/类
@Override
public void initialize(final InputSplit inputSplit,
					   final TaskAttemptContext taskAttemptContext)
		throws IOException, InterruptedException {
	this.lrr = new LineRecordReader();
	this.lrr.initialize(inputSplit, taskAttemptContext);

}
 
开发者ID:jmabuin,项目名称:BLASpark,代码行数:9,代码来源:RowPerLineRecordReader.java

示例5: initialize

import org.apache.hadoop.mapreduce.lib.input.LineRecordReader; //导入依赖的package包/类
@Override
public void initialize(InputSplit split, TaskAttemptContext context)
		throws IOException, InterruptedException {
	int halfOfBufferSize = pair.value.capacity() / 2;
	maxLineSize = context.getConfiguration().getInt(
			LineRecordReader.MAX_LINE_LENGTH, halfOfBufferSize);
	if (maxLineSize > halfOfBufferSize) {
		context.getConfiguration().setInt(LineRecordReader.MAX_LINE_LENGTH,
				halfOfBufferSize);
		maxLineSize = halfOfBufferSize;
	}
	r.initialize(split, context);
	FileSplit fs = (FileSplit) split;
	start = fs.getStart();
}
 
开发者ID:ict-carch,项目名称:hadoop-plus,代码行数:16,代码来源:ARFFManyLineRecordReader.java

示例6: ScriptRecordReader

import org.apache.hadoop.mapreduce.lib.input.LineRecordReader; //导入依赖的package包/类
public ScriptRecordReader(final VertexQueryFilter vertexQuery, final TaskAttemptContext context) throws IOException {
    this.lineRecordReader = new LineRecordReader();
    this.vertexQuery = vertexQuery;
    this.configuration = DEFAULT_COMPAT.getContextConfiguration(context);
    this.faunusConf = ModifiableHadoopConfiguration.of(configuration);
    final FileSystem fs = FileSystem.get(configuration);
    try {
        this.engine.eval(new InputStreamReader(fs.open(new Path(faunusConf.getInputConf(ROOT_NS).get(SCRIPT_FILE)))));
    } catch (Exception e) {
        throw new IOException(e.getMessage());
    }
}
 
开发者ID:graben1437,项目名称:titan0.5.4-hbase1.1.1-custom,代码行数:13,代码来源:ScriptRecordReader.java

示例7: initialize

import org.apache.hadoop.mapreduce.lib.input.LineRecordReader; //导入依赖的package包/类
@Override
public void initialize(InputSplit inputSplit, TaskAttemptContext taskAttemptContext)
  throws IOException, InterruptedException {
  lineRecordReader = new LineRecordReader();
  lineRecordReader.initialize(inputSplit, taskAttemptContext);
  currentKey = new ImmutableBytesWritable();
  parser = new JSONParser();
  skipBadLines = taskAttemptContext.getConfiguration().getBoolean(
    SKIP_LINES_CONF_KEY, true);
}
 
开发者ID:lhfei,项目名称:hbase-in-action,代码行数:11,代码来源:BulkImportJobExample.java

示例8: close

import org.apache.hadoop.mapreduce.lib.input.LineRecordReader; //导入依赖的package包/类
@Override
@SuppressWarnings("squid:S2095") // recordReader is closed explictly in the close() method
public void initialize(InputSplit split, TaskAttemptContext context) throws IOException,
    InterruptedException
{
  if (split instanceof FileSplit)
  {
    FileSplit fsplit = (FileSplit) split;
    delimitedParser = getDelimitedParser(fsplit.getPath().toString(),
        context.getConfiguration());
    recordReader = new LineRecordReader();
    recordReader.initialize(fsplit, context);
    // Skip the first
    if (delimitedParser.getSkipFirstLine())
    {
      // Only skip the first line of the first split. The other
      // splits are somewhere in the middle of the original file,
      // so their first lines should not be skipped.
      if (fsplit.getStart() != 0)
      {
        nextKeyValue();
      }
    }
  }
  else
  {
    throw new IOException("input split is not a FileSplit");
  }
}
 
开发者ID:ngageoint,项目名称:mrgeo,代码行数:30,代码来源:DelimitedVectorRecordReader.java

示例9: createRecordReader

import org.apache.hadoop.mapreduce.lib.input.LineRecordReader; //导入依赖的package包/类
@Override
public RecordReader<LongWritable, Text> createRecordReader(
		InputSplit split,
		TaskAttemptContext context )
		throws IOException,
		InterruptedException {
	return new LineRecordReader();
}
 
开发者ID:locationtech,项目名称:geowave,代码行数:9,代码来源:GeonamesDataFileInputFormat.java

示例10: initializeNextReader

import org.apache.hadoop.mapreduce.lib.input.LineRecordReader; //导入依赖的package包/类
private void initializeNextReader() throws IOException {

			rdr = new LineRecordReader();
			rdr.initialize(
					new FileSplit(split.getPath(currentSplit),

					split.getOffset(currentSplit), split
							.getLength(currentSplit), null), context);

			++currentSplit;
		}
 
开发者ID:Pivotal-Field-Engineering,项目名称:pmr-common,代码行数:12,代码来源:CombineTextInputFormat.java

示例11: initialize

import org.apache.hadoop.mapreduce.lib.input.LineRecordReader; //导入依赖的package包/类
@Override
public void initialize(InputSplit split, TaskAttemptContext context)
		throws IOException, InterruptedException {

	rdr = new LineRecordReader();
	rdr.initialize(split, context);
}
 
开发者ID:Pivotal-Field-Engineering,项目名称:pmr-common,代码行数:8,代码来源:JsonInputFormat.java

示例12: initialize

import org.apache.hadoop.mapreduce.lib.input.LineRecordReader; //导入依赖的package包/类
@Override
public void initialize(InputSplit inputSplit, TaskAttemptContext attempt)
		throws IOException, InterruptedException {
	lineReader = new LineRecordReader();
	lineReader.initialize(inputSplit, attempt);		
			
}
 
开发者ID:willddy,项目名称:bigdata_pattern,代码行数:8,代码来源:LogFileRecordReader.java

示例13: TsvRecordReader

import org.apache.hadoop.mapreduce.lib.input.LineRecordReader; //导入依赖的package包/类
public TsvRecordReader(Configuration conf, int[] keyFields) throws IOException
{
	in = new LineRecordReader();
	if (keyFields.length == 0)
	{
		cutter = null;
		builder = null;
	}
	else
	{
		cutter = new CutText( conf.get(DELIM_CONF, DELIM_DEFALT), keyFields);
		builder = new StringBuilder(1000);
	}
}
 
开发者ID:ilveroluca,项目名称:seal,代码行数:15,代码来源:TsvInputFormat.java

示例14: initialize

import org.apache.hadoop.mapreduce.lib.input.LineRecordReader; //导入依赖的package包/类
public void initialize(InputSplit genericSplit, TaskAttemptContext context) throws IOException
{
	lineReader = new LineRecordReader();
	lineReader.initialize(genericSplit, context);

	split = (FileSplit)genericSplit;
	value = null;
}
 
开发者ID:ilveroluca,项目名称:seal,代码行数:9,代码来源:SamInputFormat.java

示例15: ScriptRecordReader

import org.apache.hadoop.mapreduce.lib.input.LineRecordReader; //导入依赖的package包/类
public ScriptRecordReader() {
    this.lineRecordReader = new LineRecordReader();
}
 
开发者ID:PKUSilvester,项目名称:LiteGraph,代码行数:4,代码来源:ScriptRecordReader.java


注:本文中的org.apache.hadoop.mapreduce.lib.input.LineRecordReader类示例由纯净天空整理自Github/MSDocs等开源代码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。