当前位置: 首页>>代码示例>>Java>>正文


Java HFileBlockDefaultDecodingContext类代码示例

本文整理汇总了Java中org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultDecodingContext的典型用法代码示例。如果您正苦于以下问题:Java HFileBlockDefaultDecodingContext类的具体用法?Java HFileBlockDefaultDecodingContext怎么用?Java HFileBlockDefaultDecodingContext使用的例子?那么恭喜您, 这里精选的类代码示例或许可以为您提供帮助。


HFileBlockDefaultDecodingContext类属于org.apache.hadoop.hbase.io.encoding包,在下文中一共展示了HFileBlockDefaultDecodingContext类的12个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Java代码示例。

示例1: FSReaderImpl

import org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultDecodingContext; //导入依赖的package包/类
FSReaderImpl(FSDataInputStreamWrapper stream, long fileSize, HFileSystem hfs, Path path,
    HFileContext fileContext) throws IOException {
  this.fileSize = fileSize;
  this.hfs = hfs;
  if (path != null) {
    this.pathName = path.toString();
  }
  this.fileContext = fileContext;
  this.hdrSize = headerSize(fileContext.isUseHBaseChecksum());

  this.streamWrapper = stream;
  // Older versions of HBase didn't support checksum.
  this.streamWrapper.prepareForBlockReader(!fileContext.isUseHBaseChecksum());
  defaultDecodingCtx = new HFileBlockDefaultDecodingContext(fileContext);
  encodedBlockDecodingCtx = defaultDecodingCtx;
}
 
开发者ID:apache,项目名称:hbase,代码行数:17,代码来源:HFileBlock.java

示例2: newDataBlockDecodingContext

import org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultDecodingContext; //导入依赖的package包/类
@Override
public HFileBlockDecodingContext newDataBlockDecodingContext(HFileContext fileContext) {
  DataBlockEncoder encoder = encoding.getEncoder();
  if (encoder != null) {
    return encoder.newDataBlockDecodingContext(fileContext);
  }
  return new HFileBlockDefaultDecodingContext(fileContext);
}
 
开发者ID:fengchen8086,项目名称:ditb,代码行数:9,代码来源:HFileDataBlockEncoderImpl.java

示例3: FSReaderImpl

import org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultDecodingContext; //导入依赖的package包/类
public FSReaderImpl(FSDataInputStreamWrapper stream, long fileSize, HFileSystem hfs, Path path,
    HFileContext fileContext) throws IOException {
  super(fileSize, hfs, path, fileContext);
  this.streamWrapper = stream;
  // Older versions of HBase didn't support checksum.
  this.streamWrapper.prepareForBlockReader(!fileContext.isUseHBaseChecksum());
  defaultDecodingCtx = new HFileBlockDefaultDecodingContext(fileContext);
  encodedBlockDecodingCtx = defaultDecodingCtx;
}
 
开发者ID:fengchen8086,项目名称:ditb,代码行数:10,代码来源:HFileBlock.java

示例4: FSReaderV2

import org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultDecodingContext; //导入依赖的package包/类
public FSReaderV2(FSDataInputStreamWrapper stream, long fileSize, HFileSystem hfs, Path path,
    HFileContext fileContext) throws IOException {
  super(fileSize, hfs, path, fileContext);
  this.streamWrapper = stream;
  // Older versions of HBase didn't support checksum.
  this.streamWrapper.prepareForBlockReader(!fileContext.isUseHBaseChecksum());
  defaultDecodingCtx =
    new HFileBlockDefaultDecodingContext(fileContext);
  encodedBlockDecodingCtx =
      new HFileBlockDefaultDecodingContext(fileContext);
}
 
开发者ID:tenggyut,项目名称:HIndex,代码行数:12,代码来源:HFileBlock.java

示例5: newDataBlockDecodingContext

import org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultDecodingContext; //导入依赖的package包/类
@Override
public HFileBlockDecodingContext newDataBlockDecodingContext(
    Algorithm compressionAlgorithm) {
  DataBlockEncoder encoder = encoding.getEncoder();
  if (encoder != null) {
    return encoder.newDataBlockDecodingContext(compressionAlgorithm);
  }
  return new HFileBlockDefaultDecodingContext(compressionAlgorithm);
}
 
开发者ID:cloud-software-foundation,项目名称:c5,代码行数:10,代码来源:HFileDataBlockEncoderImpl.java

示例6: FSReaderV2

import org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultDecodingContext; //导入依赖的package包/类
public FSReaderV2(FSDataInputStreamWrapper stream, Algorithm compressAlgo, long fileSize,
    int minorVersion, HFileSystem hfs, Path path) throws IOException {
  super(compressAlgo, fileSize, minorVersion, hfs, path);
  this.streamWrapper = stream;
  // Older versions of HBase didn't support checksum.
  boolean forceNoHBaseChecksum = (this.getMinorVersion() < MINOR_VERSION_WITH_CHECKSUM);
  this.streamWrapper.prepareForBlockReader(forceNoHBaseChecksum);

  defaultDecodingCtx =
    new HFileBlockDefaultDecodingContext(compressAlgo);
  encodedBlockDecodingCtx =
      new HFileBlockDefaultDecodingContext(compressAlgo);
}
 
开发者ID:cloud-software-foundation,项目名称:c5,代码行数:14,代码来源:HFileBlock.java

示例7: newOnDiskDataBlockDecodingContext

import org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultDecodingContext; //导入依赖的package包/类
@Override
public HFileBlockDecodingContext newOnDiskDataBlockDecodingContext(
    Algorithm compressionAlgorithm) {
  if (onDisk != null) {
    DataBlockEncoder encoder = onDisk.getEncoder();
    if (encoder != null) {
      return encoder.newDataBlockDecodingContext(
          compressionAlgorithm);
    }
  }
  return new HFileBlockDefaultDecodingContext(compressionAlgorithm);
}
 
开发者ID:daidong,项目名称:DominoHBase,代码行数:13,代码来源:HFileDataBlockEncoderImpl.java

示例8: FSReaderV2

import org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultDecodingContext; //导入依赖的package包/类
public FSReaderV2(FSDataInputStream istream, 
    FSDataInputStream istreamNoFsChecksum, Algorithm compressAlgo,
    long fileSize, int minorVersion, HFileSystem hfs, Path path) 
  throws IOException {
  super(istream, istreamNoFsChecksum, compressAlgo, fileSize, 
        minorVersion, hfs, path);

  if (hfs != null) {
    // Check the configuration to determine whether hbase-level
    // checksum verification is needed or not.
    useHBaseChecksum = hfs.useHBaseChecksum();
  } else {
    // The configuration does not specify anything about hbase checksum
    // validations. Set it to true here assuming that we will verify
    // hbase checksums for all reads. For older files that do not have 
    // stored checksums, this flag will be reset later.
    useHBaseChecksum = true;
  }

  // for older versions, hbase did not store checksums.
  if (getMinorVersion() < MINOR_VERSION_WITH_CHECKSUM) {
    useHBaseChecksum = false;
  }
  this.useHBaseChecksumConfigured = useHBaseChecksum;
  defaultDecodingCtx =
    new HFileBlockDefaultDecodingContext(compressAlgo);
  encodedBlockDecodingCtx =
      new HFileBlockDefaultDecodingContext(compressAlgo);
}
 
开发者ID:daidong,项目名称:DominoHBase,代码行数:30,代码来源:HFileBlock.java

示例9: newDataBlockDecodingContext

import org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultDecodingContext; //导入依赖的package包/类
@Override
public HFileBlockDecodingContext newDataBlockDecodingContext(HFileContext meta) {
  return new HFileBlockDefaultDecodingContext(meta);
}
 
开发者ID:fengchen8086,项目名称:ditb,代码行数:5,代码来源:NoOpDataBlockEncoder.java

示例10: newDataBlockDecodingContext

import org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultDecodingContext; //导入依赖的package包/类
@Override
public HFileBlockDecodingContext newDataBlockDecodingContext(
    Algorithm compressionAlgorithm) {
  return new HFileBlockDefaultDecodingContext(compressionAlgorithm);
}
 
开发者ID:cloud-software-foundation,项目名称:c5,代码行数:6,代码来源:NoOpDataBlockEncoder.java

示例11: newDataBlockDecodingContext

import org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultDecodingContext; //导入依赖的package包/类
@Override
public HFileBlockDecodingContext newDataBlockDecodingContext(Algorithm compressionAlgorithm) {
  return new HFileBlockDefaultDecodingContext(compressionAlgorithm);
}
 
开发者ID:cloud-software-foundation,项目名称:c5,代码行数:5,代码来源:PrefixTreeCodec.java

示例12: newOnDiskDataBlockDecodingContext

import org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultDecodingContext; //导入依赖的package包/类
@Override
public HFileBlockDecodingContext newOnDiskDataBlockDecodingContext(
    Algorithm compressionAlgorithm) {
  return new HFileBlockDefaultDecodingContext(compressionAlgorithm);
}
 
开发者ID:daidong,项目名称:DominoHBase,代码行数:6,代码来源:NoOpDataBlockEncoder.java


注:本文中的org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultDecodingContext类示例由纯净天空整理自Github/MSDocs等开源代码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。