当前位置: 首页>>代码示例>>Java>>正文


Java LightWeightGSet.computeCapacity方法代码示例

本文整理汇总了Java中org.apache.hadoop.util.LightWeightGSet.computeCapacity方法的典型用法代码示例。如果您正苦于以下问题:Java LightWeightGSet.computeCapacity方法的具体用法?Java LightWeightGSet.computeCapacity怎么用?Java LightWeightGSet.computeCapacity使用的例子?那么恭喜您, 这里精选的方法代码示例或许可以为您提供帮助。您也可以进一步了解该方法所在org.apache.hadoop.util.LightWeightGSet的用法示例。


在下文中一共展示了LightWeightGSet.computeCapacity方法的8个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Java代码示例。

示例1: RetryCache

import org.apache.hadoop.util.LightWeightGSet; //导入方法依赖的package包/类
/**
 * Constructor
 * @param cacheName name to identify the cache by
 * @param percentage percentage of total java heap space used by this cache
 * @param expirationTime time for an entry to expire in nanoseconds
 */
public RetryCache(String cacheName, double percentage, long expirationTime) {
  int capacity = LightWeightGSet.computeCapacity(percentage, cacheName);
  capacity = capacity > MAX_CAPACITY ? capacity : MAX_CAPACITY;
  this.set = new LightWeightCache<CacheEntry, CacheEntry>(capacity, capacity,
      expirationTime, 0);
  this.expirationTime = expirationTime;
  this.cacheName = cacheName;
  this.retryCacheMetrics =  RetryCacheMetrics.create(this);
}
 
开发者ID:nucypher,项目名称:hadoop-oss,代码行数:16,代码来源:RetryCache.java

示例2: newInstance

import org.apache.hadoop.util.LightWeightGSet; //导入方法依赖的package包/类
static INodeMap newInstance(INodeDirectory rootDir) {
  // Compute the map capacity by allocating 1% of total memory
  int capacity = LightWeightGSet.computeCapacity(1, "INodeMap");
  GSet<INode, INodeWithAdditionalFields> map
      = new LightWeightGSet<INode, INodeWithAdditionalFields>(capacity);
  map.put(rootDir);
  return new INodeMap(map);
}
 
开发者ID:naver,项目名称:hadoop,代码行数:9,代码来源:INodeMap.java

示例3: CacheManager

import org.apache.hadoop.util.LightWeightGSet; //导入方法依赖的package包/类
CacheManager(FSNamesystem namesystem, Configuration conf,
    BlockManager blockManager) {
  this.namesystem = namesystem;
  this.blockManager = blockManager;
  this.nextDirectiveId = 1;
  this.maxListCachePoolsResponses = conf.getInt(
      DFS_NAMENODE_LIST_CACHE_POOLS_NUM_RESPONSES,
      DFS_NAMENODE_LIST_CACHE_POOLS_NUM_RESPONSES_DEFAULT);
  this.maxListCacheDirectivesNumResponses = conf.getInt(
      DFS_NAMENODE_LIST_CACHE_DIRECTIVES_NUM_RESPONSES,
      DFS_NAMENODE_LIST_CACHE_DIRECTIVES_NUM_RESPONSES_DEFAULT);
  scanIntervalMs = conf.getLong(
      DFS_NAMENODE_PATH_BASED_CACHE_REFRESH_INTERVAL_MS,
      DFS_NAMENODE_PATH_BASED_CACHE_REFRESH_INTERVAL_MS_DEFAULT);
  float cachedBlocksPercent = conf.getFloat(
        DFS_NAMENODE_PATH_BASED_CACHE_BLOCK_MAP_ALLOCATION_PERCENT,
        DFS_NAMENODE_PATH_BASED_CACHE_BLOCK_MAP_ALLOCATION_PERCENT_DEFAULT);
  if (cachedBlocksPercent < MIN_CACHED_BLOCKS_PERCENT) {
    LOG.info("Using minimum value {} for {}", MIN_CACHED_BLOCKS_PERCENT,
      DFS_NAMENODE_PATH_BASED_CACHE_BLOCK_MAP_ALLOCATION_PERCENT);
    cachedBlocksPercent = MIN_CACHED_BLOCKS_PERCENT;
  }
  this.cachedBlocks = new LightWeightGSet<CachedBlock, CachedBlock>(
        LightWeightGSet.computeCapacity(cachedBlocksPercent,
            "cachedBlocks"));

}
 
开发者ID:naver,项目名称:hadoop,代码行数:28,代码来源:CacheManager.java

示例4: RetryCache

import org.apache.hadoop.util.LightWeightGSet; //导入方法依赖的package包/类
/**
 * Constructor
 * @param cacheName name to identify the cache by
 * @param percentage percentage of total java heap space used by this cache
 * @param expirationTime time for an entry to expire in nanoseconds
 */
public RetryCache(String cacheName, double percentage, long expirationTime) {
  int capacity = LightWeightGSet.computeCapacity(percentage, cacheName);
  capacity = capacity > 16 ? capacity : 16;
  this.set = new LightWeightCache<CacheEntry, CacheEntry>(capacity, capacity,
      expirationTime, 0);
  this.expirationTime = expirationTime;
  this.cacheName = cacheName;
  this.retryCacheMetrics =  RetryCacheMetrics.create(this);
}
 
开发者ID:naver,项目名称:hadoop,代码行数:16,代码来源:RetryCache.java

示例5: newInstance

import org.apache.hadoop.util.LightWeightGSet; //导入方法依赖的package包/类
static INodeMap newInstance(INodeDirectory rootDir) {
  // Compute the map capacity by allocating 1% of total memory
  int capacity = LightWeightGSet.computeCapacity(1, "INodeMap");
  GSet<INode, INodeWithAdditionalFields> map =
      new LightWeightGSet<>(capacity);
  map.put(rootDir);
  return new INodeMap(map);
}
 
开发者ID:aliyun-beta,项目名称:aliyun-oss-hadoop-fs,代码行数:9,代码来源:INodeMap.java

示例6: RetryCache

import org.apache.hadoop.util.LightWeightGSet; //导入方法依赖的package包/类
/**
 * Constructor
 * @param cacheName name to identify the cache by
 * @param percentage percentage of total java heap space used by this cache
 * @param expirationTime time for an entry to expire in nanoseconds
 */
public RetryCache(String cacheName, double percentage, long expirationTime) {
  int capacity = LightWeightGSet.computeCapacity(percentage, cacheName);
  capacity = capacity > 16 ? capacity : 16;
  this.set = new LightWeightCache<CacheEntry, CacheEntry>(capacity, capacity,
      expirationTime, 0);
  this.expirationTime = expirationTime;
}
 
开发者ID:ict-carch,项目名称:hadoop-plus,代码行数:14,代码来源:RetryCache.java

示例7: CacheManager

import org.apache.hadoop.util.LightWeightGSet; //导入方法依赖的package包/类
CacheManager(FSNamesystem namesystem, Configuration conf,
    BlockManager blockManager) {
  this.namesystem = namesystem;
  this.blockManager = blockManager;
  this.nextDirectiveId = 1;
  this.maxListCachePoolsResponses = conf.getInt(
      DFS_NAMENODE_LIST_CACHE_POOLS_NUM_RESPONSES,
      DFS_NAMENODE_LIST_CACHE_POOLS_NUM_RESPONSES_DEFAULT);
  this.maxListCacheDirectivesNumResponses = conf.getInt(
      DFS_NAMENODE_LIST_CACHE_DIRECTIVES_NUM_RESPONSES,
      DFS_NAMENODE_LIST_CACHE_DIRECTIVES_NUM_RESPONSES_DEFAULT);
  scanIntervalMs = conf.getLong(
      DFS_NAMENODE_PATH_BASED_CACHE_REFRESH_INTERVAL_MS,
      DFS_NAMENODE_PATH_BASED_CACHE_REFRESH_INTERVAL_MS_DEFAULT);
  float cachedBlocksPercent = conf.getFloat(
        DFS_NAMENODE_PATH_BASED_CACHE_BLOCK_MAP_ALLOCATION_PERCENT,
        DFS_NAMENODE_PATH_BASED_CACHE_BLOCK_MAP_ALLOCATION_PERCENT_DEFAULT);
  if (cachedBlocksPercent < MIN_CACHED_BLOCKS_PERCENT) {
    LOG.info("Using minimum value " + MIN_CACHED_BLOCKS_PERCENT +
      " for " + DFS_NAMENODE_PATH_BASED_CACHE_BLOCK_MAP_ALLOCATION_PERCENT);
    cachedBlocksPercent = MIN_CACHED_BLOCKS_PERCENT;
  }
  this.cachedBlocks = new LightWeightGSet<CachedBlock, CachedBlock>(
        LightWeightGSet.computeCapacity(cachedBlocksPercent,
            "cachedBlocks"));

}
 
开发者ID:Seagate,项目名称:hadoop-on-lustre2,代码行数:28,代码来源:CacheManager.java

示例8: BlocksMap

import org.apache.hadoop.util.LightWeightGSet; //导入方法依赖的package包/类
BlocksMap(final float loadFactor) {
  // Use 2% of total memory to size the GSet capacity
  this.capacity = LightWeightGSet.computeCapacity(2.0, "BlocksMap");
  this.blocks = new LightWeightGSet<Block, BlockInfo>(capacity);
}
 
开发者ID:ict-carch,项目名称:hadoop-plus,代码行数:6,代码来源:BlocksMap.java


注:本文中的org.apache.hadoop.util.LightWeightGSet.computeCapacity方法示例由纯净天空整理自Github/MSDocs等开源代码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。