当前位置: 首页>>代码示例>>Java>>正文


Java Conf类代码示例

本文整理汇总了Java中org.apache.hadoop.hdfs.DFSClient.Conf的典型用法代码示例。如果您正苦于以下问题:Java Conf类的具体用法?Java Conf怎么用?Java Conf使用的例子?那么恭喜您, 这里精选的类代码示例或许可以为您提供帮助。


Conf类属于org.apache.hadoop.hdfs.DFSClient包,在下文中一共展示了Conf类的15个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Java代码示例。

示例1: ClientContext

import org.apache.hadoop.hdfs.DFSClient.Conf; //导入依赖的package包/类
private ClientContext(String name, Conf conf) {
  this.name = name;
  this.confString = confAsString(conf);
  this.shortCircuitCache = new ShortCircuitCache(
      conf.shortCircuitStreamsCacheSize,
      conf.shortCircuitStreamsCacheExpiryMs,
      conf.shortCircuitMmapCacheSize,
      conf.shortCircuitMmapCacheExpiryMs,
      conf.shortCircuitMmapCacheRetryTimeout,
      conf.shortCircuitCacheStaleThresholdMs,
      conf.shortCircuitSharedMemoryWatcherInterruptCheckMs);
  this.peerCache =
        new PeerCache(conf.socketCacheCapacity, conf.socketCacheExpiry);
  this.keyProviderCache = new KeyProviderCache(conf.keyProviderCacheExpiryMs);
  this.useLegacyBlockReaderLocal = conf.useLegacyBlockReaderLocal;
  this.domainSocketFactory = new DomainSocketFactory(conf);

  this.byteArrayManager = ByteArrayManager.newInstance(conf.writeByteArrayManagerConf);
}
 
开发者ID:naver,项目名称:hadoop,代码行数:20,代码来源:ClientContext.java

示例2: DomainSocketFactory

import org.apache.hadoop.hdfs.DFSClient.Conf; //导入依赖的package包/类
public DomainSocketFactory(Conf conf) {
  final String feature;
  if (conf.isShortCircuitLocalReads() && (!conf.isUseLegacyBlockReaderLocal())) {
    feature = "The short-circuit local reads feature";
  } else if (conf.isDomainSocketDataTraffic()) {
    feature = "UNIX domain socket data traffic";
  } else {
    feature = null;
  }

  if (feature == null) {
    PerformanceAdvisory.LOG.debug(
        "Both short-circuit local reads and UNIX domain socket are disabled.");
  } else {
    if (conf.getDomainSocketPath().isEmpty()) {
      throw new HadoopIllegalArgumentException(feature + " is enabled but "
          + DFSConfigKeys.DFS_DOMAIN_SOCKET_PATH_KEY + " is not set.");
    } else if (DomainSocket.getLoadingFailureReason() != null) {
      LOG.warn(feature + " cannot be used because "
          + DomainSocket.getLoadingFailureReason());
    } else {
      LOG.debug(feature + " is enabled.");
    }
  }
}
 
开发者ID:naver,项目名称:hadoop,代码行数:26,代码来源:DomainSocketFactory.java

示例3: getPathInfo

import org.apache.hadoop.hdfs.DFSClient.Conf; //导入依赖的package包/类
/**
 * Get information about a domain socket path.
 *
 * @param addr         The inet address to use.
 * @param conf         The client configuration.
 *
 * @return             Information about the socket path.
 */
public PathInfo getPathInfo(InetSocketAddress addr, DFSClient.Conf conf) {
  // If there is no domain socket path configured, we can't use domain
  // sockets.
  if (conf.getDomainSocketPath().isEmpty()) return PathInfo.NOT_CONFIGURED;
  // If we can't do anything with the domain socket, don't create it.
  if (!conf.isDomainSocketDataTraffic() &&
      (!conf.isShortCircuitLocalReads() || conf.isUseLegacyBlockReaderLocal())) {
    return PathInfo.NOT_CONFIGURED;
  }
  // If the DomainSocket code is not loaded, we can't create
  // DomainSocket objects.
  if (DomainSocket.getLoadingFailureReason() != null) {
    return PathInfo.NOT_CONFIGURED;
  }
  // UNIX domain sockets can only be used to talk to local peers
  if (!DFSClient.isLocalAddress(addr)) return PathInfo.NOT_CONFIGURED;
  String escapedPath = DomainSocket.getEffectivePath(
      conf.getDomainSocketPath(), addr.getPort());
  PathState status = pathMap.getIfPresent(escapedPath);
  if (status == null) {
    return new PathInfo(escapedPath, PathState.VALID);
  } else {
    return new PathInfo(escapedPath, status);
  }
}
 
开发者ID:naver,项目名称:hadoop,代码行数:34,代码来源:DomainSocketFactory.java

示例4: createProxy

import org.apache.hadoop.hdfs.DFSClient.Conf; //导入依赖的package包/类
/**
 * Creates the namenode proxy with the passed protocol. This will handle
 * creation of either HA- or non-HA-enabled proxy objects, depending upon
 * if the provided URI is a configured logical URI.
 * 
 * @param conf the configuration containing the required IPC
 *        properties, client failover configurations, etc.
 * @param nameNodeUri the URI pointing either to a specific NameNode
 *        or to a logical nameservice.
 * @param xface the IPC interface which should be created
 * @return an object containing both the proxy and the associated
 *         delegation token service it corresponds to
 * @throws IOException if there is an error creating the proxy
 **/
@SuppressWarnings("unchecked")
public static <T> ProxyAndInfo<T> createProxy(Configuration conf,
    URI nameNodeUri, Class<T> xface) throws IOException {
  Class<FailoverProxyProvider<T>> failoverProxyProviderClass =
      getFailoverProxyProviderClass(conf, nameNodeUri, xface);

  if (failoverProxyProviderClass == null) {
    // Non-HA case
    return createNonHAProxy(conf, NameNode.getAddress(nameNodeUri), xface,
        UserGroupInformation.getCurrentUser(), true);
  } else {
    // HA case
    FailoverProxyProvider<T> failoverProxyProvider = NameNodeProxies
        .createFailoverProxyProvider(conf, failoverProxyProviderClass, xface,
            nameNodeUri);
    Conf config = new Conf(conf);
    T proxy = (T) RetryProxy.create(xface, failoverProxyProvider, RetryPolicies
        .failoverOnNetworkException(RetryPolicies.TRY_ONCE_THEN_FAIL,
            config.maxFailoverAttempts, config.failoverSleepBaseMillis,
            config.failoverSleepMaxMillis));
    
    Text dtService = HAUtil.buildTokenServiceForLogicalUri(nameNodeUri);
    return new ProxyAndInfo<T>(proxy, dtService);
  }
}
 
开发者ID:ict-carch,项目名称:hadoop-plus,代码行数:40,代码来源:NameNodeProxies.java

示例5: DomainSocketFactory

import org.apache.hadoop.hdfs.DFSClient.Conf; //导入依赖的package包/类
public DomainSocketFactory(Conf conf) {
  this.conf = conf;

  final String feature;
  if (conf.shortCircuitLocalReads && (!conf.useLegacyBlockReaderLocal)) {
    feature = "The short-circuit local reads feature";
  } else if (conf.domainSocketDataTraffic) {
    feature = "UNIX domain socket data traffic";
  } else {
    feature = null;
  }

  if (feature == null) {
    LOG.debug("Both short-circuit local reads and UNIX domain socket are disabled.");
  } else {
    if (conf.domainSocketPath.isEmpty()) {
      throw new HadoopIllegalArgumentException(feature + " is enabled but "
          + DFSConfigKeys.DFS_DOMAIN_SOCKET_PATH_KEY + " is not set.");
    } else if (DomainSocket.getLoadingFailureReason() != null) {
      LOG.warn(feature + " cannot be used because "
          + DomainSocket.getLoadingFailureReason());
    } else {
      LOG.debug(feature + " is enabled.");
    }
  }
}
 
开发者ID:ict-carch,项目名称:hadoop-plus,代码行数:27,代码来源:DomainSocketFactory.java

示例6: ClientContext

import org.apache.hadoop.hdfs.DFSClient.Conf; //导入依赖的package包/类
private ClientContext(String name, Conf conf) {
  this.name = name;
  this.confString = confAsString(conf);
  this.shortCircuitCache = new ShortCircuitCache(
      conf.shortCircuitStreamsCacheSize,
      conf.shortCircuitStreamsCacheExpiryMs,
      conf.shortCircuitMmapCacheSize,
      conf.shortCircuitMmapCacheExpiryMs,
      conf.shortCircuitMmapCacheRetryTimeout,
      conf.shortCircuitCacheStaleThresholdMs,
      conf.shortCircuitSharedMemoryWatcherInterruptCheckMs);
  this.peerCache =
        new PeerCache(conf.socketCacheCapacity, conf.socketCacheExpiry);
  this.useLegacyBlockReaderLocal = conf.useLegacyBlockReaderLocal;
  this.domainSocketFactory = new DomainSocketFactory(conf);

  this.byteArrayManager = ByteArrayManager.newInstance(conf.writeByteArrayManagerConf);
}
 
开发者ID:yncxcw,项目名称:FlexMap,代码行数:19,代码来源:ClientContext.java

示例7: DomainSocketFactory

import org.apache.hadoop.hdfs.DFSClient.Conf; //导入依赖的package包/类
public DomainSocketFactory(Conf conf) {
  final String feature;
  if (conf.shortCircuitLocalReads && (!conf.useLegacyBlockReaderLocal)) {
    feature = "The short-circuit local reads feature";
  } else if (conf.domainSocketDataTraffic) {
    feature = "UNIX domain socket data traffic";
  } else {
    feature = null;
  }

  if (feature == null) {
    LOG.debug("Both short-circuit local reads and UNIX domain socket are disabled.");
  } else {
    if (conf.domainSocketPath.isEmpty()) {
      throw new HadoopIllegalArgumentException(feature + " is enabled but "
          + DFSConfigKeys.DFS_DOMAIN_SOCKET_PATH_KEY + " is not set.");
    } else if (DomainSocket.getLoadingFailureReason() != null) {
      LOG.warn(feature + " cannot be used because "
          + DomainSocket.getLoadingFailureReason());
    } else {
      LOG.debug(feature + " is enabled.");
    }
  }
}
 
开发者ID:Seagate,项目名称:hadoop-on-lustre2,代码行数:25,代码来源:DomainSocketFactory.java

示例8: getPathInfo

import org.apache.hadoop.hdfs.DFSClient.Conf; //导入依赖的package包/类
/**
 * Get information about a domain socket path.
 *
 * @param addr         The inet address to use.
 * @param conf         The client configuration.
 *
 * @return             Information about the socket path.
 */
public PathInfo getPathInfo(InetSocketAddress addr, DFSClient.Conf conf) {
  // If there is no domain socket path configured, we can't use domain
  // sockets.
  if (conf.domainSocketPath.isEmpty()) return PathInfo.NOT_CONFIGURED;
  // If we can't do anything with the domain socket, don't create it.
  if (!conf.domainSocketDataTraffic &&
      (!conf.shortCircuitLocalReads || conf.useLegacyBlockReaderLocal)) {
    return PathInfo.NOT_CONFIGURED;
  }
  // If the DomainSocket code is not loaded, we can't create
  // DomainSocket objects.
  if (DomainSocket.getLoadingFailureReason() != null) {
    return PathInfo.NOT_CONFIGURED;
  }
  // UNIX domain sockets can only be used to talk to local peers
  if (!DFSClient.isLocalAddress(addr)) return PathInfo.NOT_CONFIGURED;
  String escapedPath = DomainSocket.
      getEffectivePath(conf.domainSocketPath, addr.getPort());
  PathState status = pathMap.getIfPresent(escapedPath);
  if (status == null) {
    return new PathInfo(escapedPath, PathState.VALID);
  } else {
    return new PathInfo(escapedPath, status);
  }
}
 
开发者ID:Seagate,项目名称:hadoop-on-lustre2,代码行数:34,代码来源:DomainSocketFactory.java

示例9: ClientContext

import org.apache.hadoop.hdfs.DFSClient.Conf; //导入依赖的package包/类
private ClientContext(String name, Conf conf) {
  this.name = name;
  this.confString = confAsString(conf);
  this.shortCircuitCache = new ShortCircuitCache(
      conf.shortCircuitStreamsCacheSize,
      conf.shortCircuitStreamsCacheExpiryMs,
      conf.shortCircuitMmapCacheSize,
      conf.shortCircuitMmapCacheExpiryMs,
      conf.shortCircuitMmapCacheRetryTimeout,
      conf.shortCircuitCacheStaleThresholdMs,
      conf.shortCircuitSharedMemoryWatcherInterruptCheckMs);
  this.peerCache =
        new PeerCache(conf.socketCacheCapacity, conf.socketCacheExpiry);
  this.useLegacyBlockReaderLocal = conf.useLegacyBlockReaderLocal;
  this.domainSocketFactory = new DomainSocketFactory(conf);
}
 
开发者ID:Seagate,项目名称:hadoop-on-lustre2,代码行数:17,代码来源:ClientContext.java

示例10: createProxy

import org.apache.hadoop.hdfs.DFSClient.Conf; //导入依赖的package包/类
/**
 * Creates the namenode proxy with the passed protocol. This will handle
 * creation of either HA- or non-HA-enabled proxy objects, depending upon
 * if the provided URI is a configured logical URI.
 *
 * @param conf the configuration containing the required IPC
 *        properties, client failover configurations, etc.
 * @param nameNodeUri the URI pointing either to a specific NameNode
 *        or to a logical nameservice.
 * @param xface the IPC interface which should be created
 * @param fallbackToSimpleAuth set to true or false during calls to indicate if
 *   a secure client falls back to simple auth
 * @return an object containing both the proxy and the associated
 *         delegation token service it corresponds to
 * @throws IOException if there is an error creating the proxy
 **/
@SuppressWarnings("unchecked")
public static <T> ProxyAndInfo<T> createProxy(Configuration conf,
    URI nameNodeUri, Class<T> xface, AtomicBoolean fallbackToSimpleAuth)
    throws IOException {
  AbstractNNFailoverProxyProvider<T> failoverProxyProvider =
      createFailoverProxyProvider(conf, nameNodeUri, xface, true,
        fallbackToSimpleAuth);

  if (failoverProxyProvider == null) {
    // Non-HA case
    return createNonHAProxy(conf, NameNode.getAddress(nameNodeUri), xface,
        UserGroupInformation.getCurrentUser(), true, fallbackToSimpleAuth);
  } else {
    // HA case
    Conf config = new Conf(conf);
    T proxy = (T) RetryProxy.create(xface, failoverProxyProvider,
        RetryPolicies.failoverOnNetworkException(
            RetryPolicies.TRY_ONCE_THEN_FAIL, config.maxFailoverAttempts,
            config.maxRetryAttempts, config.failoverSleepBaseMillis,
            config.failoverSleepMaxMillis));

    Text dtService;
    if (failoverProxyProvider.useLogicalURI()) {
      dtService = HAUtil.buildTokenServiceForLogicalUri(nameNodeUri,
          HdfsConstants.HDFS_URI_SCHEME);
    } else {
      dtService = SecurityUtil.buildTokenService(
          NameNode.getAddress(nameNodeUri));
    }
    return new ProxyAndInfo<T>(proxy, dtService,
        NameNode.getAddress(nameNodeUri));
  }
}
 
开发者ID:naver,项目名称:hadoop,代码行数:50,代码来源:NameNodeProxies.java

示例11: confAsString

import org.apache.hadoop.hdfs.DFSClient.Conf; //导入依赖的package包/类
public static String confAsString(Conf conf) {
  StringBuilder builder = new StringBuilder();
  builder.append("shortCircuitStreamsCacheSize = ").
    append(conf.shortCircuitStreamsCacheSize).
    append(", shortCircuitStreamsCacheExpiryMs = ").
    append(conf.shortCircuitStreamsCacheExpiryMs).
    append(", shortCircuitMmapCacheSize = ").
    append(conf.shortCircuitMmapCacheSize).
    append(", shortCircuitMmapCacheExpiryMs = ").
    append(conf.shortCircuitMmapCacheExpiryMs).
    append(", shortCircuitMmapCacheRetryTimeout = ").
    append(conf.shortCircuitMmapCacheRetryTimeout).
    append(", shortCircuitCacheStaleThresholdMs = ").
    append(conf.shortCircuitCacheStaleThresholdMs).
    append(", socketCacheCapacity = ").
    append(conf.socketCacheCapacity).
    append(", socketCacheExpiry = ").
    append(conf.socketCacheExpiry).
    append(", shortCircuitLocalReads = ").
    append(conf.shortCircuitLocalReads).
    append(", useLegacyBlockReaderLocal = ").
    append(conf.useLegacyBlockReaderLocal).
    append(", domainSocketDataTraffic = ").
    append(conf.domainSocketDataTraffic).
    append(", shortCircuitSharedMemoryWatcherInterruptCheckMs = ").
    append(conf.shortCircuitSharedMemoryWatcherInterruptCheckMs).
    append(", keyProviderCacheExpiryMs = ").
    append(conf.keyProviderCacheExpiryMs);

  return builder.toString();
}
 
开发者ID:naver,项目名称:hadoop,代码行数:32,代码来源:ClientContext.java

示例12: get

import org.apache.hadoop.hdfs.DFSClient.Conf; //导入依赖的package包/类
public static ClientContext get(String name, Conf conf) {
  ClientContext context;
  synchronized(ClientContext.class) {
    context = CACHES.get(name);
    if (context == null) {
      context = new ClientContext(name, conf);
      CACHES.put(name, context);
    } else {
      context.printConfWarningIfNeeded(conf);
    }
  }
  return context;
}
 
开发者ID:naver,项目名称:hadoop,代码行数:14,代码来源:ClientContext.java

示例13: getFromConf

import org.apache.hadoop.hdfs.DFSClient.Conf; //导入依赖的package包/类
/**
 * Get a client context, from a Configuration object.
 *
 * This method is less efficient than the version which takes a DFSClient#Conf
 * object, and should be mostly used by tests.
 */
@VisibleForTesting
public static ClientContext getFromConf(Configuration conf) {
  return get(conf.get(DFSConfigKeys.DFS_CLIENT_CONTEXT,
      DFSConfigKeys.DFS_CLIENT_CONTEXT_DEFAULT),
          new DFSClient.Conf(conf));
}
 
开发者ID:naver,项目名称:hadoop,代码行数:13,代码来源:ClientContext.java

示例14: printConfWarningIfNeeded

import org.apache.hadoop.hdfs.DFSClient.Conf; //导入依赖的package包/类
private void printConfWarningIfNeeded(Conf conf) {
  String existing = this.getConfString();
  String requested = confAsString(conf);
  if (!existing.equals(requested)) {
    if (!printedConfWarning) {
      printedConfWarning = true;
      LOG.warn("Existing client context '" + name + "' does not match " +
          "requested configuration.  Existing: " + existing + 
          ", Requested: " + requested);
    }
  }
}
 
开发者ID:naver,项目名称:hadoop,代码行数:13,代码来源:ClientContext.java

示例15: confAsString

import org.apache.hadoop.hdfs.DFSClient.Conf; //导入依赖的package包/类
public static String confAsString(Conf conf) {
  StringBuilder builder = new StringBuilder();
  builder.append("shortCircuitStreamsCacheSize = ").
    append(conf.shortCircuitStreamsCacheSize).
    append(", shortCircuitStreamsCacheExpiryMs = ").
    append(conf.shortCircuitStreamsCacheExpiryMs).
    append(", shortCircuitMmapCacheSize = ").
    append(conf.shortCircuitMmapCacheSize).
    append(", shortCircuitMmapCacheExpiryMs = ").
    append(conf.shortCircuitMmapCacheExpiryMs).
    append(", shortCircuitMmapCacheRetryTimeout = ").
    append(conf.shortCircuitMmapCacheRetryTimeout).
    append(", shortCircuitCacheStaleThresholdMs = ").
    append(conf.shortCircuitCacheStaleThresholdMs).
    append(", socketCacheCapacity = ").
    append(conf.socketCacheCapacity).
    append(", socketCacheExpiry = ").
    append(conf.socketCacheExpiry).
    append(", shortCircuitLocalReads = ").
    append(conf.shortCircuitLocalReads).
    append(", useLegacyBlockReaderLocal = ").
    append(conf.useLegacyBlockReaderLocal).
    append(", domainSocketDataTraffic = ").
    append(conf.domainSocketDataTraffic).
    append(", shortCircuitSharedMemoryWatcherInterruptCheckMs = ").
    append(conf.shortCircuitSharedMemoryWatcherInterruptCheckMs);

  return builder.toString();
}
 
开发者ID:yncxcw,项目名称:FlexMap,代码行数:30,代码来源:ClientContext.java


注:本文中的org.apache.hadoop.hdfs.DFSClient.Conf类示例由纯净天空整理自Github/MSDocs等开源代码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。