当前位置: 首页>>代码示例>>Java>>正文


Java CountMetric类代码示例

本文整理汇总了Java中backtype.storm.metric.api.CountMetric的典型用法代码示例。如果您正苦于以下问题:Java CountMetric类的具体用法?Java CountMetric怎么用?Java CountMetric使用的例子?那么恭喜您, 这里精选的类代码示例或许可以为您提供帮助。


CountMetric类属于backtype.storm.metric.api包,在下文中一共展示了CountMetric类的11个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Java代码示例。

示例1: open

import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
@Override
public void open(Map conf, TopologyContext context,
        SpoutOutputCollector _collector) {
    collector = _collector;
    _spoutMetric = new CountMetric();
    context.registerMetric(CatMetricUtil.getSpoutMetricName(topic, group),  
            _spoutMetric, Constants.EMIT_FREQUENCY_IN_SECONDS);
    
    ConsumerConfig config = new ConsumerConfig();
    try {
        consumer = new Consumer(topic, group, config);
    } catch (LionException e) {
        throw new RuntimeException(e);
    }
    consumer.start();
    stream = consumer.getStream();
    
    fetchThread = new MessageFetcher(stream);
    new Thread(fetchThread).start();
}
 
开发者ID:kkllwww007,项目名称:jstrom,代码行数:21,代码来源:BlackholeBlockingQueueSpout.java

示例2: open

import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
@Override
public void open(Map conf, TopologyContext context,
        SpoutOutputCollector _collector) {
    collector = _collector;
    _spoutMetric = new CountMetric();
    context.registerMetric(CatMetricUtil.getSpoutMetricName(topic, group),  
            _spoutMetric, Constants.EMIT_FREQUENCY_IN_SECONDS);
    
    ConsumerConfig config = new ConsumerConfig();
    try {
        consumer = new Consumer(topic, group, config);
    } catch (LionException e) {
        throw new RuntimeException(e);
    }
    consumer.start();
    stream = consumer.getStream();
}
 
开发者ID:kkllwww007,项目名称:jstrom,代码行数:18,代码来源:BlackholeSpout.java

示例3: registerMetrics

import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
@SuppressWarnings("rawtypes")
  public void registerMetrics(Map conf, IMetricsContext context, String mapStateMetricName) {
      int bucketSize = (Integer) (conf.get(Config.TOPOLOGY_BUILTIN_METRICS_BUCKET_SIZE_SECS));
      String metricBaseName = "cassandra/" + mapStateMetricName;
_mreads = context.registerMetric(metricBaseName + "/readCount", new CountMetric(), bucketSize);
      _mwrites = context.registerMetric(metricBaseName + "/writeCount", new CountMetric(), bucketSize);
      _mexceptions = context.registerMetric(metricBaseName + "/exceptionCount", new CountMetric(), bucketSize);
  }
 
开发者ID:hpcc-systems,项目名称:storm-cassandra-cql,代码行数:9,代码来源:CassandraCqlMapState.java

示例4: registerMetrics

import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
/**
 * Register Metrics in Storm
 * @param conf A set of properties
 * @param context Metrics context
 */
public void registerMetrics(Map conf, IMetricsContext context) {
    int bucketSize = (Integer) (conf.get(Config.TOPOLOGY_BUILTIN_METRICS_BUCKET_SIZE_SECS));
    _mreads = context.registerMetric("mongodb/readCount", new CountMetric(), bucketSize);
    _mwrites = context.registerMetric("mongodb/writeCount", new CountMetric(), bucketSize);
    _mexceptions = context.registerMetric("mongodb/exceptionCount", new CountMetric(), bucketSize);
}
 
开发者ID:andressanchez,项目名称:storm-mongodb,代码行数:12,代码来源:MongoDBMapState.java

示例5: open

import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
@Override
public void open(Map conf, TopologyContext context, SpoutOutputCollector collector) {
    _throttler = new WindowedTimeThrottler((Number) conf.get(Config.TOPOLOGY_TRIDENT_BATCH_EMIT_INTERVAL_MILLIS),
            Integer.MAX_VALUE);
    for (String spoutId : _managedSpoutIds) {
        _states.add(TransactionalState.newCoordinatorState(conf, spoutId));
    }
    _currTransaction = getStoredCurrTransaction();

    _collector = collector;
    Number active = (Number) conf.get(Config.TOPOLOGY_MAX_SPOUT_PENDING);
    if (active == null) {
        _maxTransactionActive = 1;
    } else {
        _maxTransactionActive = active.intValue();
    }
    _attemptIds = getStoredCurrAttempts(_currTransaction, _maxTransactionActive);
    avgWaittingTime = context.registerMetric("commit-wait-ms", new ReducedMetric(new MeanReducer()),
            Utils.getInt(conf.get(Config.TOPOLOGY_BUILTIN_METRICS_BUCKET_SIZE_SECS)));
    commitWaittingCount = context.registerMetric("commit-wait-cnt", new CountMetric(),
            Utils.getInt(conf.get(Config.TOPOLOGY_BUILTIN_METRICS_BUCKET_SIZE_SECS)));
    activeTxCount = context.registerMetric("active-tx-cnt", new ReducedMetric(new MeanReducer()),
            Utils.getInt(conf.get(Config.TOPOLOGY_BUILTIN_METRICS_BUCKET_SIZE_SECS)));
    for (int i = 0; i < _spouts.size(); i++) {
        String txId = _managedSpoutIds.get(i);
        _coordinators.add(_spouts.get(i).getCoordinator(txId, conf, context));
    }
}
 
开发者ID:troyding,项目名称:storm-resa,代码行数:29,代码来源:MasterBatchCoordinator.java

示例6: prepare

import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
@Override
public void prepare(Map config, TopologyContext context) {
  // Optional set logging
  if (config.get(CONF_LOGGING) != null) {
    m_logging = (Boolean) config.get(CONF_LOGGING);
  } else {
    m_logging = false;
  }

  // Tuple counter metric
  if (config.get(CONF_METRIC_LOGGING_INTERVALL) != null) {
    m_countMetric = new CountMetric();
    context.registerMetric("tuple_count", m_countMetric,
        ((Long) config.get(CONF_METRIC_LOGGING_INTERVALL)).intValue());
  }

  LOG.info("Loading SVM model...");
  Dataset dataset = Configuration.getDataSetSemEval2013();
  m_model = SerializationUtils.deserialize(dataset.getDatasetPath()
      + File.separator + SVM.SVM_MODEL_FILE_SER);

  if (m_model == null) {
    LOG.error("Could not load SVM model! File: " + dataset.getDatasetPath()
        + File.separator + SVM.SVM_MODEL_FILE_SER);
    throw new RuntimeException();
  }
}
 
开发者ID:millecker,项目名称:storm-apps,代码行数:28,代码来源:SVMBolt.java

示例7: PartitionManager

import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
public PartitionManager(DynamicPartitionConnections connections, String topologyInstanceId, ZkState state, Map stormConf, SpoutConfig spoutConfig, Partition id) {
    _partition = id;
    _connections = connections;
    _spoutConfig = spoutConfig;
    _topologyInstanceId = topologyInstanceId;
    _consumer = connections.register(id.host, id.partition);
    _state = state;
    _stormConf = stormConf;
    numberAcked = numberFailed = 0;

    String jsonTopologyId = null;
    Long jsonOffset = null;
    String path = committedPath();
    try {
        Map<Object, Object> json = _state.readJSON(path);
        LOG.info("Read partition information from: " + path +  "  --> " + json );
        if (json != null) {
            jsonTopologyId = (String) ((Map<Object, Object>) json.get("topology")).get("id");
            jsonOffset = (Long) json.get("offset");
        }
    } catch (Throwable e) {
        LOG.warn("Error reading and/or parsing at ZkNode: " + path, e);
    }

    Long currentOffset = KafkaUtils.getOffset(_consumer, spoutConfig.topic, id.partition, spoutConfig);

    if (jsonTopologyId == null || jsonOffset == null) { // failed to parse JSON?
        _committedTo = currentOffset;
        LOG.info("No partition information found, using configuration to determine offset");
    } else if (!topologyInstanceId.equals(jsonTopologyId) && spoutConfig.forceFromStart) {
        _committedTo = KafkaUtils.getOffset(_consumer, spoutConfig.topic, id.partition, spoutConfig.startOffsetTime);
        LOG.info("Topology change detected and reset from start forced, using configuration to determine offset");
    } else {
        _committedTo = jsonOffset;
        LOG.info("Read last commit offset from zookeeper: " + _committedTo + "; old topology_id: " + jsonTopologyId + " - new topology_id: " + topologyInstanceId );
    }

    if (currentOffset - _committedTo > spoutConfig.maxOffsetBehind || _committedTo <= 0) {
        LOG.info("Last commit offset from zookeeper: " + _committedTo);
        _committedTo = currentOffset;
        LOG.info("Commit offset " + _committedTo + " is more than " +
                spoutConfig.maxOffsetBehind + " behind, resetting to startOffsetTime=" + spoutConfig.startOffsetTime);
    }

    LOG.info("Starting Kafka " + _consumer.host() + ":" + id.partition + " from offset " + _committedTo);
    _emittedToOffset = _committedTo;

    _fetchAPILatencyMax = new CombinedMetric(new MaxMetric());
    _fetchAPILatencyMean = new ReducedMetric(new MeanReducer());
    _fetchAPICallCount = new CountMetric();
    _fetchAPIMessageCount = new CountMetric();
}
 
开发者ID:redBorder,项目名称:rb-bi,代码行数:53,代码来源:PartitionManager.java

示例8: execute

import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
@Override
public void execute(Tuple input) {

    // triggered by the arrival of a tuple
    // be it a tick or normal one
    flushQueues();

    if (isTickTuple(input)) {
        return;
    }

    CountMetric metric = metricGauge.scope("activethreads");
    metric.getValueAndReset();
    metric.incrBy(this.activeThreads.get());

    metric = metricGauge.scope("in queues");
    metric.getValueAndReset();
    metric.incrBy(this.fetchQueues.inQueues.get());

    metric = metricGauge.scope("queues");
    metric.getValueAndReset();
    metric.incrBy(this.fetchQueues.queues.size());

    LOG.info("[Fetcher #{}] Threads : {}\tqueues : {}\tin_queues : {}",
            taskIndex, this.activeThreads.get(),
            this.fetchQueues.queues.size(), this.fetchQueues.inQueues.get());

    if (!input.contains("url")) {
        LOG.info("[Fetcher #{}] Missing field url in tuple {}", taskIndex,
                input);
        // ignore silently
        _collector.ack(input);
        return;
    }

    String url = input.getStringByField("url");

    // has one but what about the content?
    if (StringUtils.isBlank(url)) {
        LOG.info("[Fetcher #{}] Missing value for field url in tuple {}",
                taskIndex, input);
        // ignore silently
        _collector.ack(input);
        return;
    }

    fetchQueues.addFetchItem(input);
}
 
开发者ID:zaizi,项目名称:alfresco-apache-storm-demo,代码行数:49,代码来源:FetcherBolt.java

示例9: initMetrics

import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
private void initMetrics(TopologyContext context)
{
    _countMetric = new CountMetric();
    
    context.registerMetric("record_count", _countMetric, 1);
}
 
开发者ID:dsarlis,项目名称:datix,代码行数:7,代码来源:SFlowBolt.java

示例10: PartitionManager

import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
public PartitionManager(DynamicPartitionConnections connections, String topologyInstanceId, ZkState state, Map stormConf, SpoutConfig spoutConfig, Partition id) {
    _partition = id;
    _connections = connections;
    _spoutConfig = spoutConfig;
    _topologyInstanceId = topologyInstanceId;
    _consumer = connections.register(id.host, id.partition);
    _state = state;
    _stormConf = stormConf;

    String jsonTopologyId = null;
    Long jsonOffset = null;
    String path = committedPath();
    try {
        Map<Object, Object> json = _state.readJSON(path);
        LOG.info("Read partition information from: " + path +  " --> " + json );
        if (json != null) {
            jsonTopologyId = (String) ((Map<Object, Object>) json.get("topology")).get("id");
            jsonOffset = (Long) json.get("offset");
        }
    } catch (Throwable e) {
        LOG.warn("Error reading and/or parsing at ZkNode: " + path, e);
    }

    if (jsonTopologyId == null || jsonOffset == null) { // failed to parse JSON?
        _committedTo = KafkaUtils.getOffset(_consumer, spoutConfig.topic, id.partition, spoutConfig);
        LOG.info("No partition information found, using configuration to determine offset");
    } else if (!topologyInstanceId.equals(jsonTopologyId) && spoutConfig.forceFromStart) {
        _committedTo = KafkaUtils.getOffset(_consumer, spoutConfig.topic, id.partition, spoutConfig.startOffsetTime);
        LOG.info("Topology change detected and reset from start forced, using configuration to determine offset");
    } else {
        _committedTo = jsonOffset;
        LOG.info("Read last commit offset from zookeeper: " + _committedTo + "; old topology_id: " + jsonTopologyId + " - new topology_id: " + topologyInstanceId );
    }

    LOG.info("Starting " + _partition + " from offset " + _committedTo);
    _emittedToOffset = _committedTo;

    _fetchAPILatencyMax = new CombinedMetric(new MaxMetric());
    _fetchAPILatencyMean = new ReducedMetric(new MeanReducer());
    _fetchAPICallCount = new CountMetric();
    _fetchAPIMessageCount = new CountMetric();
}
 
开发者ID:metamx,项目名称:incubator-storm,代码行数:43,代码来源:PartitionManager.java

示例11: prepare

import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
@Override
public void prepare(@SuppressWarnings("rawtypes") Map stormNativeConfig,
		TopologyContext context, OutputCollector collector) {
	this.stormNativeConfig = stormNativeConfig;
	this.collector = collector;
	originId = operationClass.getName() + "." + context.getThisTaskIndex();

	// connect to the zoopkeeper configuration
	try {
		zookeeperStormConfiguration = ZookeeperStormConfigurationFactory
				.getInstance().getStormConfiguration(stormNativeConfig);
	} catch (StormConfigurationException e) {
		logger.error("Can not connect to zookeeper for get Storm configuration. Reason: "
				+ e.getMessage());
		// create empty config to avoid errors
		zookeeperStormConfiguration = new EmptyStormConfiguration();
	}

	String msg = "SensorStormBolt instance created for";
	if (fieldGrouperId == null) {
		msg = msg + " a single_instance operation class \""
				+ operationClass.getName() + "\"";
	} else {
		msg = msg + " a fieldGrouping operation class \""
				+ operationClass.getName() + "\" grouped on tuple field \""
				+ fieldGrouperId + "\"";
	}
	if (batcherClass != null) {
		msg = msg + ", with a batcher class \"" + batcherClass.getName()
				+ "\"";
	} else {
		msg = msg + ", with no batcher.";
	}
	logger.info(msg);
	syncBuffer = new FlushingSyncBuffer(syncBufferSize);
	operationManagers = new HashMap<String, OperationManager>();
	fieldGrouperValues = new HashMap<Particle, String>();
	bufferRejectMetric = new CountMetric();
	context.registerMetric("syncbuffer_rejects", bufferRejectMetric,
			TIME_BUCKET_SIZE_IN_SECS);
}
 
开发者ID:sensorstorm,项目名称:SensorStorm,代码行数:42,代码来源:SensorStormBolt.java


注:本文中的backtype.storm.metric.api.CountMetric类示例由纯净天空整理自Github/MSDocs等开源代码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。