本文整理汇总了Scala中org.apache.kafka.clients.consumer.ConsumerRecords类的典型用法代码示例。如果您正苦于以下问题:Scala ConsumerRecords类的具体用法?Scala ConsumerRecords怎么用?Scala ConsumerRecords使用的例子?那么, 这里精选的类代码示例或许可以为您提供帮助。
在下文中一共展示了ConsumerRecords类的6个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Scala代码示例。
示例1: ConsumerLoop
//设置package包名称以及导入依赖的类
package org.hpi.esb.datavalidator.consumer
import java.util.Properties
import org.apache.kafka.clients.consumer.{ConsumerConfig, ConsumerRecord, ConsumerRecords, KafkaConsumer}
import org.apache.kafka.common.TopicPartition
import org.hpi.esb.datavalidator.config.KafkaConsumerConfig
import org.hpi.esb.datavalidator.util.Logging
import scala.collection.JavaConversions._
import scala.collection.mutable.ListBuffer
class ConsumerLoop(topic: String, config: KafkaConsumerConfig, results: ListBuffer[ConsumerRecord[String, String]]) extends Runnable with Logging {
private val props = createConsumerProps()
private val consumer = new KafkaConsumer(props)
initializeConsumer()
override def run(): Unit = {
var running = true
var zeroCount = 0
while (running) {
val records = consumer.poll(1000).asInstanceOf[ConsumerRecords[String, String]]
if (records.count() == 0) {
logger.debug(s"Received 0 records from Kafka.")
zeroCount += 1
if (zeroCount == 3) {
logger.debug("Received 0 records from Kafka for the third time. We assume the stream has finished and terminate.")
running = false
}
}
for (record <- records) {
results.append(record)
}
}
consumer.close()
}
private def initializeConsumer(): Unit = {
val topicPartitions = List(new TopicPartition(topic, 0))
consumer.assign(topicPartitions)
consumer.seekToBeginning(topicPartitions)
}
private def createConsumerProps(): Properties = {
val props = new Properties()
props.put(ConsumerConfig.BOOTSTRAP_SERVERS_CONFIG, config.bootstrapServers)
props.put(ConsumerConfig.GROUP_ID_CONFIG, s"Validator")
props.put(ConsumerConfig.ENABLE_AUTO_COMMIT_CONFIG, config.autoCommit)
props.put(ConsumerConfig.AUTO_COMMIT_INTERVAL_MS_CONFIG, config.autoCommitInterval)
props.put(ConsumerConfig.SESSION_TIMEOUT_MS_CONFIG, config.sessionTimeout)
props.put(ConsumerConfig.KEY_DESERIALIZER_CLASS_CONFIG, config.keyDeserializerClass)
props.put(ConsumerConfig.VALUE_DESERIALIZER_CLASS_CONFIG, config.valueDeserializerClass)
props
}
}
示例2: TweetConsumer
//设置package包名称以及导入依赖的类
package com.knoldus.kafka
import java.util
import java.util.Properties
import com.knoldus.twitter.Tweet
import com.knoldus.utils.ConfigReader
import org.apache.kafka.clients.consumer.{ConsumerRecords, KafkaConsumer}
import scala.collection.JavaConversions._
class TweetConsumer {
def consumeTweets(groupId: String): Unit = {
val kafkaServer = ConfigReader.getKafkaServers
val kafkaTopic = ConfigReader.getKafkaTopic
val properties = new Properties()
properties.put("bootstrap.servers", kafkaServer)
properties.put("group.id", groupId)
properties.put("enable.auto.commit", "true")
properties.put("auto.commit.interval.ms", "1000")
properties.put("auto.offset.reset", "earliest")
properties.put("session.timeout.ms", "30000")
properties.put("key.deserializer", "org.apache.kafka.common.serialization.StringDeserializer")
properties.put("value.deserializer", "org.apache.kafka.common.serialization.StringDeserializer")
val kafkaConsumer = new KafkaConsumer[String, Tweet](properties)
kafkaConsumer.subscribe(util.Collections.singletonList(kafkaTopic))
while (true) {
val records: ConsumerRecords[String, Tweet] = kafkaConsumer.poll(100)
records.records(kafkaTopic).iterator().toList.foreach { record =>
println(s"Received : ${record.value()}")
}
}
}
}
示例3: ReadyKafkaConsumer
//设置package包名称以及导入依赖的类
package com.bencassedy.readykafka.consumer
import java.util
import java.util.Properties
import java.util.concurrent.Executors
import org.apache.kafka.common.serialization.StringDeserializer
import scala.collection.JavaConverters._
import org.apache.kafka.clients.consumer.{ConsumerRecord, ConsumerRecords, KafkaConsumer}
import scala.concurrent.Future
import scala.concurrent.ExecutionContext.Implicits.global
import scala.util.{Failure, Success}
class ReadyKafkaConsumer[A](
topics: List[String],
groupId: String,
brokerList: String = "localhost:9092",
msgFunc: (ConsumerRecord[String, String]) => A
) {
val consumerProperties = new Properties()
consumerProperties.put("bootstrap.servers", brokerList)
consumerProperties.put("group.id", groupId)
consumerProperties.put("auto.offset.reset", "earliest")
consumerProperties.put("key.deserializer", classOf[StringDeserializer])
consumerProperties.put("value.deserializer", classOf[StringDeserializer])
def consume(): Unit = {
val numConsumers = 2
0 until numConsumers foreach { n =>
val f = Future {
val consumer = new KafkaConsumer[String, String](consumerProperties)
consumer.subscribe(topics.asJava)
while (true) {
val crs: ConsumerRecords[String, String] = consumer.poll(1000)
crs.iterator().asScala.foreach(r => msgFunc(r))
}
}
f onComplete {
case Success(_) => println("success! consumer future shutting down")
case Failure(e) => println(s"failure! bah! exception: $e")
}
}
}
}
示例4: KafkaEventSourceTest
//设置package包名称以及导入依赖的类
package process
import java.util
import java.util.Collections
import kpi.twitter.analysis.utils.{PredictedStatus, SentimentLabel, TweetSerDe}
import org.apache.kafka.clients.consumer.{ConsumerRecord, ConsumerRecords, KafkaConsumer}
import org.apache.kafka.common.TopicPartition
import org.scalatest.FunSuite
import org.scalatest.mockito.MockitoSugar
import org.mockito.Mockito._
import org.mockito.invocation.InvocationOnMock
import org.mockito.stubbing.Answer
import twitter4j.Status
class KafkaEventSourceTest extends FunSuite with MockitoSugar {
test("subscribe should be invoked once for correct topic") {
val topicName = "fake"
val mockConsumer = mock[KafkaConsumer[SentimentLabel, Status]]
val mockTime = new MockTime
val kafkaEventSource = new KafkaEventSource(mockConsumer, topicName, mockTime)
verify(mockConsumer, times(1)).subscribe(Collections.singletonList(topicName))
}
test("poll should return on max records") {
val topicName = "fake"
val mockConsumer = mock[KafkaConsumer[SentimentLabel, Status]]
val mockTime = new MockTime
when(mockConsumer.poll(1000)).thenAnswer(new Answer[ConsumerRecords[SentimentLabel, Status]]() {
override def answer(invocation: InvocationOnMock): ConsumerRecords[SentimentLabel, Status] = {
mockTime.sleep(1)
val tp = new TopicPartition(topicName, 1)
val record = new ConsumerRecord[SentimentLabel, Status](topicName, 0, 0, mock[SentimentLabel], mock[Status])
val recordsMap = new util.HashMap[TopicPartition, util.List[ConsumerRecord[SentimentLabel, Status]]]()
val recordsList = new util.ArrayList[ConsumerRecord[SentimentLabel, Status]]()
recordsList.add(record)
recordsMap.put(tp, recordsList)
new ConsumerRecords[SentimentLabel, Status](recordsMap)
}
})
val kafkaEventSource = new KafkaEventSource(mockConsumer, topicName, mockTime)
val records = kafkaEventSource.poll(1000, 1)
assert(1 === records.size)
assert(1 === mockTime.currentMillis)
}
}
示例5: SimpleKafkaConsumer
//设置package包名称以及导入依赖的类
import java.util.Properties
import com.fasterxml.jackson.databind.KeyDeserializer
import org.apache.kafka.clients.consumer.{ConsumerRecords, KafkaConsumer}
import org.apache.kafka.common.serialization.Deserializer
import net.manub.embeddedkafka.Codecs.stringDeserializer
import net.manub.embeddedkafka.ConsumerExtensions._
class SimpleKafkaConsumer[K,V](consumerProps : Properties,
topic : String,
keyDeserializer: Deserializer[K],
valueDeserializer: Deserializer[V],
function : ConsumerRecords[K, V] => Unit,
poll : Long = 2000) {
private var running = false
private val consumer = new KafkaConsumer[K, V](consumerProps, keyDeserializer, valueDeserializer)
private val thread = new Thread {
import scala.collection.JavaConverters._
override def run: Unit = {
consumer.subscribe(List(topic).asJava)
consumer.partitionsFor(topic)
while (running) {
val record: ConsumerRecords[K, V] = consumer.poll(poll)
function(record)
}
}
}
def start(): Unit = {
if(!running) {
running = true
thread.start()
}
}
def stop(): Unit = {
if(running) {
running = false
thread.join()
consumer.close()
}
}
}
示例6: Consumer
//设置package包名称以及导入依赖的类
package com.knoldus.consumer
import java.util.{Properties, UUID}
import org.apache.kafka.clients.consumer.{ConsumerRecords, KafkaConsumer}
import org.apache.kafka.common.errors.WakeupException
import org.apache.kafka.common.serialization.StringDeserializer
import org.slf4j.LoggerFactory
import scala.collection.JavaConversions._
class Consumer(groupId: String, servers: String, topics: List[String]) {
private val timeout = 10000
val logger = LoggerFactory.getLogger(this.getClass())
private val props: Properties = new Properties
props.put("bootstrap.servers", servers)
props.put("client.id", UUID.randomUUID.toString)
props.put("group.id", groupId)
props.put("key.deserializer", classOf[StringDeserializer].getName)
props.put("value.deserializer", classOf[StringDeserializer].getName)
private val consumer = new KafkaConsumer[String, String](props)
consumer.subscribe(topics)
def read(): List[MessageFromKafka] = {
try {
logger.info("Reading from kafka queue ...... " + topics)
val consumerRecords: ConsumerRecords[String, String] = consumer.poll(timeout)
consumerRecords.map(record => MessageFromKafka(record.value())).toList
}
catch {
case wakeupException: WakeupException => {
logger.error(" Getting WakeupException ", wakeupException)
Nil
}
}
}
def close(): Unit = consumer.close()
}
case class MessageFromKafka(record: String)