本文整理汇总了Python中aiokafka.producer.AIOKafkaProducer.create_batch方法的典型用法代码示例。如果您正苦于以下问题:Python AIOKafkaProducer.create_batch方法的具体用法?Python AIOKafkaProducer.create_batch怎么用?Python AIOKafkaProducer.create_batch使用的例子?那么, 这里精选的方法代码示例或许可以为您提供帮助。您也可以进一步了解该方法所在类aiokafka.producer.AIOKafkaProducer
的用法示例。
在下文中一共展示了AIOKafkaProducer.create_batch方法的4个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Python代码示例。
示例1: test_producer_transactional_flush_2_batches_before_commit
# 需要导入模块: from aiokafka.producer import AIOKafkaProducer [as 别名]
# 或者: from aiokafka.producer.AIOKafkaProducer import create_batch [as 别名]
async def test_producer_transactional_flush_2_batches_before_commit(self):
# We need to be sure if batches that are pending and batches that are
# queued will be waited. To test this we need at least 2 batches on
# the same partition. They will be sent one at a time.
producer = AIOKafkaProducer(
loop=self.loop, bootstrap_servers=self.hosts,
transactional_id="sobaka_producer", client_id="p1")
await producer.start()
self.add_cleanup(producer.stop)
await producer.begin_transaction()
batch = producer.create_batch()
batch.append(timestamp=None, key=None, value=b"1")
batch.close()
fut1 = await producer.send_batch(batch, self.topic, partition=0)
batch = producer.create_batch()
batch.append(timestamp=None, key=None, value=b"2")
batch.close()
fut2 = await producer.send_batch(batch, self.topic, partition=0)
await producer.commit_transaction()
self.assertTrue(fut1.done())
self.assertTrue(fut2.done())
示例2: test_producer_send_batch
# 需要导入模块: from aiokafka.producer import AIOKafkaProducer [as 别名]
# 或者: from aiokafka.producer.AIOKafkaProducer import create_batch [as 别名]
def test_producer_send_batch(self):
key = b'test key'
value = b'test value'
max_batch_size = 10000
producer = AIOKafkaProducer(
loop=self.loop, bootstrap_servers=self.hosts,
max_batch_size=max_batch_size)
yield from producer.start()
partitions = yield from producer.partitions_for(self.topic)
partition = partitions.pop()
# silly method to find current offset for this partition
resp = yield from producer.send_and_wait(
self.topic, value=b'discovering offset', partition=partition)
offset = resp.offset
# only fills up to its limits, then returns None
batch = producer.create_batch()
self.assertEqual(batch.record_count(), 0)
num = 0
while True:
metadata = batch.append(key=key, value=value, timestamp=None)
if metadata is None:
break
num += 1
self.assertTrue(num > 0)
self.assertEqual(batch.record_count(), num)
# batch gets properly sent
future = yield from producer.send_batch(
batch, self.topic, partition=partition)
resp = yield from future
self.assertEqual(resp.topic, self.topic)
self.assertEqual(resp.partition, partition)
self.assertEqual(resp.offset, offset + 1)
# batch accepts a too-large message if it's the first
too_large = b'm' * (max_batch_size + 1)
batch = producer.create_batch()
metadata = batch.append(key=None, value=too_large, timestamp=None)
self.assertIsNotNone(metadata)
# batch rejects a too-large message if it's not the first
batch = producer.create_batch()
batch.append(key=None, value=b"short", timestamp=None)
metadata = batch.append(key=None, value=too_large, timestamp=None)
self.assertIsNone(metadata)
yield from producer.stop()
# batch can't be sent after closing time
with self.assertRaises(ProducerClosed):
yield from producer.send_batch(
batch, self.topic, partition=partition)
示例3: test_producer_transactional_send_batch_outside_txn
# 需要导入模块: from aiokafka.producer import AIOKafkaProducer [as 别名]
# 或者: from aiokafka.producer.AIOKafkaProducer import create_batch [as 别名]
async def test_producer_transactional_send_batch_outside_txn(self):
producer = AIOKafkaProducer(
loop=self.loop, bootstrap_servers=self.hosts,
transactional_id="sobaka_producer", client_id="p1")
await producer.start()
self.add_cleanup(producer.stop)
batch = producer.create_batch()
batch.append(timestamp=None, key=None, value=b"2")
batch.close()
# Can not send if not yet in transaction
with self.assertRaises(IllegalOperation):
await producer.send_batch(batch, self.topic, partition=0)
await producer.begin_transaction()
await producer.send(self.topic, value=b"1", partition=0)
commit_task = ensure_future(producer.commit_transaction())
await asyncio.sleep(0.001, loop=self.loop)
self.assertFalse(commit_task.done())
# Already not in transaction
with self.assertRaises(IllegalOperation):
await producer.send_batch(batch, self.topic, partition=0)
await commit_task
# Transaction needs to be restarted
with self.assertRaises(IllegalOperation):
await producer.send_batch(batch, self.topic, partition=0)
示例4: test_producer_leader_change_preserves_order
# 需要导入模块: from aiokafka.producer import AIOKafkaProducer [as 别名]
# 或者: from aiokafka.producer.AIOKafkaProducer import create_batch [as 别名]
def test_producer_leader_change_preserves_order(self):
# Before 0.5.0 we did not lock partition until a response came from
# the server, but locked the node itself.
# For example: Say the sender sent a request to node 1 and before an
# failure answer came we updated metadata and leader become node 0.
# This way we may send the next batch to node 0 without waiting for
# node 1 batch to be reenqueued, resulting in out-of-order batches
producer = AIOKafkaProducer(
loop=self.loop, bootstrap_servers=self.hosts, linger_ms=1000)
yield from producer.start()
self.add_cleanup(producer.stop)
# Alter metadata to convince the producer, that leader or partition 0
# is a different node
yield from producer.partitions_for(self.topic)
topic_meta = producer._metadata._partitions[self.topic]
real_leader = topic_meta[0].leader
topic_meta[0] = topic_meta[0]._replace(leader=real_leader + 1)
# Make sure the first request for produce takes more time
original_send = producer.client.send
@asyncio.coroutine
def mocked_send(node_id, request, *args, **kw):
if node_id != real_leader and \
request.API_KEY == ProduceResponse[0].API_KEY:
yield from asyncio.sleep(2, loop=self.loop)
result = yield from original_send(node_id, request, *args, **kw)
return result
producer.client.send = mocked_send
# Send Batch 1. This will end up waiting for some time on fake leader
batch = producer.create_batch()
meta = batch.append(key=b"key", value=b"1", timestamp=None)
batch.close()
fut = yield from producer.send_batch(
batch, self.topic, partition=0)
# Make sure we sent the request
yield from asyncio.sleep(0.1, loop=self.loop)
# Update metadata to return leader to real one
yield from producer.client.force_metadata_update()
# Send Batch 2, that if it's bugged will go straight to the real node
batch2 = producer.create_batch()
meta2 = batch2.append(key=b"key", value=b"2", timestamp=None)
batch2.close()
fut2 = yield from producer.send_batch(
batch2, self.topic, partition=0)
batch_meta = yield from fut
batch_meta2 = yield from fut2
# Check the order of messages
consumer = AIOKafkaConsumer(
self.topic, loop=self.loop,
bootstrap_servers=self.hosts,
auto_offset_reset="earliest")
yield from consumer.start()
self.add_cleanup(consumer.stop)
msg = yield from consumer.getone()
self.assertEqual(msg.offset, batch_meta.offset)
self.assertEqual(msg.timestamp or -1, meta.timestamp)
self.assertEqual(msg.value, b"1")
self.assertEqual(msg.key, b"key")
msg2 = yield from consumer.getone()
self.assertEqual(msg2.offset, batch_meta2.offset)
self.assertEqual(msg2.timestamp or -1, meta2.timestamp)
self.assertEqual(msg2.value, b"2")
self.assertEqual(msg2.key, b"key")