You need to configure a sink connector to write records that fail into a dead letter queue topic. Requirements:
Topic name: DLQ-Topic
Headers containing error context must be added to the messagesWhich three configuration parameters are necessary?(Select three.)
You are developing a Java application using a Kafka consumer.
You need to integrate Kafka’s client logs with your own application’s logs using log4j2.
Which Java library dependency must you include in your project?
You create a topic named loT-Data with 10 partitions and replication factor of three.
A producer sends 1 MB messages compressed with Gzip.
Which two statements are true in this scenario?
(Select two.)
A producer is configured with the default partitioner. It is sending records to a topic that is configured with five partitions. The record does not contain any key.
What is the result of this?
You have a topic t1 with six partitions. You use Kafka Connect to send data from topic t1 in your Kafka cluster to Amazon S3. Kafka Connect is configured for two tasks.
How many partitions will each task process?
What are two examples of performance metrics?
(Select two.)
You want to connect with username and password to a secured Kafka cluster that has SSL encryption.
Which properties must your client include?
You are building a system for a retail store selling products to customers.
Which three datasets should you model as a GlobalKTable?
(Select three.)
You need to consume messages from Kafka using the command-line interface (CLI).
Which command should you use?
Which two statements are correct about transactions in Kafka?
(Select two.)
You are experiencing low throughput from a Java producer.
Metrics show low I/O thread ratio and low I/O thread wait ratio.
What is the most likely cause of the slow producer performance?
This schema excerpt is an example of which schema format?
package com.mycorp.mynamespace;
message SampleRecord {
int32 Stock = 1;
double Price = 2;
string Product_Name = 3;
}
Where are source connector offsets stored?
Your application is consuming from a topic configured with a deserializer.
It needs to be resilient to badly formatted records ("poison pills"). You surround the poll() call with a try/catch for RecordDeserializationException.
You need to log the bad record, skip it, and continue processing.
Which action should you take in the catch block?
An application is consuming messages from Kafka.
The application logs show that partitions are frequently being reassigned within the consumer group.
Which two factors may be contributing to this?
(Select two.)
You are composing a REST request to create a new connector in a running Connect cluster. You invoke POST /connectors with a configuration and receive a 409 (Conflict) response.
What are two reasons for this response? (Select two.)
Match each configuration parameter with the correct deployment step in installing a Kafka connector.
Which configuration allows more time for the consumer poll to process records?