Spring Sale - Special 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: 70dumps

CCDAK Questions and Answers

Question # 6

You have a consumer group with default configuration settings reading messages from your Kafka cluster.

You need to optimize throughput so the consumer group processes more messages in the same amount of time.

Which change should you make?

A.

Remove some consumers from the consumer group.

B.

Increase the number of bytes the consumers read with each fetch request.

C.

Disable auto commit and have the consumers manually commit offsets.

D.

Decrease the session timeout of each consumer.

Full Access
Question # 7

(Your application consumes from a topic configured with a deserializer.

You want the application to be resilient to badly formatted records (poison pills).

You surround the poll() call with a try/catch block for RecordDeserializationException.

You need to log the bad record, skip it, and continue processing other records.

Which action should you take in the catch block?)

A.

Log the bad record and seek the consumer to the offset of the next record.

B.

Log the bad record and call consumer.skip() method.

C.

Throw a runtime exception to trigger a restart of the application.

D.

Log the bad record; no other action is needed.

Full Access
Question # 8

(An S3 source connector named s3-connector stopped running.

You use the Kafka Connect REST API to query the connector and task status.

One of the three tasks has failed.

You need to restart the connector and all currently running tasks.

Which REST request will restart the connector instance and all its tasks?)

A.

POST /connectors/s3-connector/restart?includeTasks=true

B.

POST /connectors/s3-connector/restart?includeTasks=true&onlyFailed=true

C.

POST /connectors/s3-connector/restart

D.

POST /connectors/s3-connector/tasks/0/restart

Full Access
Question # 9

(You are writing a producer application and need to ensure proper delivery.

You configure the producer with acks=all.

Which two actions should you take to ensure proper error handling?

Select two.)

A.

Check the value of ProducerRecord.status().

B.

Use a callback argument in producer.send() where you check delivery status.

C.

Check that producer.send() returned a RecordMetadata object and is not null.

D.

Surround the call to producer.send() with a try/catch block to catch KafkaException.

Full Access
Question # 10

Which two statements are correct about transactions in Kafka?

(Select two.)

A.

All messages from a failed transaction will be deleted from a Kafka topic.

B.

Transactions are only possible when writing messages to a topic with single partition.

C.

Consumers can consume both committed and uncommitted transactions.

D.

Information about producers and their transactions is stored in the _transaction_state topic.

E.

Transactions guarantee at least once delivery of messages.

Full Access
Question # 11

Your application is consuming from a topic with one consumer group.

The number of running consumers is equal to the number of partitions.

Application logs show that some consumers are leaving the consumer group during peak time, triggering a rebalance. You also notice that your application is processing many duplicates.

You need to stop consumers from leaving the consumer group.

What should you do?

A.

Reduce max.poll.records property.

B.

Increase session.timeout.ms property.

C.

Add more consumer instances.

D.

Split consumers in different consumer groups.

Full Access
Question # 12

(A stream processing application tracks user activity in online shopping carts, including items added, removed, and ordered throughout the day for each user.

You need to capture data to identify possible periods of user inactivity.

Which type of Kafka Streams window should you use?)

A.

Session

B.

Hopping

C.

Tumbling

D.

Sliding

Full Access
Question # 13

You are building a system for a retail store selling products to customers.

Which three datasets should you model as a GlobalKTable?

(Select three.)

A.

Inventory of products at a warehouse

B.

All purchases at a retail store occurring in real time

C.

Customer profile information

D.

Log of payment transactions

E.

Catalog of products

Full Access
Question # 14

(You are building real-time streaming applications using Kafka Streams.

Your application has a custom transformation.

You need to define custom processors in Kafka Streams.

Which tool should you use?)

A.

TopologyTestDriver

B.

Processor API

C.

Kafka Streams Domain Specific Language (DSL)

D.

Kafka Streams Custom Transformation Language

Full Access
Question # 15

What is the default maximum size of a message the Apache Kafka broker can accept?

A.

1MB

B.

2MB

C.

5MB

D.

10MB

Full Access
Question # 16

(You create a topic with five partitions.

What can you assume about messages read from that topic by a single consumer group?)

A.

Messages can be consumed by a maximum of five consumers in the same consumer group.

B.

The consumer group can only read the same number of messages from all the partitions.

C.

All messages will be read from exactly one broker by the consumer group.

D.

Messages from one partition can be consumed by any of the consumers in a group for faster processing.

Full Access
Question # 17

What is a consequence of increasing the number of partitions in an existing Kafka topic?

A.

Existing data will be redistributed across the new number of partitions temporarily increasing cluster load.

B.

Records with the same key could be located in different partitions.

C.

Consumers will need to process data from more partitions which will significantly increase consumer lag.

D.

The acknowledgment process will increase latency for producers using acks=all.

Full Access
Question # 18

You have a topic with four partitions. The application reads from it using two consumers in a single consumer group.

Processing is CPU-bound, and lag is increasing.

What should you do?

A.

Add more consumers to increase the level of parallelism of the processing.

B.

Add more partitions to the topic to increase the level of parallelism of the processing.

C.

Increase the max.poll.records property of consumers.

D.

Decrease the max.poll.records property of consumers.

Full Access
Question # 19

Where are source connector offsets stored?

A.

offset.storage.topic

B.

storage.offset.topic

C.

topic.offset.config

D.

offset, storage, partitions

Full Access
Question # 20

(Which configuration is valid for deploying a JDBC Source Connector to read all rows from the orders table and write them to the dbl-orders topic?)

A.

{"name": "orders-connect","connector.class": "io.confluent.connect.jdbc.DdbcSourceConnector","tasks.max": "1","connection.url": "jdbc:mysql://mysql:3306/dbl","topic.whitelist": "orders","auto.create": "true"}

B.

{"name": "dbl-orders","connector.class": "io.confluent.connect.jdbc.DdbcSourceConnector","tasks.max": "1","connection.url": "jdbc:mysql://mysql:3306/dbl?user=user&password=pas","topic.prefix": "dbl-","table.blacklist": "ord*"}

C.

{"name": "jdbc-source","connector.class": "io.confluent.connect.jdbc.DdbcSourceConnector","tasks.max": "1","connection.url": "jdbc:mysql://mysql:3306/dbl?user=user&useAutoAuth=true","topic.prefix": "dbl-","table.whitelist": "orders"}

D.

{"name": "jdbc-source","connector.class": "io.confluent.connect.jdbc.JdbcSourceConnector","tasks.max": "1","connection.url": "jdbc:mysql://mysql:3306/dbl?user=user&password=pas","topic.prefix": "dbl-","table.whitelist": "orders"}

Full Access
Question # 21

(A consumer application needs to use an at-most-once delivery semantic.

What is the best consumer configuration and code skeleton to avoid duplicate messages being read?)

A.

auto.offset.reset=latest and enable.auto.commit=truewhile (true) {final var records = consumer.poll(POLL_TIMEOUT);for (var record : records) {// Any processing}consumer.commitAsync();}

B.

auto.offset.reset=earliest and enable.auto.commit=falsewhile (true) {final var records = consumer.poll(POLL_TIMEOUT);consumer.commitAsync();for (var record : records) {// Any processing}}

C.

auto.offset.reset=earliest and enable.auto.commit=falsewhile (true) {final var records = consumer.poll(POLL_TIMEOUT);for (var record : records) {// Any processing}consumer.commitAsync();}

D.

auto.offset.reset=earliest and enable.auto.commit=truewhile (true) {final var records = consumer.poll(POLL_TIMEOUT);consumer.commitAsync();for (var record : records) {// Any processing}}

Full Access
Question # 22

(You have a topic with four partitions. The application reading this topic is using a consumer group with two consumers.

Throughput is smoothly distributed among partitions, but application lag is increasing.

Application monitoring shows that message processing is consuming all available CPU resources.

Which action should you take to resolve this issue?)

A.

Add more partitions to the topic to increase the level of parallelism of the processing.

B.

Increase the max.poll.records property of consumers.

C.

Add more consumers to increase the level of parallelism of the processing.

D.

Decrease the max.poll.records property of consumers.

Full Access
Question # 23

(Your configuration parameters for a Source connector and Connect worker are:

• offset.flush.interval.ms=60000

• offset.flush.timeout.ms=500

• offset.storage.topic=connect-offsets

• offset.storage.replication.factor=-1

Which two statements match the expected behavior?

Select two.)

A.

The offsets topic will use the broker default replication factor.

B.

The connector will commit offsets to the broker default offsets topic.

C.

The connector will commit offsets to a topic called connect-offsets.

D.

The connector will wait 500 ms before trying to commit offsets for tasks.

Full Access
Question # 24

You are working on a Kafka cluster with three nodes. You create a topic named orders with:

replication.factor = 3

min.insync.replicas = 2

acks = allWhat exception will be generated if two brokers are down due to network delay?

A.

NotEnoughReplicasException

B.

NetworkException

C.

NotCoordinatorException

D.

NotLeaderForPartitionException

Full Access
Question # 25

You need to consume messages from Kafka using the command-line interface (CLI).

Which command should you use?

A.

kafka-console-consumer

B.

kafka-consumer

C.

kafka-get-messages

D.

kafka-consume

Full Access
Question # 26

(You have a Kafka consumer in production actively reading from a critical topic from which multiple other applications are consuming.

You have a new requirement to update the offset of your consumer to start reading from the beginning of the topic.

Which action would you take?)

A.

Update the consumer group’s offset to the earliest position using the kafka-consumer-groups CLI tool.

B.

Temporarily configure the topic’s retention.ms parameter to 0 to empty the topic.

C.

Start a new consumer application with the same consumer group id.

D.

Update the consumer configuration by setting the auto.offset.reset property to earliest.

Full Access
Question # 27

Which two producer exceptions are examples of the class RetriableException? (Select two.)

A.

LeaderNotAvailableException

B.

RecordTooLargeException

C.

AuthorizationException

D.

NotEnoughReplicasException

Full Access