Summer Special - 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: c4sdisc65

CCDAK PDF

$38.5

$109.99

3 Months Free Update

  • Printable Format
  • Value of Money
  • 100% Pass Assurance
  • Verified Answers
  • Researched by Industry Experts
  • Based on Real Exams Scenarios
  • 100% Real Questions

CCDAK PDF + Testing Engine

$61.6

$175.99

3 Months Free Update

  • Exam Name: Confluent Certified Developer for Apache Kafka Certification Examination
  • Last Update: Sep 12, 2025
  • Questions and Answers: 61
  • Free Real Questions Demo
  • Recommended by Industry Experts
  • Best Economical Package
  • Immediate Access

CCDAK Engine

$46.2

$131.99

3 Months Free Update

  • Best Testing Engine
  • One Click installation
  • Recommended by Teachers
  • Easy to use
  • 3 Modes of Learning
  • State of Art Technology
  • 100% Real Questions included

CCDAK Practice Exam Questions with Answers Confluent Certified Developer for Apache Kafka Certification Examination Certification

Question # 6

You want to connect with username and password to a secured Kafka cluster that has SSL encryption.

Which properties must your client include?

A.

security.protocol=SASL_SSL

sasl.jaas.config=org.apache.kafka.common.security.plain.PlainLoginModule required username='myUser' password='myPassword';

B.

security.protocol=SSL

sasl.jaas.config=org.apache.kafka.common.security.plain.PlainLoginModule required username='myUser' password='myPassword';

C.

security.protocol=SASL_PLAINTEXT

sasl.jaas.config=org.apache.kafka.common.security.plain.PlainLoginModule required username='myUser' password='myPassword';

D.

security.protocol=PLAINTEXT

sasl.jaas.config=org.apache.kafka.common.security.ssl.TlsLoginModule required username='myUser' password='myPassword';

Full Access
Question # 7

You create a producer that writes messages about bank account transactions from tens of thousands of different customers into a topic.

    Your consumers must process these messages with low latency and minimize consumer lag

    Processing takes ~6x longer than producing

    Transactions for each bank account must be processedin orderWhich strategy should you use?

A.

Use the timestamp of the message's arrival as its key.

B.

Use the bank account number found in the message as the message key.

C.

Use a combination of the bank account number and the transaction timestamp as the message key.

D.

Use a unique identifier such as a universally unique identifier (UUID) as the message key.

Full Access
Question # 8

What is accomplished by producing data to a topic with a message key?

A.

Messages with the same key are routed to a deterministically selected partition, enabling order guarantees within that partition.

B.

Kafka brokers allow you to add more partitions to a given topic, without impacting the data flow for existing keys.

C.

It provides a mechanism for encrypting messages at the partition level to ensure secure data transmission.

D.

Consumers can filter messages in real time based on the message key without processing unrelated messages.

Full Access
Question # 9

Your Kafka cluster has five brokers. The topic t1 on the cluster has:

    Two partitions

    Replication factor = 4

    min.insync.replicas = 3You need strong durability guarantees for messages written to topic t1.You configure a producer acks=all and all the replicas for t1 are in-sync.How many brokers need to acknowledge a message before it is considered committed?

A.

2

B.

3

C.

4

D.

5

Full Access
Question # 10

Which tool can you use to modify the replication factor of an existing topic?

A.

kafka-reassign-partitions.sh

B.

kafka-recreate-topic.sh

C.

kafka-topics.sh

D.

kafka-reassign-topics.sh

Full Access
Question # 11

A stream processing application is tracking user activity in online shopping carts.

You want to identify periods of user inactivity.

Which type of Kafka Streams window should you use?

A.

Sliding

B.

Tumbling

C.

Hopping

D.

Session

Full Access
Question # 12

A stream processing application is consuming from a topic with five partitions. You run three instances of the application. Each instance has num.stream.threads=5.

You need to identify the number of stream tasks that will be created and how many will actively consume messages from the input topic.

A.

5 created, 1 actively consuming

B.

5 created, 5 actively consuming

C.

15 created, 5 actively consuming

D.

15 created, 15 actively consuming

Full Access
Question # 13

You need to explain the best reason to implement the consumer callback interface ConsumerRebalanceListener prior to a Consumer Group Rebalance.

Which statement is correct?

A.

Partitions assigned to a consumer may change.

B.

Previous log files are deleted.

C.

Offsets are compacted.

D.

Partition leaders may change.

Full Access
Question # 14

Your application is consuming from a topic with one consumer group.

The number of running consumers is equal to the number of partitions.

Application logs show that some consumers are leaving the consumer group during peak time, triggering a rebalance. You also notice that your application is processing many duplicates.

You need to stop consumers from leaving the consumer group.

What should you do?

A.

Reduce max.poll.records property.

B.

Increase session.timeout.ms property.

C.

Add more consumer instances.

D.

Split consumers in different consumer groups.

Full Access
Question # 15

What is the default maximum size of a message the Apache Kafka broker can accept?

A.

1MB

B.

2MB

C.

5MB

D.

10MB

Full Access
Question # 16

You are writing to a topic with acks=all.

The producer receives acknowledgments but you notice duplicate messages.

You find that timeouts due to network delay are causing resends.

Which configuration should you use to prevent duplicates?

A.

enable.auto.commit=true

B.

retries=2147483647

max.in.flight.requests.per.connection=5

enable.idempotence=true

C.

retries=0

max.in.flight.requests.per.connection=5

enable.idempotence=true

D.

retries=2147483647

max.in.flight.requests.per.connection=1

enable.idempotence=false

Full Access
Question # 17

You need to correctly join data from two Kafka topics.

Which two scenarios will allow for co-partitioning?

(Select two.)

A.

Both topics have the same number of partitions.

B.

Both topics have the same key and partitioning strategy.

C.

Both topics have the same value schema.

D.

Both topics have the same retention time.

Full Access
Question # 18

What are three built-in abstractions in the Kafka Streams DSL?

(Select three.)

A.

KStream

B.

KTable

C.

GlobalKTable

D.

GlobalKStream

E.

StreamTable

Full Access