Skip to content

tombentley/klog

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

19 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Klog - analyse kafka segment dumps

What is it?

A tool to analyse dumps of Kafka log segments and producer snapshots, as produced by kafka-dump-logs.sh. This is complementary to the official tools already provided by the Apache Kafka distribution.

Building

You can build a native executable:

./mvnw package -Pnative

Use the following command if you have Docker but not GraalVM:

./mvnw clean package -DskipTests -Pnative -Dquarkus.native.container-build=true

Installing

You can copy the executable to some directory in your $PATH, for example:

sudo cp target/klog-*-SNAPSHOT-runner ~/usr/bin/klog

Alternatively you can use alias in your shell session:

alias klog=target/klog-*-SNAPSHOT-runner

Or add it to your .bashrc to make it permanent:

echo alias klog=target/klog-*-runner >> ~/.bashrc

Usage

There are two main subcommands named segment and snapshot.

Segment

The segment subcommand which itself takes two subcommands.

cat

This will echo a dumped segment file (or files) to standard output.

klog segment cat 00000000000002226093.log.dump

Example output:

Batch(baseOffset=2253037, lastOffset=2253037, count=1, baseSequence=143, lastSequence=143, producerId=895510, producerEpoch=1, partitionLeaderEpoch=539, isTransactional=true, isControl=false, position=37567204, createTime=2021-09-09T08:45:55.998Z, size=4526, magic=2, compressCodec='ZSTD', crc=-1607451119, isValid=false)
  DataMessage(offset=2253037, createTime=2021-09-09T08:45:55.998Z, keySize=52, valueSize=184452, sequence=143, headerKeys='foo,bar')
Batch(baseOffset=2253038, lastOffset=2253038, count=1, baseSequence=-1, lastSequence=-1, producerId=895510, producerEpoch=1, partitionLeaderEpoch=539, isTransactional=true, isControl=true, position=37571730, createTime=2021-09-09T08:45:56.044Z, size=78, magic=2, compressCodec='NONE', crc=-170113429, isValid=false)
  ControlMessage(offset=2253038, createTime=2021-09-09T08:45:56.044Z, keySize=4, valueSize=6, sequence=-1, headers='', commit=true, coordinatorEpoch=887)

There's not much value in this above regular cat except it will interpret all timestamps in a human readable way. (Plain segment dump represent these as a millisecond offset since the UNIX epoch) and colourise the output.

Filtering options are supported: --pid, --producer-epoch, leader-epoch and, for __transaction_state dumps, --transactional-id. When multiple options are present a batch or message must support all the filters to be included in the output.

txn-stat

This will report a statistics transactions in the given segment dumps of normal partitions.

klog segment txn-stat *.log.dump

Example output:

num_committed: 12683
num_aborted: 2
txn_size_stats: IntSummaryStatistics{count=12683, sum=12772, min=1, average=1.007017, max=6}
txn_duration_stats_ms: IntSummaryStatistics{count=12683, sum=643672, min=11, average=50.750769, max=32189}
empty_txn: EmptyTransaction[closingBatch=Batch(baseOffset=2241851, lastOffset=2241851, count=1, baseSequence=-1, lastSequence=-1, producerId=895428, producerEpoch=12, partitionLeaderEpoch=531, isTransactional=true, isControl=true, position=19143380, createTime=2021-09-06T07:47:42.540Z, size=78, magic=2, compressCodec='NONE', crc=-1540206536, isValid=true), controlMessage=ControlMessage(offset=2241851, createTime=2021-09-06T07:47:42.540Z, keySize=4, valueSize=6, sequence=-1, headers='', commit=false, coordinatorEpoch=612)]
empty_txn: EmptyTransaction[closingBatch=Batch(baseOffset=2250125, lastOffset=2250125, count=1, baseSequence=-1, lastSequence=-1, producerId=894436, producerEpoch=4, partitionLeaderEpoch=534, isTransactional=true, isControl=true, position=32948458, createTime=2021-09-08T13:22:27.087Z, size=78, magic=2, compressCodec='NONE', crc=448547950, isValid=true), controlMessage=ControlMessage(offset=2250125, createTime=2021-09-08T13:22:27.087Z, keySize=4, valueSize=6, sequence=-1, headers='', commit=false, coordinatorEpoch=911)]
open_txn: ProducerSession[producerId=894436, producerEpoch=4]->FirstBatchInTxn[firstBatchInTxn=Batch(baseOffset=2250126, lastOffset=2250126, count=1, baseSequence=660, lastSequence=660, producerId=894436, producerEpoch=4, partitionLeaderEpoch=534, isTransactional=true, isControl=false, position=32948536, createTime=2021-09-08T13:20:26.964Z, size=6764, magic=2, compressCodec='ZSTD', crc=-1999558231, isValid=true), numDataBatches=1]

Currently, this includes:

  • num_committed The number of transactional commits.
  • num_aborted The number of transactional aborts.
  • txn_size_stats Some stats about the number of data batches in each transaction.
  • txn_duraction_stats_ms Some stats about the duration of transactions.
  • empty_txn (multiple occurrences) Info about each empty transaction in the log. An empty transaction is one where, for a given producer session (identified by a PID and producer epoch), a transaction is ended by a control batch without the previous batch for that session being a data batch.
  • open_txn (multiple occurrences) Info about any open transactions in the log An open transaction is one where there's a data batch for a given producer session that's not followed by a control batch.

As for klog segment cat filtering options are supported: --pid, --producer-epoch, leader-epoch.

Snapshot

cat

This will echo a dumped snapshot file (or files) to standard output:

klog snapshot cat 00000000000933607637.snapshot.dump

Example output:

ProducerState(producerId=171101, producerEpoch=12, coordinatorEpoch=54, currentTxnFirstOffset=0, lastTimestamp=1970-01-01T00:00:00Z, firstSequence=0, lastSequence=0, lastOffset=932442537, offsetDelta=0, timestamp=2022-06-19T08:40:30.307Z)
ProducerState(producerId=199398, producerEpoch=0, coordinatorEpoch=57, currentTxnFirstOffset=933607621, lastTimestamp=1970-01-01T00:00:00Z, firstSequence=0, lastSequence=0, lastOffset=933607621, offsetDelta=0, timestamp=2022-06-20T21:41:07.833Z)
ProducerState(producerId=173102, producerEpoch=16, coordinatorEpoch=59, currentTxnFirstOffset=0, lastTimestamp=1970-01-01T00:00:00Z, firstSequence=0, lastSequence=0, lastOffset=933203854, offsetDelta=0, timestamp=2022-06-20T05:50:54.875Z)

Filtering options are supported: --pid, --producer-epoch. When multiple options are present a batch or message must support all the filters to be included in the output.

abort-cmd

This will emit the kafka-transactions.sh command to use to abort the transaction.

klog snapshot abort-cmd 00000000000933607637.snapshot.dump --pid 173101 --producer-epoch 14

Example output:

KAFKA_HOME/bin/kafka-transactions.sh --bootstrap-server $BOOTSTRAP_URL abort --topic $TOPIC_NAME --partition $PART_NUM --producer-id 173101 --producer-epoch 14 --coordinator-epoch 53

Post-mortem analysis of hanging transactions

Confirm that read_committed consumer groups are stuck on one or more topic partitions and their lag is growing by using kafka-consumer-groups.sh.

Get the raw partition folders and dump all segments containing the stuck last stable offset (LSO) using kafka-dump-log.sh. It is required to run this command inside the partition directory containing the row segments, as shown in the following example:

cd /path/to/my-topic-9
kafka-dump-log.sh --deep-iteration --files 00000000000912375285.log > 00000000000912375285.log.dump

Find the hanging transactions running klog segment txn-stat on these dumps to find the PID and producer epoch of the session lacking a control record.

If the partition is stuck but there is no open_txn found, it probably means that the retention policy has already kicked in, i.e. the log segments in which data records were appended in a transaction which lacked a following end marker have been deleted, but the producer snapshot retains knowledge of their having existed. In this case you can simply delete all .snapshot files from partition folder in all brokers and do a rolling restart. That way, the in-memory PID map will be recreated by scanning the full log and will have no memory of the hanging transaction.

If open_txn are found, use klog snapshot abort-cmd to get the abort transaction command to run.

Development

This project uses Quarkus.

Running the application in dev mode

You can run your application in dev mode that enables live coding using:

./mvnw compile quarkus:dev

To seed the command line arguments, pass the -Dquarkus.args option:

./mvnw compile quarkus:dev -Dquarkus.args='patch get connectors'

In dev mode, remote debuggers can connect to the running application on port 5005. In order to wait for a debugger to connect, pass the -Dsuspend option.

Packaging and running the application

The application can be packaged using:

./mvnw package

It produces the quarkus-run.jar file in the target/quarkus-app/ directory. Be aware that it's not an über-jar as the dependencies are copied into the target/quarkus-app/lib/ directory.

The application is now runnable using java -jar target/quarkus-app/quarkus-run.jar.

About

Analyse Kafka log dumps

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages