kafkajs
Version:
A modern Apache Kafka client for node.js
66 lines (57 loc) • 2.12 kB
JavaScript
const Decoder = require('../../../decoder')
const { KafkaJSDeleteTopicRecordsError } = require('../../../../errors')
const { failure, createErrorFromCode } = require('../../../error')
/**
* DeleteRecords Response (Version: 0) => throttle_time_ms [topics]
* throttle_time_ms => INT32
* topics => name [partitions]
* name => STRING
* partitions => partition low_watermark error_code
* partition => INT32
* low_watermark => INT64
* error_code => INT16
*/
const topicNameComparator = (a, b) => a.topic.localeCompare(b.topic)
const decode = async rawData => {
const decoder = new Decoder(rawData)
return {
throttleTime: decoder.readInt32(),
topics: decoder
.readArray(decoder => ({
topic: decoder.readString(),
partitions: decoder.readArray(decoder => ({
partition: decoder.readInt32(),
lowWatermark: decoder.readInt64(),
errorCode: decoder.readInt16(),
})),
}))
.sort(topicNameComparator),
}
}
const parse = requestTopics => async data => {
const topicsWithErrors = data.topics
.map(({ partitions }) => ({
partitionsWithErrors: partitions.filter(({ errorCode }) => failure(errorCode)),
}))
.filter(({ partitionsWithErrors }) => partitionsWithErrors.length)
if (topicsWithErrors.length > 0) {
// at present we only ever request one topic at a time, so can destructure the arrays
const [{ topic }] = data.topics // topic name
const [{ partitions: requestPartitions }] = requestTopics // requested offset(s)
const [{ partitionsWithErrors }] = topicsWithErrors // partition(s) + error(s)
throw new KafkaJSDeleteTopicRecordsError({
topic,
partitions: partitionsWithErrors.map(({ partition, errorCode }) => ({
partition,
error: createErrorFromCode(errorCode),
// attach the original offset from the request, onto the error response
offset: requestPartitions.find(p => p.partition === partition).offset,
})),
})
}
return data
}
module.exports = ({ topics }) => ({
decode,
parse: parse(topics),
})