org.apache.kafka.common.requests.TxnOffsetCommitResponse Maven / Gradle / Ivy
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.kafka.common.requests;
import org.apache.kafka.common.TopicPartition;
import org.apache.kafka.common.message.TxnOffsetCommitResponseData;
import org.apache.kafka.common.message.TxnOffsetCommitResponseData.TxnOffsetCommitResponsePartition;
import org.apache.kafka.common.message.TxnOffsetCommitResponseData.TxnOffsetCommitResponseTopic;
import org.apache.kafka.common.protocol.ApiKeys;
import org.apache.kafka.common.protocol.ByteBufferAccessor;
import org.apache.kafka.common.protocol.Errors;
import java.nio.ByteBuffer;
import java.util.ArrayList;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
import java.util.function.Function;
/**
* Possible error codes:
*
* - {@link Errors#INVALID_PRODUCER_EPOCH}
* - {@link Errors#NOT_COORDINATOR}
* - {@link Errors#COORDINATOR_NOT_AVAILABLE}
* - {@link Errors#COORDINATOR_LOAD_IN_PROGRESS}
* - {@link Errors#OFFSET_METADATA_TOO_LARGE}
* - {@link Errors#GROUP_AUTHORIZATION_FAILED}
* - {@link Errors#INVALID_COMMIT_OFFSET_SIZE}
* - {@link Errors#TRANSACTIONAL_ID_AUTHORIZATION_FAILED}
* - {@link Errors#UNSUPPORTED_FOR_MESSAGE_FORMAT}
* - {@link Errors#REQUEST_TIMED_OUT}
* - {@link Errors#UNKNOWN_MEMBER_ID}
* - {@link Errors#FENCED_INSTANCE_ID}
* - {@link Errors#ILLEGAL_GENERATION}
*/
public class TxnOffsetCommitResponse extends AbstractResponse {
public static class Builder {
TxnOffsetCommitResponseData data = new TxnOffsetCommitResponseData();
HashMap byTopicName = new HashMap<>();
private TxnOffsetCommitResponseTopic getOrCreateTopic(
String topicName
) {
TxnOffsetCommitResponseTopic topic = byTopicName.get(topicName);
if (topic == null) {
topic = new TxnOffsetCommitResponseTopic().setName(topicName);
data.topics().add(topic);
byTopicName.put(topicName, topic);
}
return topic;
}
public Builder addPartition(
String topicName,
int partitionIndex,
Errors error
) {
final TxnOffsetCommitResponseTopic topicResponse = getOrCreateTopic(topicName);
topicResponse.partitions().add(new TxnOffsetCommitResponsePartition()
.setPartitionIndex(partitionIndex)
.setErrorCode(error.code()));
return this;
}
public Builder addPartitions(
String topicName,
List
partitions,
Function
partitionIndex,
Errors error
) {
final TxnOffsetCommitResponseTopic topicResponse = getOrCreateTopic(topicName);
partitions.forEach(partition -> {
topicResponse.partitions().add(new TxnOffsetCommitResponsePartition()
.setPartitionIndex(partitionIndex.apply(partition))
.setErrorCode(error.code()));
});
return this;
}
public Builder merge(
TxnOffsetCommitResponseData newData
) {
if (data.topics().isEmpty()) {
// If the current data is empty, we can discard it and use the new data.
data = newData;
} else {
// Otherwise, we have to merge them together.
newData.topics().forEach(newTopic -> {
TxnOffsetCommitResponseTopic existingTopic = byTopicName.get(newTopic.name());
if (existingTopic == null) {
// If no topic exists, we can directly copy the new topic data.
data.topics().add(newTopic);
byTopicName.put(newTopic.name(), newTopic);
} else {
// Otherwise, we add the partitions to the existing one. Note we
// expect non-overlapping partitions here as we don't verify
// if the partition is already in the list before adding it.
existingTopic.partitions().addAll(newTopic.partitions());
}
});
}
return this;
}
public TxnOffsetCommitResponse build() {
return new TxnOffsetCommitResponse(data);
}
}
private final TxnOffsetCommitResponseData data;
public TxnOffsetCommitResponse(TxnOffsetCommitResponseData data) {
super(ApiKeys.TXN_OFFSET_COMMIT);
this.data = data;
}
public TxnOffsetCommitResponse(int requestThrottleMs, Map responseData) {
super(ApiKeys.TXN_OFFSET_COMMIT);
Map responseTopicDataMap = new HashMap<>();
for (Map.Entry entry : responseData.entrySet()) {
TopicPartition topicPartition = entry.getKey();
String topicName = topicPartition.topic();
TxnOffsetCommitResponseTopic topic = responseTopicDataMap.getOrDefault(
topicName, new TxnOffsetCommitResponseTopic().setName(topicName));
topic.partitions().add(new TxnOffsetCommitResponsePartition()
.setErrorCode(entry.getValue().code())
.setPartitionIndex(topicPartition.partition())
);
responseTopicDataMap.put(topicName, topic);
}
data = new TxnOffsetCommitResponseData()
.setTopics(new ArrayList<>(responseTopicDataMap.values()))
.setThrottleTimeMs(requestThrottleMs);
}
@Override
public TxnOffsetCommitResponseData data() {
return data;
}
@Override
public int throttleTimeMs() {
return data.throttleTimeMs();
}
@Override
public void maybeSetThrottleTimeMs(int throttleTimeMs) {
data.setThrottleTimeMs(throttleTimeMs);
}
@Override
public Map errorCounts() {
return errorCounts(data.topics().stream().flatMap(topic ->
topic.partitions().stream().map(partition ->
Errors.forCode(partition.errorCode()))));
}
public Map errors() {
Map errorMap = new HashMap<>();
for (TxnOffsetCommitResponseTopic topic : data.topics()) {
for (TxnOffsetCommitResponsePartition partition : topic.partitions()) {
errorMap.put(new TopicPartition(topic.name(), partition.partitionIndex()),
Errors.forCode(partition.errorCode()));
}
}
return errorMap;
}
public static TxnOffsetCommitResponse parse(ByteBuffer buffer, short version) {
return new TxnOffsetCommitResponse(new TxnOffsetCommitResponseData(new ByteBufferAccessor(buffer), version));
}
@Override
public String toString() {
return data.toString();
}
@Override
public boolean shouldClientThrottle(short version) {
return version >= 1;
}
}