org.apache.kafka.streams.kstream.internals.KTableKTableJoinMerger Maven / Gradle / Ivy
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.kafka.streams.kstream.internals;
import org.apache.kafka.streams.state.TimestampedKeyValueStore;
import org.apache.kafka.streams.state.ValueAndTimestamp;
import java.util.Collections;
import java.util.HashSet;
import java.util.Set;
@SuppressWarnings("deprecation") // Old PAPI. Needs to be migrated.
public class KTableKTableJoinMerger implements KTableProcessorSupplier {
private final KTableProcessorSupplier parent1;
private final KTableProcessorSupplier parent2;
private final String queryableName;
private boolean sendOldValues = false;
KTableKTableJoinMerger(final KTableProcessorSupplier parent1,
final KTableProcessorSupplier parent2,
final String queryableName) {
this.parent1 = parent1;
this.parent2 = parent2;
this.queryableName = queryableName;
}
public String getQueryableName() {
return queryableName;
}
@Override
public org.apache.kafka.streams.processor.Processor> get() {
return new KTableKTableJoinMergeProcessor();
}
@Override
public KTableValueGetterSupplier view() {
// if the result KTable is materialized, use the materialized store to return getter value;
// otherwise rely on the parent getter and apply join on-the-fly
if (queryableName != null) {
return new KTableMaterializedValueGetterSupplier<>(queryableName);
} else {
return new KTableValueGetterSupplier() {
public KTableValueGetter get() {
return parent1.view().get();
}
@Override
public String[] storeNames() {
final String[] storeNames1 = parent1.view().storeNames();
final String[] storeNames2 = parent2.view().storeNames();
final Set stores = new HashSet<>(storeNames1.length + storeNames2.length);
Collections.addAll(stores, storeNames1);
Collections.addAll(stores, storeNames2);
return stores.toArray(new String[0]);
}
};
}
}
@Override
public boolean enableSendingOldValues(final boolean forceMaterialization) {
// Table-table joins require upstream materialization:
parent1.enableSendingOldValues(true);
parent2.enableSendingOldValues(true);
sendOldValues = true;
return true;
}
public static KTableKTableJoinMerger of(final KTableProcessorSupplier parent1,
final KTableProcessorSupplier parent2) {
return of(parent1, parent2, null);
}
public static KTableKTableJoinMerger of(final KTableProcessorSupplier parent1,
final KTableProcessorSupplier parent2,
final String queryableName) {
return new KTableKTableJoinMerger<>(parent1, parent2, queryableName);
}
private class KTableKTableJoinMergeProcessor extends org.apache.kafka.streams.processor.AbstractProcessor> {
private TimestampedKeyValueStore store;
private TimestampedTupleForwarder tupleForwarder;
@SuppressWarnings("unchecked")
@Override
public void init(final org.apache.kafka.streams.processor.ProcessorContext context) {
super.init(context);
if (queryableName != null) {
store = (TimestampedKeyValueStore) context.getStateStore(queryableName);
tupleForwarder = new TimestampedTupleForwarder<>(
store,
context,
new TimestampedCacheFlushListener<>(context),
sendOldValues);
}
}
@Override
public void process(final K key, final Change value) {
if (queryableName != null) {
store.put(key, ValueAndTimestamp.make(value.newValue, context().timestamp()));
tupleForwarder.maybeForward(key, value.newValue, sendOldValues ? value.oldValue : null);
} else {
if (sendOldValues) {
context().forward(key, value);
} else {
context().forward(key, new Change<>(value.newValue, null));
}
}
}
}
}