
org.elasticsearch.action.fieldcaps.TransportFieldCapabilitiesAction Maven / Gradle / Ivy
/*
* Copyright Elasticsearch B.V. and/or licensed to Elasticsearch B.V. under one
* or more contributor license agreements. Licensed under the Elastic License
* 2.0 and the Server Side Public License, v 1; you may not use this file except
* in compliance with, at your election, the Elastic License 2.0 or the Server
* Side Public License, v 1.
*/
package org.elasticsearch.action.fieldcaps;
import org.apache.lucene.util.ArrayUtil;
import org.elasticsearch.ExceptionsHelper;
import org.elasticsearch.action.ActionListener;
import org.elasticsearch.action.ActionRunnable;
import org.elasticsearch.action.OriginalIndices;
import org.elasticsearch.action.support.ActionFilters;
import org.elasticsearch.action.support.ChannelActionListener;
import org.elasticsearch.action.support.HandledTransportAction;
import org.elasticsearch.action.support.RefCountingRunnable;
import org.elasticsearch.client.internal.Client;
import org.elasticsearch.cluster.ClusterState;
import org.elasticsearch.cluster.block.ClusterBlockLevel;
import org.elasticsearch.cluster.metadata.IndexNameExpressionResolver;
import org.elasticsearch.cluster.service.ClusterService;
import org.elasticsearch.common.Strings;
import org.elasticsearch.common.inject.Inject;
import org.elasticsearch.common.util.set.Sets;
import org.elasticsearch.core.Tuple;
import org.elasticsearch.index.shard.ShardId;
import org.elasticsearch.indices.IndicesService;
import org.elasticsearch.logging.LogManager;
import org.elasticsearch.logging.Logger;
import org.elasticsearch.search.SearchService;
import org.elasticsearch.tasks.CancellableTask;
import org.elasticsearch.tasks.Task;
import org.elasticsearch.threadpool.ThreadPool;
import org.elasticsearch.transport.RemoteClusterAware;
import org.elasticsearch.transport.TransportChannel;
import org.elasticsearch.transport.TransportRequestHandler;
import org.elasticsearch.transport.TransportService;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.Collections;
import java.util.Comparator;
import java.util.HashMap;
import java.util.HashSet;
import java.util.List;
import java.util.Map;
import java.util.Optional;
import java.util.Set;
import java.util.concurrent.atomic.AtomicBoolean;
import java.util.function.BiConsumer;
import java.util.function.Consumer;
import java.util.function.Function;
import java.util.stream.Collectors;
import java.util.stream.Stream;
import static org.elasticsearch.action.search.TransportSearchHelper.checkCCSVersionCompatibility;
public class TransportFieldCapabilitiesAction extends HandledTransportAction {
public static final String ACTION_NODE_NAME = FieldCapabilitiesAction.NAME + "[n]";
public static final Logger LOGGER = LogManager.getLogger(TransportFieldCapabilitiesAction.class);
private final ThreadPool threadPool;
private final TransportService transportService;
private final ClusterService clusterService;
private final IndexNameExpressionResolver indexNameExpressionResolver;
private final IndicesService indicesService;
private final boolean ccsCheckCompatibility;
@Inject
public TransportFieldCapabilitiesAction(
TransportService transportService,
ClusterService clusterService,
ThreadPool threadPool,
ActionFilters actionFilters,
IndicesService indicesService,
IndexNameExpressionResolver indexNameExpressionResolver
) {
super(FieldCapabilitiesAction.NAME, transportService, actionFilters, FieldCapabilitiesRequest::new);
this.threadPool = threadPool;
this.transportService = transportService;
this.clusterService = clusterService;
this.indexNameExpressionResolver = indexNameExpressionResolver;
this.indicesService = indicesService;
transportService.registerRequestHandler(
ACTION_NODE_NAME,
ThreadPool.Names.SEARCH_COORDINATION,
FieldCapabilitiesNodeRequest::new,
new NodeTransportHandler()
);
this.ccsCheckCompatibility = SearchService.CCS_VERSION_CHECK_SETTING.get(clusterService.getSettings());
}
@Override
protected void doExecute(Task task, FieldCapabilitiesRequest request, final ActionListener listener) {
if (ccsCheckCompatibility) {
checkCCSVersionCompatibility(request);
}
assert task instanceof CancellableTask;
final CancellableTask fieldCapTask = (CancellableTask) task;
// retrieve the initial timestamp in case the action is a cross cluster search
long nowInMillis = request.nowInMillis() == null ? System.currentTimeMillis() : request.nowInMillis();
final ClusterState clusterState = clusterService.state();
final Map remoteClusterIndices = transportService.getRemoteClusterService()
.groupIndices(request.indicesOptions(), request.indices());
final OriginalIndices localIndices = remoteClusterIndices.remove(RemoteClusterAware.LOCAL_CLUSTER_GROUP_KEY);
final String[] concreteIndices;
if (localIndices == null) {
// in the case we have one or more remote indices but no local we don't expand to all local indices and just do remote indices
concreteIndices = Strings.EMPTY_ARRAY;
} else {
concreteIndices = indexNameExpressionResolver.concreteIndexNames(clusterState, localIndices);
}
if (concreteIndices.length == 0 && remoteClusterIndices.isEmpty()) {
listener.onResponse(new FieldCapabilitiesResponse(new String[0], Collections.emptyMap()));
return;
}
checkIndexBlocks(clusterState, concreteIndices);
final FailureCollector indexFailures = new FailureCollector();
final Map indexResponses = Collections.synchronizedMap(new HashMap<>());
// This map is used to share the index response for indices which have the same index mapping hash to reduce the memory usage.
final Map indexMappingHashToResponses = Collections.synchronizedMap(new HashMap<>());
final Runnable releaseResourcesOnCancel = () -> {
LOGGER.trace("clear index responses on cancellation");
indexFailures.clear();
indexResponses.clear();
indexMappingHashToResponses.clear();
};
final Consumer handleIndexResponse = resp -> {
if (fieldCapTask.isCancelled()) {
releaseResourcesOnCancel.run();
return;
}
if (resp.canMatch() && resp.getIndexMappingHash() != null) {
FieldCapabilitiesIndexResponse curr = indexMappingHashToResponses.putIfAbsent(resp.getIndexMappingHash(), resp);
if (curr != null) {
resp = new FieldCapabilitiesIndexResponse(resp.getIndexName(), curr.getIndexMappingHash(), curr.get(), true);
}
}
indexResponses.putIfAbsent(resp.getIndexName(), resp);
if (fieldCapTask.isCancelled()) {
releaseResourcesOnCancel.run();
}
};
final BiConsumer handleIndexFailure = (index, error) -> {
if (fieldCapTask.isCancelled()) {
releaseResourcesOnCancel.run();
return;
}
indexFailures.collect(index, error);
if (fieldCapTask.isCancelled()) {
releaseResourcesOnCancel.run();
}
};
final var finishedOrCancelled = new AtomicBoolean();
fieldCapTask.addListener(() -> {
if (finishedOrCancelled.compareAndSet(false, true)) {
releaseResourcesOnCancel.run();
}
});
try (RefCountingRunnable refs = new RefCountingRunnable(() -> {
finishedOrCancelled.set(true);
if (fieldCapTask.notifyIfCancelled(listener)) {
releaseResourcesOnCancel.run();
} else {
mergeIndexResponses(request, fieldCapTask, indexResponses, indexFailures, listener);
}
})) {
// local cluster
final RequestDispatcher requestDispatcher = new RequestDispatcher(
clusterService,
transportService,
task,
request,
localIndices,
nowInMillis,
concreteIndices,
threadPool.executor(ThreadPool.Names.SEARCH_COORDINATION),
handleIndexResponse,
handleIndexFailure,
refs.acquire()::close
);
requestDispatcher.execute();
// this is the cross cluster part of this API - we force the other cluster to not merge the results but instead
// send us back all individual index results.
for (Map.Entry remoteIndices : remoteClusterIndices.entrySet()) {
String clusterAlias = remoteIndices.getKey();
OriginalIndices originalIndices = remoteIndices.getValue();
Client remoteClusterClient = transportService.getRemoteClusterService().getRemoteClusterClient(threadPool, clusterAlias);
FieldCapabilitiesRequest remoteRequest = prepareRemoteRequest(request, originalIndices, nowInMillis);
ActionListener remoteListener = ActionListener.wrap(response -> {
for (FieldCapabilitiesIndexResponse resp : response.getIndexResponses()) {
String indexName = RemoteClusterAware.buildRemoteIndexName(clusterAlias, resp.getIndexName());
handleIndexResponse.accept(
new FieldCapabilitiesIndexResponse(indexName, resp.getIndexMappingHash(), resp.get(), resp.canMatch())
);
}
for (FieldCapabilitiesFailure failure : response.getFailures()) {
Exception ex = failure.getException();
for (String index : failure.getIndices()) {
handleIndexFailure.accept(RemoteClusterAware.buildRemoteIndexName(clusterAlias, index), ex);
}
}
}, ex -> {
for (String index : originalIndices.indices()) {
handleIndexFailure.accept(RemoteClusterAware.buildRemoteIndexName(clusterAlias, index), ex);
}
});
remoteClusterClient.fieldCaps(remoteRequest, ActionListener.releaseAfter(remoteListener, refs.acquire()));
}
}
}
private static void checkIndexBlocks(ClusterState clusterState, String[] concreteIndices) {
clusterState.blocks().globalBlockedRaiseException(ClusterBlockLevel.READ);
for (String index : concreteIndices) {
clusterState.blocks().indexBlockedRaiseException(ClusterBlockLevel.READ, index);
}
}
private void mergeIndexResponses(
FieldCapabilitiesRequest request,
CancellableTask task,
Map indexResponses,
FailureCollector indexFailures,
ActionListener listener
) {
List failures = indexFailures.build(indexResponses.keySet());
if (indexResponses.size() > 0) {
if (request.isMergeResults()) {
// fork off to the management pool for merging the responses as the operation can run for longer than is acceptable
// on a transport thread in case of large numbers of indices and/or fields
threadPool.executor(ThreadPool.Names.SEARCH_COORDINATION)
.submit(ActionRunnable.supply(listener, () -> merge(indexResponses, task, request, new ArrayList<>(failures))));
} else {
listener.onResponse(new FieldCapabilitiesResponse(new ArrayList<>(indexResponses.values()), new ArrayList<>(failures)));
}
} else {
// we have no responses at all, maybe because of errors
if (indexFailures.isEmpty() == false) {
// throw back the first exception
listener.onFailure(failures.iterator().next().getException());
} else {
listener.onResponse(new FieldCapabilitiesResponse(Collections.emptyList(), Collections.emptyList()));
}
}
}
private static FieldCapabilitiesRequest prepareRemoteRequest(
FieldCapabilitiesRequest request,
OriginalIndices originalIndices,
long nowInMillis
) {
FieldCapabilitiesRequest remoteRequest = new FieldCapabilitiesRequest();
remoteRequest.setMergeResults(false); // we need to merge on this node
remoteRequest.indicesOptions(originalIndices.indicesOptions());
remoteRequest.indices(originalIndices.indices());
remoteRequest.fields(request.fields());
remoteRequest.filters(request.filters());
remoteRequest.types(request.types());
remoteRequest.runtimeFields(request.runtimeFields());
remoteRequest.indexFilter(request.indexFilter());
remoteRequest.nowInMillis(nowInMillis);
return remoteRequest;
}
private static boolean hasSameMappingHash(FieldCapabilitiesIndexResponse r1, FieldCapabilitiesIndexResponse r2) {
return r1.getIndexMappingHash() != null
&& r2.getIndexMappingHash() != null
&& r1.getIndexMappingHash().equals(r2.getIndexMappingHash());
}
private FieldCapabilitiesResponse merge(
Map indexResponsesMap,
CancellableTask task,
FieldCapabilitiesRequest request,
List failures
) {
task.ensureNotCancelled();
final FieldCapabilitiesIndexResponse[] indexResponses = indexResponsesMap.values()
.stream()
.sorted(Comparator.comparing(FieldCapabilitiesIndexResponse::getIndexName))
.toArray(FieldCapabilitiesIndexResponse[]::new);
final String[] indices = Arrays.stream(indexResponses).map(FieldCapabilitiesIndexResponse::getIndexName).toArray(String[]::new);
final Map> responseMapBuilder = new HashMap<>();
int lastPendingIndex = 0;
for (int i = 1; i <= indexResponses.length; i++) {
if (i == indexResponses.length || hasSameMappingHash(indexResponses[lastPendingIndex], indexResponses[i]) == false) {
final String[] subIndices;
if (lastPendingIndex == 0 && i == indexResponses.length) {
subIndices = indices;
} else {
subIndices = ArrayUtil.copyOfSubArray(indices, lastPendingIndex, i);
}
innerMerge(subIndices, responseMapBuilder, request, indexResponses[lastPendingIndex]);
lastPendingIndex = i;
}
}
task.ensureNotCancelled();
Map> responseMap = new HashMap<>();
for (Map.Entry> entry : responseMapBuilder.entrySet()) {
Map typeMapBuilder = entry.getValue();
Optional> unmapped = Optional.empty();
if (request.includeUnmapped()) {
// do this directly, rather than using the builder, to save creating a whole lot of objects we don't need
unmapped = getUnmappedFields(
indexResponsesMap.keySet(),
entry.getKey(),
typeMapBuilder.values().stream().flatMap(FieldCapabilities.Builder::getIndices).collect(Collectors.toSet())
);
}
boolean multiTypes = typeMapBuilder.size() + unmapped.map(f -> 1).orElse(0) > 1;
responseMap.put(
entry.getKey(),
Collections.unmodifiableMap(
Stream.concat(
typeMapBuilder.entrySet()
.stream()
.map(e -> Map.>entry(e.getKey(), e.getValue()::build)),
unmapped.stream().map(f -> Map.entry("unmapped", f))
).collect(Collectors.toMap(Map.Entry::getKey, e -> e.getValue().apply(multiTypes)))
)
);
}
return new FieldCapabilitiesResponse(indices, Collections.unmodifiableMap(responseMap), failures);
}
private static Optional> getUnmappedFields(
Set indices,
String field,
Set mappedIndices
) {
if (mappedIndices.size() != indices.size()) {
return Optional.of(
mt -> FieldCapabilities.buildBasic(
field,
"unmapped",
mt ? Sets.difference(indices, mappedIndices).toArray(String[]::new) : null
)
);
}
return Optional.empty();
}
private void innerMerge(
String[] indices,
Map> responseMapBuilder,
FieldCapabilitiesRequest request,
FieldCapabilitiesIndexResponse response
) {
Map fields = ResponseRewriter.rewriteOldResponses(
response.getOriginVersion(),
response.get(),
request.filters(),
request.types()
);
for (Map.Entry entry : fields.entrySet()) {
final String field = entry.getKey();
final IndexFieldCapabilities fieldCap = entry.getValue();
Map typeMap = responseMapBuilder.computeIfAbsent(field, f -> new HashMap<>());
FieldCapabilities.Builder builder = typeMap.computeIfAbsent(
fieldCap.getType(),
key -> new FieldCapabilities.Builder(field, key)
);
builder.add(
indices,
fieldCap.isMetadatafield(),
fieldCap.isSearchable(),
fieldCap.isAggregatable(),
fieldCap.isDimension(),
fieldCap.getMetricType(),
fieldCap.meta()
);
}
}
/**
* Collects failures from all the individual index requests, then builds a failure list grouped by the underlying cause.
*
* This collector can contain a failure for an index even if one of its shards was successful. When building the final
* list, these failures will be skipped because they have no affect on the final response.
*/
private static final class FailureCollector {
private final Map failuresByIndex = Collections.synchronizedMap(new HashMap<>());
List build(Set successfulIndices) {
Map, FieldCapabilitiesFailure> indexFailures = Collections.synchronizedMap(new HashMap<>());
for (Map.Entry failure : failuresByIndex.entrySet()) {
String index = failure.getKey();
Exception e = failure.getValue();
if (successfulIndices.contains(index) == false) {
// we deduplicate exceptions on the underlying causes message and classname
// we unwrap the cause to e.g. group RemoteTransportExceptions coming from different nodes if the cause is the same
Throwable cause = ExceptionsHelper.unwrapCause(e);
Tuple groupingKey = new Tuple<>(cause.getMessage(), cause.getClass().getName());
indexFailures.compute(
groupingKey,
(k, v) -> v == null ? new FieldCapabilitiesFailure(new String[] { index }, e) : v.addIndex(index)
);
}
}
return new ArrayList<>(indexFailures.values());
}
void collect(String index, Exception e) {
failuresByIndex.putIfAbsent(index, e);
}
void clear() {
failuresByIndex.clear();
}
boolean isEmpty() {
return failuresByIndex.isEmpty();
}
}
private class NodeTransportHandler implements TransportRequestHandler {
@Override
public void messageReceived(FieldCapabilitiesNodeRequest request, TransportChannel channel, Task task) throws Exception {
assert task instanceof CancellableTask;
final ActionListener listener = new ChannelActionListener<>(channel);
ActionListener.completeWith(listener, () -> {
final List allResponses = new ArrayList<>();
final Map allFailures = new HashMap<>();
final Set allUnmatchedShardIds = new HashSet<>();
// If the request has an index filter, it may contain several shards belonging to the same
// index. We make sure to skip over a shard if we already found a match for that index.
final Map> groupedShardIds = request.shardIds()
.stream()
.collect(Collectors.groupingBy(ShardId::getIndexName));
final FieldCapabilitiesFetcher fetcher = new FieldCapabilitiesFetcher(indicesService);
for (List shardIds : groupedShardIds.values()) {
final Map failures = new HashMap<>();
final Set unmatched = new HashSet<>();
for (ShardId shardId : shardIds) {
try {
final FieldCapabilitiesIndexResponse response = fetcher.fetch(
(CancellableTask) task,
shardId,
request.fields(),
request.filters(),
request.allowedTypes(),
request.indexFilter(),
request.nowInMillis(),
request.runtimeFields()
);
if (response.canMatch()) {
unmatched.clear();
failures.clear();
allResponses.add(response);
break;
} else {
unmatched.add(shardId);
}
} catch (Exception e) {
failures.put(shardId, e);
}
}
allUnmatchedShardIds.addAll(unmatched);
allFailures.putAll(failures);
}
return new FieldCapabilitiesNodeResponse(allResponses, allFailures, allUnmatchedShardIds);
});
}
}
}
© 2015 - 2025 Weber Informatics LLC | Privacy Policy