All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.opensearch.action.search.CreatePitController Maven / Gradle / Ivy

There is a newer version: 2.18.0
Show newest version
/*
 * SPDX-License-Identifier: Apache-2.0
 *
 * The OpenSearch Contributors require contributions made to
 * this file be licensed under the Apache-2.0 license or a
 * compatible open source license.
 */

package org.opensearch.action.search;

import org.apache.logging.log4j.LogManager;
import org.apache.logging.log4j.Logger;
import org.apache.logging.log4j.message.ParameterizedMessage;
import org.opensearch.OpenSearchException;
import org.opensearch.action.StepListener;
import org.opensearch.action.support.GroupedActionListener;
import org.opensearch.cluster.ClusterState;
import org.opensearch.cluster.node.DiscoveryNode;
import org.opensearch.cluster.service.ClusterService;
import org.opensearch.common.inject.Inject;
import org.opensearch.common.settings.Setting;
import org.opensearch.common.unit.TimeValue;
import org.opensearch.core.action.ActionListener;
import org.opensearch.core.common.Strings;
import org.opensearch.core.common.io.stream.NamedWriteableRegistry;
import org.opensearch.core.index.shard.ShardId;
import org.opensearch.search.SearchPhaseResult;
import org.opensearch.search.SearchShardTarget;
import org.opensearch.tasks.Task;
import org.opensearch.transport.Transport;

import java.util.ArrayList;
import java.util.Arrays;
import java.util.Collection;
import java.util.Collections;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
import java.util.Set;
import java.util.function.BiFunction;
import java.util.stream.Collectors;

import static org.opensearch.common.unit.TimeValue.timeValueSeconds;

/**
 * Controller for creating PIT reader context
 * Phase 1 of create PIT request : Create PIT reader contexts in the associated shards with a temporary keep alive
 * Phase 2 of create PIT : Update PIT reader context with PIT ID and keep alive from request and
 * fail user request if any of the updates in this phase are failed - we clean up PITs in case of such failures.
 * This two phase approach is used to save PIT ID as part of context which is later used for other use cases like list PIT etc.
 */
public class CreatePitController {
    private final SearchTransportService searchTransportService;
    private final ClusterService clusterService;
    private final TransportSearchAction transportSearchAction;
    private final NamedWriteableRegistry namedWriteableRegistry;
    private final PitService pitService;
    private static final Logger logger = LogManager.getLogger(CreatePitController.class);
    public static final Setting PIT_INIT_KEEP_ALIVE = Setting.positiveTimeSetting(
        "point_in_time.init.keep_alive",
        timeValueSeconds(30),
        Setting.Property.NodeScope
    );

    @Inject
    public CreatePitController(
        SearchTransportService searchTransportService,
        ClusterService clusterService,
        TransportSearchAction transportSearchAction,
        NamedWriteableRegistry namedWriteableRegistry,
        PitService pitService
    ) {
        this.searchTransportService = searchTransportService;
        this.clusterService = clusterService;
        this.transportSearchAction = transportSearchAction;
        this.namedWriteableRegistry = namedWriteableRegistry;
        this.pitService = pitService;
    }

    /**
     * This method creates PIT reader context
     */
    public void executeCreatePit(
        CreatePitRequest request,
        Task task,
        StepListener createPitListener,
        ActionListener updatePitIdListener
    ) {
        SearchRequest searchRequest = new SearchRequest(request.getIndices());
        searchRequest.preference(request.getPreference());
        searchRequest.routing(request.getRouting());
        searchRequest.indicesOptions(request.getIndicesOptions());
        searchRequest.allowPartialSearchResults(request.shouldAllowPartialPitCreation());
        SearchTask searchTask = searchRequest.createTask(
            task.getId(),
            task.getType(),
            task.getAction(),
            task.getParentTaskId(),
            Collections.emptyMap()
        );
        /*
          This is needed for cross cluster functionality to work with PITs and current ccsMinimizeRoundTrips is
          not supported for point in time
         */
        searchRequest.setCcsMinimizeRoundtrips(false);
        /*
          Phase 1 of create PIT
         */
        executeCreatePit(searchTask, searchRequest, createPitListener);

        /*
          Phase 2 of create PIT where we update pit id in pit contexts
         */
        createPitListener.whenComplete(
            searchResponse -> { executeUpdatePitId(request, searchRequest, searchResponse, updatePitIdListener); },
            updatePitIdListener::onFailure
        );
    }

    /**
     * Creates PIT reader context with temporary keep alive
     */
    void executeCreatePit(Task task, SearchRequest searchRequest, StepListener createPitListener) {
        logger.debug(
            () -> new ParameterizedMessage("Executing creation of PIT context for indices [{}]", Arrays.toString(searchRequest.indices()))
        );
        transportSearchAction.executeRequest(
            task,
            searchRequest,
            TransportCreatePitAction.CREATE_PIT_ACTION,
            true,
            new TransportSearchAction.SinglePhaseSearchAction() {
                @Override
                public void executeOnShardTarget(
                    SearchTask searchTask,
                    SearchShardTarget target,
                    Transport.Connection connection,
                    ActionListener searchPhaseResultActionListener
                ) {
                    searchTransportService.createPitContext(
                        connection,
                        new TransportCreatePitAction.CreateReaderContextRequest(
                            target.getShardId(),
                            PIT_INIT_KEEP_ALIVE.get(clusterService.getSettings())
                        ),
                        searchTask,
                        ActionListener.wrap(r -> searchPhaseResultActionListener.onResponse(r), searchPhaseResultActionListener::onFailure)
                    );
                }
            },
            createPitListener
        );
    }

    /**
     * Updates PIT ID, keep alive and createdTime of PIT reader context
     */
    void executeUpdatePitId(
        CreatePitRequest request,
        SearchRequest searchRequest,
        SearchResponse searchResponse,
        ActionListener updatePitIdListener
    ) {
        logger.debug(
            () -> new ParameterizedMessage(
                "Updating PIT context with PIT ID [{}], creation time and keep alive",
                searchResponse.pointInTimeId()
            )
        );
        /*
          store the create time ( same create time for all PIT contexts across shards ) to be used
          for list PIT api
         */
        final long relativeStartNanos = System.nanoTime();
        final TransportSearchAction.SearchTimeProvider timeProvider = new TransportSearchAction.SearchTimeProvider(
            searchRequest.getOrCreateAbsoluteStartMillis(),
            relativeStartNanos,
            System::nanoTime
        );
        final long creationTime = timeProvider.getAbsoluteStartMillis();
        CreatePitResponse createPITResponse = new CreatePitResponse(
            searchResponse.pointInTimeId(),
            creationTime,
            searchResponse.getTotalShards(),
            searchResponse.getSuccessfulShards(),
            searchResponse.getSkippedShards(),
            searchResponse.getFailedShards(),
            searchResponse.getShardFailures()
        );
        SearchContextId contextId = SearchContextId.decode(namedWriteableRegistry, createPITResponse.getId());
        final StepListener> lookupListener = getConnectionLookupListener(contextId);
        lookupListener.whenComplete(nodelookup -> {
            final ActionListener groupedActionListener = getGroupedListener(
                updatePitIdListener,
                createPITResponse,
                contextId.shards().size(),
                contextId.shards().values()
            );
            for (Map.Entry entry : contextId.shards().entrySet()) {
                DiscoveryNode node = nodelookup.apply(entry.getValue().getClusterAlias(), entry.getValue().getNode());
                if (node == null) {
                    node = this.clusterService.state().getNodes().get(entry.getValue().getNode());
                }
                if (node == null) {
                    logger.error(
                        () -> new ParameterizedMessage(
                            "Create pit update phase for PIT ID [{}] failed " + "because node [{}] not found",
                            searchResponse.pointInTimeId(),
                            entry.getValue().getNode()
                        )
                    );
                    groupedActionListener.onFailure(
                        new OpenSearchException(
                            "Create pit update phase for PIT ID ["
                                + searchResponse.pointInTimeId()
                                + "] failed because node["
                                + entry.getValue().getNode()
                                + "] "
                                + "not found"
                        )
                    );
                    return;
                }
                try {
                    final Transport.Connection connection = searchTransportService.getConnection(entry.getValue().getClusterAlias(), node);
                    searchTransportService.updatePitContext(
                        connection,
                        new UpdatePitContextRequest(
                            entry.getValue().getSearchContextId(),
                            createPITResponse.getId(),
                            request.getKeepAlive().millis(),
                            creationTime
                        ),
                        groupedActionListener
                    );
                } catch (Exception e) {
                    String nodeName = node.getName();
                    logger.error(
                        () -> new ParameterizedMessage(
                            "Create pit update phase failed for PIT ID [{}] on node [{}]",
                            searchResponse.pointInTimeId(),
                            nodeName
                        ),
                        e
                    );
                    groupedActionListener.onFailure(
                        new OpenSearchException(
                            "Create pit update phase for PIT ID [" + searchResponse.pointInTimeId() + "] failed on node[" + node + "]",
                            e
                        )
                    );
                }
            }
        }, updatePitIdListener::onFailure);
    }

    private StepListener> getConnectionLookupListener(SearchContextId contextId) {
        ClusterState state = clusterService.state();
        final Set clusters = contextId.shards()
            .values()
            .stream()
            .filter(ctx -> Strings.isEmpty(ctx.getClusterAlias()) == false)
            .map(SearchContextIdForNode::getClusterAlias)
            .collect(Collectors.toSet());
        return (StepListener>) SearchUtils.getConnectionLookupListener(
            searchTransportService.getRemoteClusterService(),
            state,
            clusters
        );
    }

    private ActionListener getGroupedListener(
        ActionListener updatePitIdListener,
        CreatePitResponse createPITResponse,
        int size,
        Collection contexts
    ) {
        return new GroupedActionListener<>(new ActionListener<>() {
            @Override
            public void onResponse(final Collection responses) {
                updatePitIdListener.onResponse(createPITResponse);
            }

            @Override
            public void onFailure(final Exception e) {
                cleanupContexts(contexts, createPITResponse.getId());
                updatePitIdListener.onFailure(e);
            }
        }, size);
    }

    /**
     * Cleanup all created PIT contexts in case of failure
     */
    private void cleanupContexts(Collection contexts, String pitId) {
        ActionListener deleteListener = new ActionListener<>() {
            @Override
            public void onResponse(DeletePitResponse response) {
                // this is invoke and forget call
                final StringBuilder failedPitsStringBuilder = new StringBuilder();
                response.getDeletePitResults()
                    .stream()
                    .filter(r -> !r.isSuccessful())
                    .forEach(r -> failedPitsStringBuilder.append(r.getPitId()).append(","));
                logger.warn(() -> new ParameterizedMessage("Failed to delete PIT IDs {}", failedPitsStringBuilder.toString()));
                if (logger.isDebugEnabled()) {
                    final StringBuilder successfulPitsStringBuilder = new StringBuilder();
                    response.getDeletePitResults()
                        .stream()
                        .filter(r -> r.isSuccessful())
                        .forEach(r -> successfulPitsStringBuilder.append(r.getPitId()).append(","));
                    logger.debug(() -> new ParameterizedMessage("Deleted PIT with IDs {}", successfulPitsStringBuilder.toString()));
                }
            }

            @Override
            public void onFailure(Exception e) {
                logger.error("Cleaning up PIT contexts failed ", e);
            }
        };
        Map> nodeToContextsMap = new HashMap<>();
        for (SearchContextIdForNode context : contexts) {
            List contextIdsForNode = nodeToContextsMap.getOrDefault(context.getNode(), new ArrayList<>());
            contextIdsForNode.add(new PitSearchContextIdForNode(pitId, context));
            nodeToContextsMap.put(context.getNode(), contextIdsForNode);
        }
        pitService.deletePitContexts(nodeToContextsMap, deleteListener);
    }
}




© 2015 - 2024 Weber Informatics LLC | Privacy Policy