All Downloads are FREE. Search and download functionalities are using the official Maven repository.

io.trino.plugin.deltalake.procedure.VacuumProcedure Maven / Gradle / Ivy

There is a newer version: 458
Show newest version
/*
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package io.trino.plugin.deltalake.procedure;

import com.google.common.collect.ImmutableList;
import com.google.common.collect.ImmutableSet;
import com.google.inject.Inject;
import com.google.inject.Provider;
import io.airlift.log.Logger;
import io.airlift.units.Duration;
import io.trino.filesystem.FileEntry;
import io.trino.filesystem.FileIterator;
import io.trino.filesystem.Location;
import io.trino.filesystem.TrinoFileSystem;
import io.trino.filesystem.TrinoFileSystemFactory;
import io.trino.plugin.base.util.UncheckedCloseable;
import io.trino.plugin.deltalake.DeltaLakeConfig;
import io.trino.plugin.deltalake.DeltaLakeMetadata;
import io.trino.plugin.deltalake.DeltaLakeMetadataFactory;
import io.trino.plugin.deltalake.DeltaLakeSessionProperties;
import io.trino.plugin.deltalake.DeltaLakeTableHandle;
import io.trino.plugin.deltalake.transactionlog.AddFileEntry;
import io.trino.plugin.deltalake.transactionlog.DeltaLakeTransactionLogEntry;
import io.trino.plugin.deltalake.transactionlog.ProtocolEntry;
import io.trino.plugin.deltalake.transactionlog.RemoveFileEntry;
import io.trino.plugin.deltalake.transactionlog.TableSnapshot;
import io.trino.plugin.deltalake.transactionlog.TransactionLogAccess;
import io.trino.spi.TrinoException;
import io.trino.spi.catalog.CatalogName;
import io.trino.spi.classloader.ThreadContextClassLoader;
import io.trino.spi.connector.ConnectorAccessControl;
import io.trino.spi.connector.ConnectorSession;
import io.trino.spi.connector.ConnectorTableHandle;
import io.trino.spi.connector.SchemaTableName;
import io.trino.spi.predicate.TupleDomain;
import io.trino.spi.procedure.Procedure;
import io.trino.spi.procedure.Procedure.Argument;

import java.io.IOException;
import java.lang.invoke.MethodHandle;
import java.time.Instant;
import java.util.ArrayList;
import java.util.List;
import java.util.Objects;
import java.util.Set;
import java.util.stream.Stream;

import static com.google.common.base.Preconditions.checkState;
import static com.google.common.base.Predicates.alwaysFalse;
import static com.google.common.collect.ImmutableList.toImmutableList;
import static com.google.common.collect.ImmutableSet.toImmutableSet;
import static io.trino.plugin.base.util.Procedures.checkProcedureArgument;
import static io.trino.plugin.deltalake.DeltaLakeMetadata.MAX_WRITER_VERSION;
import static io.trino.plugin.deltalake.DeltaLakeMetadata.checkValidTableHandle;
import static io.trino.plugin.deltalake.DeltaLakeSessionProperties.getVacuumMinRetention;
import static io.trino.plugin.deltalake.transactionlog.DeltaLakeSchemaSupport.unsupportedWriterFeatures;
import static io.trino.plugin.deltalake.transactionlog.TransactionLogUtil.TRANSACTION_LOG_DIRECTORY;
import static io.trino.plugin.deltalake.transactionlog.TransactionLogUtil.getTransactionLogDir;
import static io.trino.spi.StandardErrorCode.NOT_SUPPORTED;
import static io.trino.spi.type.VarcharType.VARCHAR;
import static java.lang.String.format;
import static java.lang.invoke.MethodHandles.lookup;
import static java.util.Comparator.naturalOrder;
import static java.util.Objects.requireNonNull;

public class VacuumProcedure
        implements Provider
{
    private static final Logger log = Logger.get(VacuumProcedure.class);
    private static final int DELETE_BATCH_SIZE = 1000;

    private static final MethodHandle VACUUM;

    static {
        try {
            VACUUM = lookup().unreflect(VacuumProcedure.class.getMethod("vacuum", ConnectorSession.class, ConnectorAccessControl.class, String.class, String.class, String.class));
        }
        catch (ReflectiveOperationException e) {
            throw new AssertionError(e);
        }
    }

    private final CatalogName catalogName;
    private final TrinoFileSystemFactory fileSystemFactory;
    private final DeltaLakeMetadataFactory metadataFactory;
    private final TransactionLogAccess transactionLogAccess;

    @Inject
    public VacuumProcedure(
            CatalogName catalogName,
            TrinoFileSystemFactory fileSystemFactory,
            DeltaLakeMetadataFactory metadataFactory,
            TransactionLogAccess transactionLogAccess)
    {
        this.catalogName = requireNonNull(catalogName, "catalogName is null");
        this.fileSystemFactory = requireNonNull(fileSystemFactory, "fileSystemFactory is null");
        this.metadataFactory = requireNonNull(metadataFactory, "metadataFactory is null");
        this.transactionLogAccess = requireNonNull(transactionLogAccess, "transactionLogAccess is null");
    }

    @Override
    public Procedure get()
    {
        return new Procedure(
                "system",
                "vacuum",
                ImmutableList.of(
                        new Argument("SCHEMA_NAME", VARCHAR),
                        new Argument("TABLE_NAME", VARCHAR),
                        // Databricks's VACUUM has 7d default. We decided not to pick the default here.
                        new Argument("RETENTION", VARCHAR)),
                VACUUM.bindTo(this));
    }

    public void vacuum(
            ConnectorSession session,
            ConnectorAccessControl accessControl,
            String schema,
            String table,
            String retention)
    {
        try (ThreadContextClassLoader ignored = new ThreadContextClassLoader(getClass().getClassLoader())) {
            doVacuum(session, accessControl, schema, table, retention);
        }
        catch (TrinoException e) {
            throw e;
        }
        catch (Exception e) {
            // This is not categorized as TrinoException. All possible external failures should be handled explicitly.
            throw new RuntimeException(format("Failure when vacuuming %s.%s with retention %s: %s", schema, table, retention, e), e);
        }
    }

    private void doVacuum(
            ConnectorSession session,
            ConnectorAccessControl accessControl,
            String schema,
            String table,
            String retention)
            throws IOException
    {
        checkProcedureArgument(schema != null, "schema_name cannot be null");
        checkProcedureArgument(!schema.isEmpty(), "schema_name cannot be empty");
        checkProcedureArgument(table != null, "table_name cannot be null");
        checkProcedureArgument(!table.isEmpty(), "table_name cannot be empty");
        checkProcedureArgument(retention != null, "retention cannot be null");

        Duration retentionDuration = Duration.valueOf(retention);
        Duration minRetention = getVacuumMinRetention(session);
        checkProcedureArgument(
                retentionDuration.compareTo(minRetention) >= 0,
                "Retention specified (%s) is shorter than the minimum retention configured in the system (%s). " +
                        "Minimum retention can be changed with %s configuration property or %s.%s session property",
                retentionDuration,
                minRetention,
                DeltaLakeConfig.VACUUM_MIN_RETENTION,
                catalogName,
                DeltaLakeSessionProperties.VACUUM_MIN_RETENTION);

        Instant threshold = Instant.now().minusMillis(retentionDuration.toMillis());

        DeltaLakeMetadata metadata = metadataFactory.create(session.getIdentity());
        metadata.beginQuery(session);
        try (UncheckedCloseable ignore = () -> metadata.cleanupQuery(session)) {
            SchemaTableName tableName = new SchemaTableName(schema, table);
            ConnectorTableHandle connectorTableHandle = metadata.getTableHandle(session, tableName);
            checkProcedureArgument(connectorTableHandle != null, "Table '%s' does not exist", tableName);
            DeltaLakeTableHandle handle = checkValidTableHandle(connectorTableHandle);

            accessControl.checkCanInsertIntoTable(null, tableName);
            accessControl.checkCanDeleteFromTable(null, tableName);

            TableSnapshot tableSnapshot = metadata.getSnapshot(session, tableName, handle.getLocation(), handle.getReadVersion());
            ProtocolEntry protocolEntry = transactionLogAccess.getProtocolEntry(session, tableSnapshot);
            if (protocolEntry.minWriterVersion() > MAX_WRITER_VERSION) {
                throw new TrinoException(NOT_SUPPORTED, "Cannot execute vacuum procedure with %d writer version".formatted(protocolEntry.minWriterVersion()));
            }
            Set unsupportedWriterFeatures = unsupportedWriterFeatures(protocolEntry.writerFeatures().orElse(ImmutableSet.of()));
            if (!unsupportedWriterFeatures.isEmpty()) {
                throw new TrinoException(NOT_SUPPORTED, "Cannot execute vacuum procedure with %s writer features".formatted(unsupportedWriterFeatures));
            }

            String tableLocation = tableSnapshot.getTableLocation();
            String transactionLogDir = getTransactionLogDir(tableLocation);
            TrinoFileSystem fileSystem = fileSystemFactory.create(session);
            String commonPathPrefix = tableLocation.endsWith("/") ? tableLocation : tableLocation + "/";
            String queryId = session.getQueryId();

            // Retain all active files and every file removed by a "recent" transaction (except for the oldest "recent").
            // Any remaining file are not live, and not needed to read any "recent" snapshot.
            List recentVersions = transactionLogAccess.getPastTableVersions(fileSystem, transactionLogDir, threshold, tableSnapshot.getVersion());
            Set retainedPaths;
            try (Stream activeAddEntries = transactionLogAccess.getActiveFiles(
                    session,
                    tableSnapshot,
                    handle.getMetadataEntry(),
                    handle.getProtocolEntry(),
                    TupleDomain.all(),
                    alwaysFalse())) {
                retainedPaths = Stream.concat(
                                activeAddEntries
                                        // paths can be absolute as well in case of shallow-cloned tables, and they shouldn't be deleted as part of vacuum because according to
                                        // delta-protocol absolute paths are inherited from base table and the vacuum procedure should only list and delete local file references
                                        .map(AddFileEntry::getPath),
                                transactionLogAccess.getJsonEntries(
                                                fileSystem,
                                                transactionLogDir,
                                                // discard oldest "recent" snapshot, since we take RemoveFileEntry only, to identify files that are no longer
                                                // active files, but still needed to read a "recent" snapshot
                                                recentVersions.stream().sorted(naturalOrder())
                                                        .skip(1)
                                                        .collect(toImmutableList()))
                                        .map(DeltaLakeTransactionLogEntry::getRemove)
                                        .filter(Objects::nonNull)
                                        .map(RemoveFileEntry::path))
                        .peek(path -> checkState(!path.startsWith(tableLocation), "Unexpected absolute path in transaction log: %s", path))
                        .collect(toImmutableSet());
            }

            log.debug(
                    "[%s] attempting to vacuum table %s [%s] with %s retention (expiry threshold %s). %s data file paths marked for retention",
                    queryId,
                    tableName,
                    tableLocation,
                    retention,
                    threshold,
                    retainedPaths.size());

            long allPathsChecked = 0;
            long transactionLogFiles = 0;
            long retainedKnownFiles = 0;
            long retainedUnknownFiles = 0;
            long removedFiles = 0;

            List filesToDelete = new ArrayList<>();
            FileIterator listing = fileSystem.listFiles(Location.of(tableLocation));
            while (listing.hasNext()) {
                FileEntry entry = listing.next();
                String location = entry.location().toString();
                checkState(
                        location.startsWith(commonPathPrefix),
                        "Unexpected path [%s] returned when listing files under [%s]",
                        location,
                        tableLocation);
                String relativePath = location.substring(commonPathPrefix.length());
                if (relativePath.isEmpty()) {
                    // A file returned for "tableLocation/", might be possible on S3.
                    continue;
                }
                allPathsChecked++;

                // ignore tableLocation/_delta_log/**
                if (relativePath.equals(TRANSACTION_LOG_DIRECTORY) || relativePath.startsWith(TRANSACTION_LOG_DIRECTORY + "/")) {
                    log.debug("[%s] skipping a file inside transaction log dir: %s", queryId, location);
                    transactionLogFiles++;
                    continue;
                }

                // skip retained files
                if (retainedPaths.contains(relativePath)) {
                    log.debug("[%s] retaining a known file: %s", queryId, location);
                    retainedKnownFiles++;
                    continue;
                }

                // ignore recently created files
                Instant modificationTime = entry.lastModified();
                if (!modificationTime.isBefore(threshold)) {
                    log.debug("[%s] retaining an unknown file %s with modification time %s", queryId, location, modificationTime);
                    retainedUnknownFiles++;
                    continue;
                }

                log.debug("[%s] deleting file [%s] with modification time %s", queryId, location, modificationTime);
                filesToDelete.add(entry.location());
                if (filesToDelete.size() == DELETE_BATCH_SIZE) {
                    fileSystem.deleteFiles(filesToDelete);
                    removedFiles += filesToDelete.size();
                    filesToDelete.clear();
                }
            }

            if (!filesToDelete.isEmpty()) {
                fileSystem.deleteFiles(filesToDelete);
                removedFiles += filesToDelete.size();
            }

            log.info(
                    "[%s] finished vacuuming table %s [%s]: files checked: %s; metadata files: %s; retained known files: %s; retained unknown files: %s; removed files: %s",
                    queryId,
                    tableName,
                    tableLocation,
                    allPathsChecked,
                    transactionLogFiles,
                    retainedKnownFiles,
                    retainedUnknownFiles,
                    removedFiles);
        }
    }
}




© 2015 - 2024 Weber Informatics LLC | Privacy Policy