All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.kylin.rest.service.MetaStoreService Maven / Gradle / Ivy

The newest version!
/*
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

package org.apache.kylin.rest.service;

import static org.apache.kylin.common.constant.Constants.KE_VERSION;
import static org.apache.kylin.common.exception.ServerErrorCode.FAILED_CREATE_MODEL;
import static org.apache.kylin.common.exception.ServerErrorCode.MODEL_EXPORT_ERROR;
import static org.apache.kylin.common.exception.ServerErrorCode.MODEL_IMPORT_ERROR;
import static org.apache.kylin.common.exception.ServerErrorCode.MODEL_METADATA_FILE_ERROR;
import static org.apache.kylin.common.exception.code.ErrorCodeServer.MODEL_ID_NOT_EXIST;
import static org.apache.kylin.common.exception.code.ErrorCodeServer.MODEL_NAME_DUPLICATE;
import static org.apache.kylin.common.exception.code.ErrorCodeServer.MODEL_NAME_INVALID;
import static org.apache.kylin.common.persistence.ResourceStore.VERSION_FILE;
import static org.apache.kylin.common.persistence.ResourceStore.VERSION_FILE_META_KEY_TAG;
import static org.apache.kylin.common.persistence.metadata.FileSystemMetadataStore.JSON_SUFFIX;
import static org.apache.kylin.metadata.model.schema.SchemaNodeType.MODEL_DIM;
import static org.apache.kylin.metadata.model.schema.SchemaNodeType.MODEL_FACT;

import java.io.ByteArrayOutputStream;
import java.io.IOException;
import java.io.InputStream;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.Collection;
import java.util.Collections;
import java.util.Comparator;
import java.util.LinkedHashMap;
import java.util.List;
import java.util.Locale;
import java.util.Map;
import java.util.Objects;
import java.util.Set;
import java.util.regex.Matcher;
import java.util.regex.Pattern;
import java.util.stream.Collectors;
import java.util.stream.Stream;
import java.util.zip.ZipEntry;
import java.util.zip.ZipOutputStream;

import javax.servlet.http.HttpServletRequest;
import javax.xml.bind.DatatypeConverter;

import org.apache.commons.collections.CollectionUtils;
import org.apache.commons.lang3.StringUtils;
import org.apache.kylin.common.KylinConfig;
import org.apache.kylin.common.exception.KylinException;
import org.apache.kylin.common.msg.Message;
import org.apache.kylin.common.msg.MsgPicker;
import org.apache.kylin.common.persistence.InMemResourceStore;
import org.apache.kylin.common.persistence.MetadataType;
import org.apache.kylin.common.persistence.RawResource;
import org.apache.kylin.common.persistence.RawResourceFilter;
import org.apache.kylin.common.persistence.ResourceStore;
import org.apache.kylin.common.persistence.StringEntity;
import org.apache.kylin.common.persistence.metadata.FileSystemMetadataStore;
import org.apache.kylin.common.persistence.metadata.MetadataStore;
import org.apache.kylin.common.persistence.transaction.UnitOfWork;
import org.apache.kylin.common.util.JsonUtil;
import org.apache.kylin.common.util.MetadataChecker;
import org.apache.kylin.common.util.Pair;
import org.apache.kylin.common.util.RandomUtil;
import org.apache.kylin.guava30.shaded.common.annotations.VisibleForTesting;
import org.apache.kylin.guava30.shaded.common.collect.Lists;
import org.apache.kylin.guava30.shaded.common.collect.Maps;
import org.apache.kylin.guava30.shaded.common.collect.Sets;
import org.apache.kylin.guava30.shaded.common.io.ByteSource;
import org.apache.kylin.helper.RoutineToolHelper;
import org.apache.kylin.metadata.Manager;
import org.apache.kylin.metadata.cube.model.IndexEntity;
import org.apache.kylin.metadata.cube.model.IndexPlan;
import org.apache.kylin.metadata.cube.model.NDataflowManager;
import org.apache.kylin.metadata.cube.model.NIndexPlanManager;
import org.apache.kylin.metadata.cube.model.RuleBasedIndex;
import org.apache.kylin.metadata.model.CcModelRelationDesc;
import org.apache.kylin.metadata.model.ComputedColumnDesc;
import org.apache.kylin.metadata.model.ComputedColumnManager;
import org.apache.kylin.metadata.model.JoinTableDesc;
import org.apache.kylin.metadata.model.MultiPartitionDesc;
import org.apache.kylin.metadata.model.NDataModel;
import org.apache.kylin.metadata.model.NDataModelManager;
import org.apache.kylin.metadata.model.SegmentConfig;
import org.apache.kylin.metadata.model.SegmentStatusEnum;
import org.apache.kylin.metadata.model.TableDesc;
import org.apache.kylin.metadata.model.TableRef;
import org.apache.kylin.metadata.model.schema.ImportModelContext;
import org.apache.kylin.metadata.model.schema.ModelImportChecker;
import org.apache.kylin.metadata.model.schema.SchemaChangeCheckResult;
import org.apache.kylin.metadata.model.schema.SchemaNodeType;
import org.apache.kylin.metadata.model.schema.SchemaUtil;
import org.apache.kylin.metadata.project.NProjectManager;
import org.apache.kylin.metadata.project.ProjectInstance;
import org.apache.kylin.metadata.realization.RealizationStatusEnum;
import org.apache.kylin.metadata.recommendation.candidate.JdbcRawRecStore;
import org.apache.kylin.metadata.recommendation.candidate.RawRecItem;
import org.apache.kylin.metadata.recommendation.candidate.RawRecManager;
import org.apache.kylin.metadata.recommendation.entity.RecItemSet;
import org.apache.kylin.metadata.recommendation.ref.OptRecManagerV2;
import org.apache.kylin.metadata.view.LogicalView;
import org.apache.kylin.metadata.view.LogicalViewManager;
import org.apache.kylin.rest.aspect.Transaction;
import org.apache.kylin.rest.constant.ModelStatusToDisplayEnum;
import org.apache.kylin.rest.request.ModelImportRequest;
import org.apache.kylin.rest.request.ModelImportRequest.ImportType;
import org.apache.kylin.rest.request.StorageCleanupRequest;
import org.apache.kylin.rest.request.UpdateRuleBasedCuboidRequest;
import org.apache.kylin.rest.response.LoadTableResponse;
import org.apache.kylin.rest.response.ModelPreviewResponse;
import org.apache.kylin.rest.response.SimplifiedTablePreviewResponse;
import org.apache.kylin.rest.util.AclEvaluate;
import org.apache.kylin.rest.util.AclPermissionUtil;
import org.apache.kylin.source.ISourceMetadataExplorer;
import org.apache.kylin.source.SourceFactory;
import org.apache.kylin.tool.garbage.CleanTaskExecutorService;
import org.apache.kylin.tool.util.HashFunction;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.stereotype.Component;
import org.springframework.web.multipart.MultipartFile;

import com.fasterxml.jackson.core.JsonProcessingException;

import lombok.Setter;
import lombok.val;
import lombok.var;

@Component("metaStoreService")
public class MetaStoreService extends BasicService {
    private static final Logger logger = LoggerFactory.getLogger(MetaStoreService.class);
    private static final String BASE_CUBOID_ALWAYS_VALID_KEY = "kylin.cube.aggrgroup.is-base-cuboid-always-valid";
    private static final Pattern MD5_PATTERN = Pattern.compile(".*([a-fA-F\\d]{32})\\.zip");
    private static final String RULE_SCHEDULER_DATA_KEY = "kylin.index.rule-scheduler-data";

    @Autowired
    public AclEvaluate aclEvaluate;

    @Autowired
    public ModelService modelService;

    @Autowired
    public IndexPlanService indexPlanService;

    @Autowired
    public TableExtService tableExtService;

    @Autowired
    private RouteService routeService;

    @Setter
    @Autowired(required = false)
    private List modelChangeSupporters = Lists.newArrayList();

    public List getPreviewModels(String project, List ids) {
        aclEvaluate.checkProjectWritePermission(project);
        return modelService.getManager(NDataflowManager.class, project).listAllDataflows(true).stream()
                .filter(df -> ids.isEmpty() || ids.contains(df.getUuid())).map(df -> {
                    if (df.checkBrokenWithRelatedInfo()) {
                        NDataModel dataModel = getManager(NDataModelManager.class, project)
                                .getDataModelDescWithoutInit(df.getUuid());
                        dataModel.setBroken(true);
                        return dataModel;
                    } else {
                        return df.getModel();
                    }
                }).filter(model -> !model.isFusionModel() && model.getModelType() != NDataModel.ModelType.STREAMING)
                .map(modelDesc -> getSimplifiedModelResponse(project, modelDesc)).collect(Collectors.toList());
    }

    private ModelPreviewResponse getSimplifiedModelResponse(String project, NDataModel modelDesc) {
        val projectManager = getManager(NProjectManager.class);
        val projectInstance = projectManager.getProject(project);
        ModelPreviewResponse modelPreviewResponse = new ModelPreviewResponse();
        modelPreviewResponse.setName(modelDesc.getAlias());
        modelPreviewResponse.setUuid(modelDesc.getUuid());
        NDataflowManager dfManager = NDataflowManager.getInstance(KylinConfig.getInstanceFromEnv(),
                modelDesc.getProject());
        if (modelDesc.isBroken()) {
            modelPreviewResponse.setStatus(ModelStatusToDisplayEnum.BROKEN);
            return modelPreviewResponse;
        }
        long inconsistentSegmentCount = dfManager.getDataflow(modelDesc.getId()).getSegments(SegmentStatusEnum.WARNING)
                .size();
        ModelStatusToDisplayEnum status = modelService.convertModelStatusToDisplay(modelDesc, modelDesc.getProject(),
                inconsistentSegmentCount);
        modelPreviewResponse.setStatus(status);

        if (!projectInstance.isExpertMode()) {
            int rawRecItemCount = modelChangeSupporters.stream()
                    .map(listener -> listener.getRecItemSize(project, modelDesc.getUuid())).reduce((a, b) -> a + b)
                    .orElse(0);
            if (rawRecItemCount > 0) {
                modelPreviewResponse.setHasRecommendation(true);
            }
        }

        if (projectInstance.getConfig().isMultiPartitionEnabled() && modelDesc.isMultiPartitionModel()) {
            modelPreviewResponse
                    .setHasMultiplePartitionValues(!modelDesc.getMultiPartitionDesc().getPartitions().isEmpty());
        }

        NIndexPlanManager indexPlanManager = getManager(NIndexPlanManager.class, modelDesc.getProject());
        IndexPlan indexPlan = indexPlanManager.getIndexPlan(modelDesc.getUuid());
        if (!isEmptyAfterExcludeBlockData(indexPlan)
                || (modelDesc.getSegmentConfig() != null && modelDesc.getSegmentConfig().getAutoMergeEnabled() != null
                        && modelDesc.getSegmentConfig().getAutoMergeEnabled())) {
            modelPreviewResponse.setHasOverrideProps(true);
        }

        List tables = new ArrayList<>();
        SimplifiedTablePreviewResponse factTable = new SimplifiedTablePreviewResponse(modelDesc.getRootFactTableName(),
                NDataModel.TableKind.FACT);
        tables.add(factTable);
        List joinTableDescs = modelDesc.getJoinTables();
        for (JoinTableDesc joinTableDesc : joinTableDescs) {
            SimplifiedTablePreviewResponse lookupTable = new SimplifiedTablePreviewResponse(joinTableDesc.getTable(),
                    joinTableDesc.getKind());
            tables.add(lookupTable);
        }
        modelPreviewResponse.setTables(tables);
        return modelPreviewResponse;
    }

    private boolean isEmptyAfterExcludeBlockData(IndexPlan indexPlan) {
        val overrideProps = indexPlan.getOverrideProps();
        boolean isEmpty = overrideProps.isEmpty();
        if (overrideProps.size() == 1 && overrideProps.containsKey(RULE_SCHEDULER_DATA_KEY)) {
            isEmpty = true;
        }
        return isEmpty;
    }

    public ByteArrayOutputStream getCompressedModelMetadata(String project, List modelList,
            boolean exportRecommendations, boolean exportOverProps, boolean exportMultiplePartition) throws Exception {
        aclEvaluate.checkProjectWritePermission(project);
        NDataModelManager modelManager = modelService.getManager(NDataModelManager.class, project);
        NIndexPlanManager indexPlanManager = modelService.getManager(NIndexPlanManager.class, project);

        ByteArrayOutputStream byteArrayOutputStream = new ByteArrayOutputStream();
        try (ZipOutputStream zipOutputStream = new ZipOutputStream(byteArrayOutputStream)) {
            ResourceStore oldResourceStore = modelManager.getStore();
            KylinConfig newConfig = KylinConfig.createKylinConfig(KylinConfig.getInstanceFromEnv());
            ResourceStore newResourceStore = new InMemResourceStore(newConfig);
            ResourceStore.setRS(newConfig, newResourceStore);

            RawResourceFilter projectFilter = RawResourceFilter.equalFilter("project", project);
            for (String modelId : modelList) {
                NDataModel dataModelDesc = modelManager.getDataModelDesc(modelId);
                if (Objects.isNull(dataModelDesc)) {
                    throw new KylinException(MODEL_ID_NOT_EXIST, modelId);
                }
                if (dataModelDesc.isBroken()) {
                    throw new KylinException(MODEL_EXPORT_ERROR,
                            String.format(Locale.ROOT, MsgPicker.getMsg().getExportBrokenModel(), modelId));
                }

                NDataModel modelDesc = modelManager.copyForWrite(dataModelDesc);

                IndexPlan copyIndexPlan = indexPlanManager.copy(indexPlanManager.getIndexPlan(modelId));

                if (!exportOverProps) {
                    LinkedHashMap overridePropes = Maps.newLinkedHashMap();
                    if (copyIndexPlan.getOverrideProps().get(BASE_CUBOID_ALWAYS_VALID_KEY) != null) {
                        overridePropes.put(BASE_CUBOID_ALWAYS_VALID_KEY,
                                copyIndexPlan.getOverrideProps().get(BASE_CUBOID_ALWAYS_VALID_KEY));
                    }
                    if (copyIndexPlan.getOverrideProps().containsKey(RULE_SCHEDULER_DATA_KEY)) {
                        overridePropes.put(RULE_SCHEDULER_DATA_KEY,
                                copyIndexPlan.getOverrideProps().get(RULE_SCHEDULER_DATA_KEY));
                    }
                    copyIndexPlan.setOverrideProps(overridePropes);
                    modelDesc.setSegmentConfig(new SegmentConfig());
                }

                if (!exportMultiplePartition && modelDesc.isMultiPartitionModel()) {
                    modelDesc.setMultiPartitionDesc(
                            new MultiPartitionDesc(modelDesc.getMultiPartitionDesc().getColumns()));
                }

                newResourceStore.putResourceWithoutCheck(modelDesc.getResourcePath(),
                        ByteSource.wrap(JsonUtil.writeValueAsIndentBytes(modelDesc)), modelDesc.getLastModified(),
                        modelDesc.getMvcc());

                newResourceStore.putResourceWithoutCheck(copyIndexPlan.getResourcePath(),
                        ByteSource.wrap(JsonUtil.writeValueAsIndentBytes(copyIndexPlan)),
                        copyIndexPlan.getLastModified(), copyIndexPlan.getMvcc());

                // Broken model can't use getAllTables method, will be intercepted in BrokenEntityProxy
                Set tables = modelDesc.getAllTables().stream().map(TableRef::getTableDesc)
                        .map(TableDesc::getResourcePath)
                        .filter(resPath -> !newResourceStore
                                .listResourcesRecursively(MetadataType.TABLE_INFO.name(), projectFilter)
                                .contains(resPath))
                        .collect(Collectors.toSet());
                tables.forEach(resourcePath -> oldResourceStore.copy(resourcePath, newResourceStore));

                if (exportRecommendations) {
                    exportRecommendations(project, modelId, newResourceStore);
                }
            }
            if (CollectionUtils.isEmpty(newResourceStore.listResourcesRecursively(MetadataType.MODEL.name()))) {
                throw new KylinException(MODEL_METADATA_FILE_ERROR, MsgPicker.getMsg().getExportAtLeastOneModel());
            }

            addComputedColumns(project, modelList, newResourceStore);

            // add version file
            String version = System.getProperty(KE_VERSION) == null ? "unknown" : System.getProperty(KE_VERSION);
            StringEntity versionEntity = new StringEntity(VERSION_FILE_META_KEY_TAG, version);
            newResourceStore.putResourceWithoutCheck(VERSION_FILE,
                    ByteSource.wrap(JsonUtil.writeValueAsIndentBytes(versionEntity)), System.currentTimeMillis(), -1);

            oldResourceStore.copy(ResourceStore.METASTORE_UUID_TAG, newResourceStore);
            writeMetadataToZipOutputStream(zipOutputStream, newResourceStore);
        }
        return byteArrayOutputStream;
    }

    private void addComputedColumns(String project, List modelList, ResourceStore newResourceStore)
            throws JsonProcessingException {
        ComputedColumnManager ccManager = modelService.getManager(ComputedColumnManager.class, project);
        Manager relationManager = Manager.getInstance(getConfig(), project,
                CcModelRelationDesc.class);
        List relations = relationManager.listByFilter(new RawResourceFilter()
                .addConditions("modelUuid", new ArrayList<>(modelList), RawResourceFilter.Operator.IN));
        List usedCcs = ccManager.listByFilter(new RawResourceFilter().addConditions("metaKey",
                relations.stream().map(CcModelRelationDesc::getCcUuid).collect(Collectors.toList()),
                RawResourceFilter.Operator.IN));
        for (ComputedColumnDesc cc : usedCcs) {
            newResourceStore.putResourceWithoutCheck(cc.getResourcePath(),
                    ByteSource.wrap(JsonUtil.writeValueAsIndentBytes(cc)), cc.getLastModified(), cc.getMvcc());
        }
    }

    private void exportRecommendations(String project, String modelId, ResourceStore resourceStore) throws Exception {
        val projectManager = getManager(NProjectManager.class);
        val projectInstance = projectManager.getProject(project);
        if (projectInstance.isExpertMode()) {
            logger.info("Skip export recommendations because project {} is expert mode.", project);
            return;
        }
        JdbcRawRecStore jdbcRawRecStore = new JdbcRawRecStore(KylinConfig.getInstanceFromEnv());

        val optRecV2 = OptRecManagerV2.getInstance(project).loadOptRecV2(modelId);
        val rawRecIds = Stream
                .of(optRecV2.getCcRefs().keySet(), optRecV2.getMeasureRefs().keySet(),
                        optRecV2.getDimensionRefs().keySet(), optRecV2.getAdditionalLayoutRefs().keySet(),
                        optRecV2.getRemovalLayoutRefs().keySet()) //
                .flatMap(Collection::stream) //
                .filter(dependId -> dependId < 0) //
                .map(dependId -> -dependId) //
                .filter(dependId -> !optRecV2.getBrokenRefIds().contains(dependId)) //
                .collect(Collectors.toSet());

        if (rawRecIds.isEmpty()) {
            return;
        }

        List rawRecItems = jdbcRawRecStore.list(rawRecIds).stream()
                .sorted(Comparator.comparingInt(RawRecItem::getId)).collect(Collectors.toList());
        RecItemSet recEntity = new RecItemSet(modelId, project, rawRecItems);

        resourceStore.putResourceWithoutCheck(recEntity.getResourcePath(),
                ByteSource.wrap(JsonUtil.writeValueAsIndentBytes(recEntity)), System.currentTimeMillis(), -1);
    }

    private void writeMetadataToZipOutputStream(ZipOutputStream zipOutputStream, ResourceStore resourceStore)
            throws IOException {
        for (String resPath : resourceStore.listResourcesRecursively(MetadataType.ALL.name())) {
            zipOutputStream.putNextEntry(new ZipEntry(resPath + JSON_SUFFIX));
            zipOutputStream.write(resourceStore.getResource(resPath).getByteSource().read());
        }
    }

    @VisibleForTesting
    protected static Map getRawResourceFromUploadFile(MultipartFile uploadFile)
            throws IOException {
        val resourceMap = FileSystemMetadataStore.getFilesFromCompressedFileByStream(uploadFile.getInputStream(),
                new FileSystemMetadataStore.CompressHandler());
        val filesFromCompressedFile = Maps. newHashMap();
        resourceMap.forEach((k, v) -> filesFromCompressedFile.put(k.replaceAll(".json", ""), v));
        return filesFromCompressedFile;
    }

    private ImportModelContext getImportModelContext(String targetProject, Map rawResourceMap,
            ModelImportRequest request) {
        String srcProject = getModelMetadataProjectName(rawResourceMap);

        if (request != null) {
            val newModels = request.getModels().stream()
                    .filter(modelImport -> modelImport.getImportType() == ImportType.NEW)
                    .collect(Collectors.toMap(ModelImportRequest.ModelImport::getOriginalName,
                            ModelImportRequest.ModelImport::getTargetName));

            val unImportModels = request.getModels().stream()
                    .filter(modelImport -> modelImport.getImportType() == ImportType.UN_IMPORT)
                    .map(ModelImportRequest.ModelImport::getOriginalName).collect(Collectors.toList());

            return new ImportModelContext(targetProject, srcProject, rawResourceMap, newModels, unImportModels);
        } else {
            return new ImportModelContext(targetProject, srcProject, rawResourceMap);
        }
    }

    public SchemaChangeCheckResult checkModelMetadata(String targetProject, MultipartFile uploadFile,
            ModelImportRequest request) throws IOException {
        String originalFilename = uploadFile.getOriginalFilename();
        Matcher matcher = MD5_PATTERN.matcher(originalFilename);
        boolean valid = false;
        if (matcher.matches()) {
            String signature = matcher.group(1);
            try (InputStream inputStream = uploadFile.getInputStream()) {
                byte[] md5 = HashFunction.MD5.checksum(inputStream);
                valid = StringUtils.equalsIgnoreCase(signature, DatatypeConverter.printHexBinary(md5));
            }
        }

        if (!valid) {
            throw new KylinException(MODEL_METADATA_FILE_ERROR, MsgPicker.getMsg().getIllegalModelMetadataFile());
        }

        Map rawResourceMap = getRawResourceFromUploadFile(uploadFile);

        try (ImportModelContext context = getImportModelContext(targetProject, rawResourceMap, request)) {
            return checkModelMetadata(targetProject, context, uploadFile);
        }
    }

    public SchemaChangeCheckResult checkModelMetadata(String targetProject, ImportModelContext context,
            MultipartFile uploadFile) throws IOException {

        KylinConfig targetKylinConfig = context.getTargetKylinConfig();
        Map rawResourceMap = getRawResourceFromUploadFile(uploadFile);
        checkModelMetadataFile(ResourceStore.getKylinMetaStore(targetKylinConfig).getMetadataStore(),
                rawResourceMap.keySet());

        // check missing table exists in datasource
        List existTableList = searchTablesInDataSource(targetProject, context.getTargetMissTableList());
        // diff (local metadata + searched tables) and import metadata
        val diff = SchemaUtil.diff(targetProject, KylinConfig.getInstanceFromEnv(), targetKylinConfig, existTableList);
        SchemaChangeCheckResult checkResult = ModelImportChecker.check(diff, context);
        checkResult.getExistTableList().addAll(existTableList);
        return checkResult;
    }

    public List searchTablesInDataSource(String targetProject, List missTableList) {
        if (CollectionUtils.isEmpty(missTableList)) {
            return Collections.emptyList();
        }
        ProjectInstance projectInstance = NProjectManager.getInstance(KylinConfig.getInstanceFromEnv())
                .getProject(targetProject);
        ISourceMetadataExplorer explorer = SourceFactory.getSource(projectInstance).getSourceMetadataExplorer();
        KylinConfig config = KylinConfig.getInstanceFromEnv();
        List existTableSet = Lists.newArrayList();
        for (TableDesc missTableDesc : missTableList) {
            try {
                // check datasource exist table
                // no need to check column
                TableDesc newTableDesc = explorer
                        .loadTableMetadata(missTableDesc.getDatabase(), missTableDesc.getName(), targetProject)
                        .getFirst();
                newTableDesc.init(targetProject);
                existTableSet.add(newTableDesc);
            } catch (Exception e) {
                logger.warn("try load table: {} failed.", missTableDesc.getIdentity(), e);
            }
            if (config.isDDLLogicalViewEnabled() && missTableDesc.isLogicalView()) {
                LogicalView logicalView = LogicalViewManager.getInstance(config).get(missTableDesc.getName());
                if (logicalView != null && !targetProject.equalsIgnoreCase(logicalView.getCreatedProject())) {
                    throw new KylinException(FAILED_CREATE_MODEL,
                            String.format(Locale.ROOT, " Logical View %s can only add in project %s",
                                    missTableDesc.getName(), logicalView.getCreatedProject()));
                }
            }
        }
        return existTableSet;
    }

    private void checkModelMetadataFile(MetadataStore metadataStore, Set rawResourceList) {
        MetadataChecker metadataChecker = new MetadataChecker(metadataStore);
        MetadataChecker.VerifyResult verifyResult = metadataChecker
                .verifyModelMetadata(Lists.newArrayList(rawResourceList));
        if (!verifyResult.isModelMetadataQualified()) {
            throw new KylinException(MODEL_METADATA_FILE_ERROR, MsgPicker.getMsg().getModelMetadataPackageInvalid());
        }
    }

    @VisibleForTesting
    public static String getModelMetadataProjectName(Map rawResourceMap) {
        RawResource raw = rawResourceMap.values().stream()
                .filter(rawResource -> rawResource != null && rawResource.getProject() != null).findAny().orElse(null);
        if (raw == null) {
            throw new KylinException(MODEL_METADATA_FILE_ERROR, MsgPicker.getMsg().getModelMetadataPackageInvalid());
        }
        return raw.getProject();
    }

    private void createNewModel(NDataModel nDataModel, ModelImportRequest.ModelImport modelImport, String project,
            NIndexPlanManager importIndexPlanManager) {
        NDataModelManager dataModelManager = getManager(NDataModelManager.class, project);

        nDataModel.setProject(project);
        nDataModel.setAlias(modelImport.getTargetName());
        nDataModel.setUuid(RandomUtil.randomUUIDStr());
        nDataModel.setLastModified(System.currentTimeMillis());
        nDataModel.setMvcc(-1);
        dataModelManager.createDataModelDesc(nDataModel, AclPermissionUtil.getCurrentUsername());

        NIndexPlanManager indexPlanManager = getManager(NIndexPlanManager.class, project);
        NDataflowManager dataflowManager = getManager(NDataflowManager.class, project);
        var indexPlan = importIndexPlanManager.getIndexPlanByModelAlias(modelImport.getTargetName()).copy();
        indexPlan.setUuid(nDataModel.getUuid());
        indexPlan = indexPlanManager.copy(indexPlan);
        indexPlan.setLastModified(System.currentTimeMillis());
        indexPlan.setMvcc(-1);

        indexPlanManager.createIndexPlan(indexPlan);
        dataflowManager.createDataflow(indexPlan, nDataModel.getOwner(), RealizationStatusEnum.OFFLINE);
        indexPlanService.checkPartitionDimensionForV3Storage(project, nDataModel.getId(), getConfig());
    }

    private void updateModel(String project, NDataModel nDataModel, ModelImportRequest.ModelImport modelImport,
            boolean hasModelOverrideProps) {
        NDataModelManager dataModelManager = getManager(NDataModelManager.class, project);
        NDataModel originalDataModel = dataModelManager.getDataModelDescByAlias(modelImport.getOriginalName());
        nDataModel.setProject(project);
        nDataModel.setUuid(originalDataModel.getUuid());
        nDataModel.setLastModified(System.currentTimeMillis());

        // multiple partition column
        if (nDataModel.isMultiPartitionModel()) {
            if (!nDataModel.getMultiPartitionDesc().getPartitions().isEmpty()) {
                originalDataModel = modelService.batchUpdateMultiPartition(project, nDataModel.getUuid(),
                        nDataModel.getMultiPartitionDesc().getPartitions().stream()
                                .map(MultiPartitionDesc.PartitionInfo::getValues).collect(Collectors.toList()));
            } else {
                // keep original mapping
                nDataModel.setMultiPartitionKeyMapping(originalDataModel.getMultiPartitionKeyMapping());
            }
            nDataModel.setMultiPartitionDesc(originalDataModel.getMultiPartitionDesc());
        }

        if (!hasModelOverrideProps) {
            nDataModel.setSegmentConfig(originalDataModel.getSegmentConfig());
        }

        nDataModel.setMvcc(originalDataModel.getMvcc());

        dataModelManager.updateDataModelDesc(nDataModel);
    }

    private void updateIndexPlan(String project, NDataModel nDataModel, IndexPlan targetIndexPlan,
            boolean hasModelOverrideProps) {
        NIndexPlanManager indexPlanManager = getManager(NIndexPlanManager.class, project);
        indexPlanManager.updateIndexPlan(nDataModel.getUuid(), copyForWrite -> {
            List toBeDeletedIndexes = copyForWrite.getToBeDeletedIndexes();
            toBeDeletedIndexes.clear();
            toBeDeletedIndexes.addAll(targetIndexPlan.getToBeDeletedIndexes());
            copyForWrite.updateNextId();
        });
        if (targetIndexPlan.getRuleBasedIndex() != null) {
            indexPlanService.updateRuleBasedCuboid(project, UpdateRuleBasedCuboidRequest.convertToRequest(project,
                    nDataModel.getUuid(), false, targetIndexPlan.getRuleBasedIndex()));
        } else {
            indexPlanService.updateRuleBasedCuboid(project, UpdateRuleBasedCuboidRequest.convertToRequest(project,
                    nDataModel.getUuid(), false, new RuleBasedIndex()));
        }

        indexPlanManager.updateIndexPlan(nDataModel.getUuid(), copyForWrite -> {
            if (hasModelOverrideProps) {
                copyForWrite.setOverrideProps(targetIndexPlan.getOverrideProps());
            }

            if (targetIndexPlan.getAggShardByColumns() != null) {
                copyForWrite.setRuleBasedIndex(targetIndexPlan.getRuleBasedIndex());
                copyForWrite.setAggShardByColumns(targetIndexPlan.getAggShardByColumns());
            }
        });
    }

    private void removeIndexes(String project, SchemaChangeCheckResult.ModelSchemaChange modelSchemaChange,
            IndexPlan targetIndexPlan) {
        if (modelSchemaChange != null) {
            val newLockedItems = modelSchemaChange.getNewItems().stream()
                    .filter(item -> item.getType() == SchemaNodeType.TO_BE_DELETED_INDEX).collect(Collectors.toSet());
            val newLockedItemKeyAttrMap = Maps.newHashMap();
            newLockedItems.forEach(newLockedItem -> newLockedItemKeyAttrMap.put(newLockedItem.getSchemaNode().getKey(),
                    newLockedItem.getAttributes()));

            // filter indexes which should be removed form dataflow
            val toBeRemovedIndexes = Stream
                    .concat(modelSchemaChange.getReduceItems().stream()
                            .filter(schemaChange -> schemaChange.getType() == SchemaNodeType.WHITE_LIST_INDEX
                                    || schemaChange.getType() == SchemaNodeType.RULE_BASED_INDEX)
                            .filter(schemaChange -> {
                                val reduceItemKey = schemaChange.getSchemaNode().getKey();
                                val reduceItemAttr = schemaChange.getAttributes();
                                // 'toBeRemovedIndexes' should not contain locked indexes
                                return !reduceItemAttr.equals(newLockedItemKeyAttrMap.get(reduceItemKey));
                            }).map(SchemaChangeCheckResult.ChangedItem::getDetail),
                            modelSchemaChange.getUpdateItems().stream()
                                    .filter(schemaUpdate -> schemaUpdate.getType() == SchemaNodeType.WHITE_LIST_INDEX
                                            || schemaUpdate.getType() == SchemaNodeType.RULE_BASED_INDEX)
                                    .map(SchemaChangeCheckResult.UpdatedItem::getFirstDetail))
                    .map(Long::parseLong).collect(Collectors.toSet());
            if (!toBeRemovedIndexes.isEmpty()) {
                indexPlanService.removeIndexes(project, targetIndexPlan.getId(), toBeRemovedIndexes);
            }

            // for locked layout, just remove from 'indexes' json fields, keep in dataflow
            Set newLockedIndexIds = newLockedItems.stream().map(SchemaChangeCheckResult.ChangedItem::getDetail)
                    .map(Long::parseLong).collect(Collectors.toSet());
            removeLockedLayoutFromIndexes(newLockedIndexIds, targetIndexPlan, project);
        }
    }

    private void removeLockedLayoutFromIndexes(Set newLockedIndexIds, IndexPlan targetIndexPlan, String project) {
        NIndexPlanManager indexPlanManager = NIndexPlanManager.getInstance(getConfig(), project);
        indexPlanManager.updateIndexPlan(targetIndexPlan.getId(), copyForWrite -> {
            copyForWrite.removeLayouts(newLockedIndexIds, true, true);
            // set locked indexes, to be avoided from deleting in dataflow
            List toBeDeletedIndexes = copyForWrite.getToBeDeletedIndexes();
            toBeDeletedIndexes.clear();
            toBeDeletedIndexes.addAll(targetIndexPlan.getToBeDeletedIndexes());
        });
    }

    private void addWhiteListIndex(String project, SchemaChangeCheckResult.ModelSchemaChange modelSchemaChange,
            IndexPlan targetIndexPlan) {
        if (modelSchemaChange != null) {
            val newIndexes = Stream
                    .concat(modelSchemaChange.getNewItems().stream()
                            .filter(schemaChange -> schemaChange.getType() == SchemaNodeType.WHITE_LIST_INDEX)
                            .map(SchemaChangeCheckResult.ChangedItem::getDetail),
                            modelSchemaChange.getUpdateItems().stream()
                                    .filter(schemaUpdate -> schemaUpdate.getType() == SchemaNodeType.WHITE_LIST_INDEX)
                                    .map(SchemaChangeCheckResult.UpdatedItem::getSecondDetail))
                    .map(Long::parseLong).collect(Collectors.toList());

            val indexPlanManager = NIndexPlanManager.getInstance(KylinConfig.getInstanceFromEnv(), project);
            indexPlanManager.updateIndexPlan(targetIndexPlan.getUuid(), copyForWrite -> {
                IndexPlan.IndexPlanUpdateHandler updateHandler = copyForWrite.createUpdateHandler();
                targetIndexPlan.getWhitelistLayouts().stream().filter(layout -> newIndexes.contains(layout.getId()))
                        .forEach(layout -> updateHandler.add(layout, IndexEntity.isAggIndex(layout.getId())));
                updateHandler.complete();
            });
        }
    }

    private void addRuleBasedIndex(String project, SchemaChangeCheckResult.ModelSchemaChange modelSchemaChange,
            IndexPlan targetIndexPlan) {
        if (modelSchemaChange != null) {
            val newIndexes = Stream
                    .concat(modelSchemaChange.getNewItems().stream()
                            .filter(schemaChange -> schemaChange.getType() == SchemaNodeType.RULE_BASED_INDEX)
                            .map(SchemaChangeCheckResult.ChangedItem::getDetail),
                            modelSchemaChange.getUpdateItems().stream()
                                    .filter(schemaUpdate -> schemaUpdate.getType() == SchemaNodeType.RULE_BASED_INDEX)
                                    .map(SchemaChangeCheckResult.UpdatedItem::getSecondDetail))
                    .map(Long::parseLong).collect(Collectors.toList());
            val indexPlanManager = NIndexPlanManager.getInstance(KylinConfig.getInstanceFromEnv(), project);
            indexPlanManager.updateIndexPlan(targetIndexPlan.getUuid(), copyForWrite -> {
                IndexPlan.IndexPlanUpdateHandler updateHandler = copyForWrite.createUpdateHandler();
                targetIndexPlan.getRuleBaseLayouts().stream().filter(layout -> newIndexes.contains(layout.getId()))
                        .forEach(layout -> updateHandler.add(layout, IndexEntity.isAggIndex(layout.getId())));
                updateHandler.complete();
            });
        }
    }

    @Transaction(project = 0, retry = 1)
    public void importModelMetadata(String project, MultipartFile metadataFile, ModelImportRequest request)
            throws Exception {
        aclEvaluate.checkProjectWritePermission(project);

        List exceptions = new ArrayList<>();
        val rawResourceMap = getRawResourceFromUploadFile(metadataFile);
        try (val importModelContext = getImportModelContext(project, rawResourceMap, request)) {
            innerImportModelMetadata(project, metadataFile, request, importModelContext, exceptions);
        }
        if (!exceptions.isEmpty()) {
            String details = exceptions.stream().map(Exception::getMessage).collect(Collectors.joining("\n"));

            throw new KylinException(MODEL_IMPORT_ERROR,
                    String.format(Locale.ROOT, "%s%n%s", MsgPicker.getMsg().getImportModelException(), details),
                    exceptions);
        }
    }

    public LoadTableResponse innerLoadTables(String project, Set needLoadTables) throws Exception {
        return tableExtService.loadDbTables(needLoadTables.toArray(new String[0]), project, false);
    }

    public Pair, Map>> checkNewModelTables(SchemaChangeCheckResult checkResult,
            ModelImportRequest request) {
        List existTableList = checkResult.getExistTableList().stream().map(TableDesc::getIdentity)
                .collect(Collectors.toList());
        List newImportModelList = request.getModels().stream()
                .filter(modelRequest -> modelRequest.getImportType() == ImportType.NEW)
                .map(ModelImportRequest.ModelImport::getTargetName).collect(Collectors.toList());
        // all tables need to be loaded
        Set needLoadTableSet = Sets.newHashSet();
        // every model need to be loaded tables
        Map> modelTablesMap = Maps.newHashMap();

        checkResult.getModels().forEach((modelName, change) -> {
            if (!newImportModelList.contains(modelName) || !change.creatable()) {
                return;
            }
            Set modelTables = Sets.newHashSet();
            change.getNewItems().stream()//
                    .filter(item -> item.getSchemaNode().getType() == MODEL_DIM
                            || item.getSchemaNode().getType() == MODEL_FACT)
                    .map(SchemaChangeCheckResult.ChangedItem::getDetail).filter(existTableList::contains)
                    .forEach(table -> {
                        needLoadTableSet.add(table);
                        modelTables.add(table);
                    });
            modelTablesMap.put(modelName, modelTables);
        });
        return Pair.newPair(needLoadTableSet, modelTablesMap);
    }

    private void innerImportModelMetadata(String project, MultipartFile metadataFile, ModelImportRequest request,
            ImportModelContext context, List exceptions) throws Exception {
        val schemaChangeCheckResult = checkModelMetadata(project, context, metadataFile);

        val pair = checkNewModelTables(schemaChangeCheckResult, request);
        Set needLoadTableSet = pair.getFirst();
        Map> modelTablesMap = pair.getSecond();

        LoadTableResponse loadTableResponse = null;
        boolean needLoadTable = CollectionUtils.isNotEmpty(needLoadTableSet);
        if (needLoadTable) {
            // try load tables
            String needLoadTableStr = String.join(",", needLoadTableSet);
            logger.info("try load tables: [{}]", needLoadTableStr);
            loadTableResponse = innerLoadTables(project, needLoadTableSet);
            if (CollectionUtils.isNotEmpty(loadTableResponse.getFailed())) {
                String loadFailedTables = String.join(",", loadTableResponse.getFailed());
                logger.warn("Load Table failed: [{}]", loadFailedTables);
            }
        }

        KylinConfig targetKylinConfig = context.getTargetKylinConfig();
        val importDataModelManager = NDataModelManager.getInstance(targetKylinConfig, project);
        val importIndexPlanManager = NIndexPlanManager.getInstance(targetKylinConfig, project);

        for (ModelImportRequest.ModelImport modelImport : request.getModels()) {
            try {
                validateModelImport(project, modelImport, schemaChangeCheckResult);
                if (modelImport.getImportType() == ImportType.NEW) {
                    if (needLoadTable) {
                        Set needLoadTables = modelTablesMap.getOrDefault(modelImport.getTargetName(),
                                Collections.emptySet());
                        if (!loadTableResponse.getLoaded().containsAll(needLoadTables)) {
                            logger.warn("Import model [{}] failed, skip import.", modelImport.getOriginalName());
                            continue;
                        }
                    }
                    var importDataModel = importDataModelManager.getDataModelDescByAlias(modelImport.getTargetName());
                    var nDataModel = importDataModelManager.copyForWrite(importDataModel);
                    createNewModel(nDataModel, modelImport, project, importIndexPlanManager);
                    importRecommendations(project, nDataModel.getUuid(), importDataModel.getUuid(), targetKylinConfig);
                } else if (modelImport.getImportType() == ImportType.OVERWRITE) {
                    val importDataModel = importDataModelManager.getDataModelDescByAlias(modelImport.getOriginalName());
                    val nDataModel = importDataModelManager.copyForWrite(importDataModel);

                    // delete index, then remove dimension or measure
                    indexPlanService.checkPartitionDimensionForV3Storage(project, importDataModel.getId(),
                            targetKylinConfig);
                    val targetIndexPlan = importIndexPlanManager.getIndexPlanByModelAlias(modelImport.getOriginalName())
                            .copy();

                    boolean hasModelOverrideProps = (nDataModel.getSegmentConfig() != null
                            && nDataModel.getSegmentConfig().getAutoMergeEnabled() != null
                            && nDataModel.getSegmentConfig().getAutoMergeEnabled())
                            || (!targetIndexPlan.getOverrideProps().isEmpty());

                    val modelSchemaChange = schemaChangeCheckResult.getModels().get(modelImport.getTargetName());

                    removeIndexes(project, modelSchemaChange, targetIndexPlan);
                    updateModel(project, nDataModel, modelImport, hasModelOverrideProps);
                    updateIndexPlan(project, nDataModel, targetIndexPlan, hasModelOverrideProps);
                    addWhiteListIndex(project, modelSchemaChange, targetIndexPlan);
                    addRuleBasedIndex(project, modelSchemaChange, targetIndexPlan);

                    importRecommendations(project, nDataModel.getUuid(), importDataModel.getUuid(), targetKylinConfig);
                }
            } catch (Exception e) {
                logger.warn("Import model {} exception", modelImport.getOriginalName(), e);
                exceptions.add(e);
            }
        }
    }

    private void validateModelImport(String project, ModelImportRequest.ModelImport modelImport,
            SchemaChangeCheckResult checkResult) {

        Message msg = MsgPicker.getMsg();

        if (modelImport.getImportType() == ImportType.OVERWRITE) {
            NDataModel dataModel = NDataModelManager.getInstance(KylinConfig.getInstanceFromEnv(), project)
                    .getDataModelDescByAlias(modelImport.getOriginalName());

            if (dataModel == null) {
                throw new KylinException(MODEL_IMPORT_ERROR, String.format(Locale.ROOT, msg.getCanNotOverwriteModel(),
                        modelImport.getOriginalName(), modelImport.getImportType()));
            }

            val modelSchemaChange = checkResult.getModels().get(modelImport.getOriginalName());

            if (modelSchemaChange == null || !modelSchemaChange.overwritable()) {
                String createType = null;
                if (modelSchemaChange != null && modelSchemaChange.creatable()) {
                    createType = "NEW";
                }
                throw new KylinException(MODEL_IMPORT_ERROR,
                        String.format(Locale.ROOT, msg.getUnSuitableImportType(createType), modelImport.getImportType(),
                                modelImport.getOriginalName()));
            }
        } else if (modelImport.getImportType() == ImportType.NEW) {

            if (!StringUtils.containsOnly(modelImport.getTargetName(), ModelService.VALID_NAME_FOR_MODEL)) {
                throw new KylinException(MODEL_NAME_INVALID, modelImport.getTargetName());
            }

            NDataModel dataModel = NDataModelManager.getInstance(KylinConfig.getInstanceFromEnv(), project)
                    .getDataModelDescByAlias(modelImport.getTargetName());

            if (dataModel != null) {
                throw new KylinException(MODEL_NAME_DUPLICATE, modelImport.getTargetName());
            }

            val modelSchemaChange = checkResult.getModels().get(modelImport.getTargetName());

            if (modelSchemaChange == null || !modelSchemaChange.creatable()) {
                throw new KylinException(MODEL_IMPORT_ERROR, String.format(Locale.ROOT,
                        msg.getUnSuitableImportType(null), modelImport.getImportType(), modelImport.getTargetName()));
            }

        }
    }

    private void importRecommendations(String project, String targetModelId, String srcModelId, KylinConfig kylinConfig)
            throws IOException {
        val projectManager = getManager(NProjectManager.class);
        val projectInstance = projectManager.getProject(project);
        if (projectInstance.isExpertMode()) {
            modelChangeSupporters.forEach(listener -> listener.onUpdateSingle(project, targetModelId));
            logger.info("Skip import recommendations because project {} is expert mode.", project);
            return;
        }

        val manager = RawRecManager.getInstance(project);

        List rawRecItems = ImportModelContext.parseRawRecItems(ResourceStore.getKylinMetaStore(kylinConfig),
                project, srcModelId);

        manager.importRecommendations(project, targetModelId, rawRecItems);

        modelChangeSupporters.forEach(listener -> listener.onUpdateSingle(project, targetModelId));
    }

    public void cleanupMeta(String project) {
        if (project.equals(UnitOfWork.GLOBAL_UNIT)) {
            RoutineToolHelper.cleanGlobalSourceUsage();
            RoutineToolHelper.cleanQueryHistoriesAsync().join();
        } else {
            RoutineToolHelper.cleanMetaByProject(project);
        }
    }

    public void cleanupStorage(String[] projectsToClean, boolean cleanupStorage) {
        CleanTaskExecutorService.getInstance().cleanStorageForService(cleanupStorage, Arrays.asList(projectsToClean),
                0D, 0);
    }

    public void cleanupStorage(StorageCleanupRequest request, HttpServletRequest servletRequest) {
        if (routeService.needRoute()) {
            String url = StringUtils.stripEnd(servletRequest.getRequestURI(), "/") + "/tenant_node";
            routeService.asyncRouteForMultiTenantMode(servletRequest, url);
            return;
        }
        cleanupStorage(request.getProjectsToClean(), request.isCleanupStorage());
    }
}




© 2015 - 2025 Weber Informatics LLC | Privacy Policy