All Downloads are FREE. Search and download functionalities are using the official Maven repository.

au.csiro.pathling.security.ga4gh.ScopeAwareDatabase Maven / Gradle / Ivy

There is a newer version: 7.0.1
Show newest version
/*
 * Copyright 2023 Commonwealth Scientific and Industrial Research
 * Organisation (CSIRO) ABN 41 687 119 230.
 *
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

package au.csiro.pathling.security.ga4gh;

import au.csiro.pathling.config.ServerConfiguration;
import au.csiro.pathling.encoders.FhirEncoders;
import au.csiro.pathling.io.CacheableDatabase;
import au.csiro.pathling.io.Database;
import au.csiro.pathling.io.source.DataSource;
import au.csiro.pathling.terminology.TerminologyServiceFactory;
import ca.uhn.fhir.context.FhirContext;
import java.util.Optional;
import javax.annotation.Nonnull;
import javax.annotation.Nullable;
import lombok.extern.slf4j.Slf4j;
import org.apache.spark.sql.Dataset;
import org.apache.spark.sql.Row;
import org.apache.spark.sql.SparkSession;
import org.hl7.fhir.r4.model.Enumerations.ResourceType;
import org.springframework.context.annotation.Profile;
import org.springframework.scheduling.concurrent.ThreadPoolTaskExecutor;
import org.springframework.stereotype.Component;

/**
 * A special type of {@link Database} that is capable of limiting the scope of resources based upon
 * the passport scopes that have been applied to this request.
 *
 * @author John Grimes
 */
@Component
@Profile("core & ga4gh")
@Slf4j
public class ScopeAwareDatabase extends CacheableDatabase {

  @Nonnull
  private final ServerConfiguration configuration;

  @Nonnull
  private final FhirContext fhirContext;

  @Nonnull
  private final Optional terminologyServiceFactory;

  @Nonnull
  private final Optional passportScope;

  /**
   * @param configuration a {@link ServerConfiguration} object to control the behaviour of the
   * executor
   * @param fhirContext a {@link FhirContext} for doing FHIR stuff
   * @param spark a {@link SparkSession} for interacting with Spark
   * @param fhirEncoders {@link FhirEncoders} object for creating empty datasets
   * @param terminologyServiceFactory a {@link TerminologyServiceFactory} for resolving terminology
   * queries
   * @param passportScope a {@link PassportScope} that can be used to limit the scope of resources,
   * @param executor a {@link ThreadPoolTaskExecutor} for executing background tasks
   */
  @SuppressWarnings("WeakerAccess")
  public ScopeAwareDatabase(@Nonnull final ServerConfiguration configuration,
      @Nonnull final FhirContext fhirContext,
      @Nonnull final SparkSession spark,
      @Nonnull final FhirEncoders fhirEncoders,
      @Nonnull final Optional terminologyServiceFactory,
      @Nonnull final Optional passportScope,
      @Nonnull final ThreadPoolTaskExecutor executor) {
    super(configuration.getStorage(), spark, fhirEncoders, executor);
    log.debug("Initializing passport scope-aware resource reader");

    this.configuration = configuration;
    this.fhirContext = fhirContext;
    this.terminologyServiceFactory = terminologyServiceFactory;
    this.passportScope = passportScope;
  }

  @Nonnull
  @Override
  @SuppressWarnings("unused")
  public Dataset read(@Nullable final ResourceType resourceType) {
    final Dataset resources = super.read(resourceType);

    // If a passport scope is present, enforce the filters within it before returning the final
    // dataset.
    return passportScope
        .map(scope -> {
          // We need to create a non-scope-aware reader here for the parsing of the filters, so that
          // we don't have recursive application of the filters.
          final DataSource dataSource = Database.forConfiguration(spark, fhirEncoders,
              configuration.getStorage());
          final PassportScopeEnforcer scopeEnforcer = new PassportScopeEnforcer(
              configuration.getQuery(),
              fhirContext, spark, dataSource, terminologyServiceFactory, scope);
          return scopeEnforcer.enforce(resourceType, resources);
        })
        .orElse(resources);
  }
}




© 2015 - 2024 Weber Informatics LLC | Privacy Policy