Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance. Project price only 1 $
You can buy this project and download/modify it how often you want.
/*
* Copyright (C) 2014 Information Management Services, Inc.
*/
package com.imsweb.staging;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.Calendar;
import java.util.Collections;
import java.util.HashMap;
import java.util.HashSet;
import java.util.List;
import java.util.Map;
import java.util.Map.Entry;
import java.util.Set;
import java.util.stream.Collectors;
import com.imsweb.decisionengine.ColumnDefinition.ColumnType;
import com.imsweb.decisionengine.DecisionEngine;
import com.imsweb.decisionengine.Error;
import com.imsweb.decisionengine.Output;
import com.imsweb.decisionengine.Result;
import com.imsweb.decisionengine.Result.Type;
import com.imsweb.decisionengine.Table;
import com.imsweb.staging.entities.StagingColumnDefinition;
import com.imsweb.staging.entities.StagingMapping;
import com.imsweb.staging.entities.StagingSchema;
import com.imsweb.staging.entities.StagingSchemaInput;
import com.imsweb.staging.entities.StagingTable;
import com.imsweb.staging.entities.StagingTablePath;
public final class Staging {
// context keys definitions
public static final String CTX_ALGORITHM_VERSION = "ctx_alg_version";
public static final String CTX_YEAR_CURRENT = "ctx_year_current";
// list of all context keys
public static final List CONTEXT_KEYS = Collections.unmodifiableList(Arrays.asList(CTX_ALGORITHM_VERSION, CTX_YEAR_CURRENT));
private DecisionEngine _engine;
private StagingDataProvider _provider;
/**
* Private constructor
* @param provider data provider
*/
private Staging(StagingDataProvider provider) {
_provider = provider;
_engine = new DecisionEngine(provider);
}
/**
* Create an instance of the Staging object with a pre-constructed provider
* @param provider StagingDataProvider
* @return a Staging instance
*/
public static Staging getInstance(StagingDataProvider provider) {
return new Staging(provider);
}
/**
* Return the algorithm name
* @return an Algorithm
*/
public String getAlgorithm() {
return _provider.getAlgorithm();
}
/**
* Return the data version
* @return a version string
*/
public String getVersion() {
return _provider.getVersion();
}
/**
* Return the list a schema ids
* @return a Set of Strings of schema ids
*/
public Set getSchemaIds() {
return _provider.getSchemaIds();
}
/**
* Return a schema based on schema id
* @param id Schema identifier
* @return an Algorithm object
*/
public StagingSchema getSchema(String id) {
return _provider.getDefinition(id);
}
/**
* Return true if the site is valid
* @param site primary site
* @return true if the side is valid
*/
public boolean isValidSite(String site) {
return _provider.isValidSite(site);
}
/**
* Return true if the histology is valid
* @param histology histology
* @return true if the histology is valid
*/
public boolean isValidHistology(String histology) {
return _provider.isValidHistology(histology);
}
/**
* Look up a schema based on site, histology and an optional discriminator.
* @param lookup schema lookup input
* @return a list of StagingSchema objects
*/
public List lookupSchema(SchemaLookup lookup) {
return _provider.lookupSchema(lookup);
}
/**
* Return a complete list of mapping table names
* @return Set of String objects containing table names
*/
public Set getTableIds() {
return _provider.getTableIds();
}
/**
* Return a mapping table based on table name
* @param id table identifier
* @return Table object
*/
public StagingTable getTable(String id) {
return _provider.getTable(id);
}
/**
* Check the code validity of a single field in a schema. If the schema or field do no exist, false will be returned.
* @param schemaId schema identifier
* @param key input key
* @param value value to check validity
* @return a boolean indicating whether the code exists for the the passed schema field
*/
public boolean isCodeValid(String schemaId, String key, String value) {
Map context = new HashMap<>();
// add the value to the context
context.put(key, value);
// add the context keys
addContextKeys(context);
return isContextValid(schemaId, key, context);
}
/**
* Check the validity of a single field of a schema based on the supplied context. The value of this key should be in the context as well
* as any other properties needed to evaluation validity. If the schema or field do no exist, false will be returned.
* @param schemaId schema identifier
* @param key input key
* @param context Map of keys/values to validate against
* @return a boolean indicating whether the code exists for the the passed schema field
*/
public boolean isContextValid(String schemaId, String key, Map context) {
// first get the algorithm
StagingSchema schema = getSchema(schemaId);
if (schema == null)
return false;
// get the table id from the schema
StagingSchemaInput input = schema.getInputMap().get(key);
if (input == null)
return false;
// missing context will always return false
if (context == null || context.isEmpty())
return false;
// all context input needs to be trimmed
Map testContext = new HashMap<>();
for (Entry entry : context.entrySet())
testContext.put(entry.getKey(), entry.getValue() != null ? entry.getValue().trim() : "");
// if the input specifies a table for validation, test against it
if (input.getTable() != null) {
Table table = getTable(input.getTable());
return table != null && (DecisionEngine.matchTable(table, testContext) != null);
}
return true;
}
/**
* For a given table, return the index of the row that matches the key/value combination. If you need to match on more than
* one input, see the other findMatchingTableRow.
* @param tableId table identifier
* @param key input key
* @param value value
* @return the index of the matching table row or null if no match was found
*/
public Integer findMatchingTableRow(String tableId, String key, String value) {
Map context = new HashMap<>();
context.put(key, value);
return findMatchingTableRow(tableId, context);
}
/**
* For a given table, return the index of the row that matches supplied context.
* @param tableId table identifier
* @param context context of input key/values to use to match
* @return the index of the matching table row or null if no match was found
*/
public Integer findMatchingTableRow(String tableId, Map context) {
Integer rowIndex = null;
// add the context keys
addContextKeys(context);
Table table = getTable(tableId);
if (table != null)
rowIndex = DecisionEngine.findMatchingTableRow(table, context);
return rowIndex;
}
/**
* Return a list of tables identifiers involved in the specified schema
* @param schemaId schema identifier
* @return a Set of table identifiers; if the schema is not found the set will be empty
*/
public Set getInvolvedTables(String schemaId) {
Set tables = new HashSet<>();
StagingSchema schema = getSchema(schemaId);
if (schema != null && schema.getInvolvedTables() != null)
tables = schema.getInvolvedTables();
return tables;
}
/**
* Return a list of schema identifiers which the specified table is used in
* @param tableId table identifier
* @return a Set of schema identifiers
*/
public Set getInvolvedSchemas(String tableId) {
// loop over all schemas and find the ones that have the passed table identifier in the involved table set
return getSchemaIds().stream().filter(schemaId -> getInvolvedTables(schemaId).contains(tableId)).collect(Collectors.toSet());
}
/**
* Looks at all tables involved in the table path and returns a list of input keys that could be used. This also includes the input keys
* used in jumps tables.
* @param path a StagingTablePath
* @return a Set of unique input keys
*/
public Set getInputs(StagingTablePath path) {
return getInputs(path, new HashSet<>());
}
/**
* Looks at all tables involved in the table path and returns a list of input keys that could be used. This also includes the input keys
* used in jumps tables.
* @param path a StagingTablePath
* @param excludedInputs a list of input keys to not consider as inputs
* @return a Set of unique input keys
*/
public Set getInputs(StagingTablePath path, Set excludedInputs) {
Set inputs = new HashSet<>();
if (path != null)
inputs.addAll(_engine.getInputs(path, excludedInputs));
// always remove all context variables since they are never needed to be supplied
inputs.removeAll(CONTEXT_KEYS);
return inputs;
}
/**
* Looks at all tables involved in the mapping and returns a list of input keys that could be used. This also includes the input keys
* used in the inclusion and exclusion tables if any.
* @param mapping a StagingMapping
* @return a Set of unique input keys
*/
public Set getInputs(StagingMapping mapping) {
return getInputs(mapping, null, new HashSet<>());
}
/**
* Looks at all tables involved in the mapping and returns a list of input keys that could be used. This always includes the input keys
* used in the inclusion and exclusion tables if any. The inputs from each table path will only be included if it passes the inclusion/exclusion
* criteria based on the context.
* @param mapping a StagingMapping
* @param context a context of values used to to check mapping inclusion/exclusion
* @param excludedInputs a list of input keys to not consider as inputs
* @return a Set of unique input keys
*/
public Set getInputs(StagingMapping mapping, Map context, Set excludedInputs) {
Set inputs = new HashSet<>();
// the inclusion tables are always evaluated so any inputs there should be added always
if (mapping.getInclusionTables() != null)
for (StagingTablePath path : mapping.getInclusionTables())
inputs.addAll(getInputs(path, excludedInputs));
// the exclusion tables are always evaluated so any inputs there should be added always
if (mapping.getExclusionTables() != null)
for (StagingTablePath path : mapping.getExclusionTables())
inputs.addAll(getInputs(path, excludedInputs));
// if there are tables paths and if the mapping is involved, add the inputs
if (mapping.getTablePaths() != null)
if (context == null || _engine.isMappingInvolved(mapping, context))
inputs.addAll(_engine.getInputs(mapping, excludedInputs));
// always remove all context variables since they are never needed to be supplied
inputs.removeAll(CONTEXT_KEYS);
return inputs;
}
/**
* Looks at all tables involved in all the mappings in the definition and returns a list of input keys that could be used. It will also deal with mapped
* inputs. Note that if an input to a table was not a supplied input (i.e. it was created as an output of a previous table) it will not be included
* in the list of inputs. The inputs will also include any used in schema selection. All inputs returned from this method should be in the schema input
* list otherwise there is a problem with the schema.
* @param schema a StagingSchema
* @return a Set of unique input keys
*/
public Set getInputs(StagingSchema schema) {
return getInputs(schema, null);
}
/**
* Looks at all tables involved in all the mappings in the definition and returns a list of input keys that could be used. It will also deal with mapped
* inputs. The inputs from each mapping will only be included if it passes the inclusion/exclusion criteria based on the context. Note that if an input
* to a table was not a supplied input (i.e. it was created as an output of a previous table) it will not be included in the list of inputs. The inputs will
* also include any used in schema selection. All inputs returned from this method should be in the schema input list otherwise there is a problem with the
* schema.
* @param schema a StagingSchema
* @param context a context of values used to to check mapping inclusion/exclusion
* @return a Set of unique input keys
*/
public Set getInputs(StagingSchema schema, Map context) {
Set inputs = new HashSet<>();
// add schema selection fields
if (schema.getSchemaSelectionTable() != null) {
StagingTable table = getTable(schema.getSchemaSelectionTable());
if (table != null) {
for (StagingColumnDefinition def : table.getColumnDefinitions())
if (ColumnType.INPUT.equals(def.getType()))
inputs.add(def.getKey());
}
}
// process all mappings
if (schema.getMappings() != null) {
Set excludedInputs = new HashSet<>();
for (StagingMapping mapping : schema.getMappings())
inputs.addAll(getInputs(mapping, context, excludedInputs));
}
// always remove all context variables since they are never needed to be supplied
inputs.removeAll(CONTEXT_KEYS);
return inputs;
}
/**
* Looks at a table path (and all jump tables within in) and returns a list of output keys that could be created.
* @param path a StagingTablePath
* @return a Set of unique output keys
*/
public Set getOutputs(StagingTablePath path) {
return _engine.getOutputs(path);
}
/**
* Looks at all tables involved in the mapping and returns a list of output keys that could be created. This also includes the output keys
* used in the inclusion and exclusion tables if any.
* @param mapping a StagingMapping
* @return a Set of unique output keys
*/
public Set getOutputs(StagingMapping mapping) {
return getOutputs(mapping, null);
}
/**
* Looks at all tables involved in the mapping and returns a list of output keys that could be created. This also includes the output keys
* used in the inclusion and exclusion tables if any. The outputs from each mapping will only be included if it passes the inclusion/exclusion
* criteria based on the context.
* @param mapping a StagingMapping
* @param context a context of values used to to check mapping inclusion/exclusion
* @return a Set of unique output keys
*/
public Set getOutputs(StagingMapping mapping, Map context) {
Set outputs = new HashSet<>();
if (mapping.getTablePaths() != null) {
if (context == null || _engine.isMappingInvolved(mapping, context))
outputs.addAll(_engine.getOutputs(mapping));
}
return outputs;
}
/**
* Looks at all tables involved in all the mappings in the definition and returns a list of output keys that could be created. It will also deal with mapped
* outputs.
* @param schema a StagingSchema
* @return a Set of unique output keys
*/
public Set getOutputs(StagingSchema schema) {
return getOutputs(schema, null);
}
/**
* Looks at all tables involved in all the mappings in the definition and returns a list of output keys that will be created. It will also deal with mapped
* outputs. The outputs from each mapping will only be included if it passes the inclusion/exclusion criteria based on the context. If the schema has StagingOutputs
* defined, then the calulated output list is exactly the same as the schema output list.
* @param schema a StagingSchema
* @param context a context of values used to to check mapping inclusion/exclusion
* @return a Set of unique output keys
*/
public Set getOutputs(StagingSchema schema, Map context) {
Set outputs = new HashSet<>();
// if outputs are defined in the schema, then there is no reason to look any further into the mappings; the output defines exactly what keys will
// be returned and it doesn't matter what context is passed in that case
if (schema.getOutputMap() != null) {
for (Entry entry : schema.getOutputMap().entrySet())
outputs.add(entry.getKey());
return outputs;
}
// if outputs were not defined, then the tables involved in the mappings will be used to determine the possible outputs
if (schema.getMappings() != null)
for (StagingMapping mapping : schema.getMappings())
outputs.addAll(getOutputs(mapping, context));
// if valid outputs are defined on the schema level, only return outputs that defined; this removed "temporary" outputs that may be defined during the
// staging process
if (schema.getOutputMap() != null)
outputs.removeIf(entry -> !schema.getOutputMap().containsKey(entry));
return outputs;
}
/**
* Stage the passed case.
* @param data all input values are passed through this database
* @return the same StagingData with output values filled in
*/
public StagingData stage(StagingData data) {
// first clear out schema/output/errors/path
data.setSchemaId(null);
data.setOutput(new HashMap<>());
data.setErrors(new ArrayList<>());
data.setPath(new ArrayList<>());
// make sure site and histology are supplied
if (data.getInput(StagingData.PRIMARY_SITE_KEY) == null || data.getInput(StagingData.HISTOLOGY_KEY) == null) {
data.setResult(StagingData.Result.FAILED_MISSING_SITE_OR_HISTOLOGY);
return data;
}
// get the schema; if a single schema is not found, return right away with an error
List schemas = lookupSchema(new SchemaLookup(data.getInput()));
if (schemas.size() != 1) {
if (schemas.size() == 0)
data.setResult(StagingData.Result.FAILED_NO_MATCHING_SCHEMA);
else
data.setResult(StagingData.Result.FAILED_MULITPLE_MATCHING_SCHEMAS);
return data;
}
StagingSchema schema = schemas.get(0);
// add schema id to result
data.setSchemaId(schema.getId());
// copy the input into a new context
Map context = new HashMap<>(data.getInput());
// make sure all supplied inputs are defined in the definition
for (Entry entry : context.entrySet())
if (!schema.getInputMap().containsKey(entry.getKey()))
data.addError(new Error.ErrorBuilder(Error.Type.UNKNOWN_INPUT).message("Unknown input key supplied: " + entry.getKey()).key(entry.getKey()).build());
if (data.getErrors().size() > 0) {
data.setResult(StagingData.Result.FAILED_INVALID_INPUT);
return data;
}
// add context variables
addContextKeys(context);
// check that year of DX is valid
if (!isContextValid(schema.getId(), StagingData.YEAR_DX_KEY, context)) {
data.setResult(StagingData.Result.FAILED_INVALID_YEAR_DX);
return data;
}
// perform the staging
Result result = _engine.process(schemas.get(0).getId(), context);
// remove the context variables
removeContextKeys(context);
// set the staging data result based on the Result returned from the DecisionEngine
if (Type.FAILED_INPUT.equals(result.getType()))
data.setResult(StagingData.Result.FAILED_INVALID_INPUT);
else
data.setResult(StagingData.Result.STAGED);
// remove the original input keys from the resulting context; in addition, we want to remove any input keys
// from the resulting context that were set with a default value; to accomplish this remove all keys that are
// defined as input in the selected schema
for (Entry entry : data.getInput().entrySet())
context.remove(entry.getKey());
for (StagingSchemaInput input : schemas.get(0).getInputs())
context.remove(input.getKey());
// add the results to the data card
data.setOutput(result.getContext());
data.setErrors(result.getErrors());
data.setPath(result.getPath());
return data;
}
/**
* Add the context keys which are used in staging and other calls
* @param context Map of context
*/
private void addContextKeys(Map context) {
// make the algorithm version available in the context
context.put(CTX_ALGORITHM_VERSION, getVersion());
// put the current year in the context
Calendar now = Calendar.getInstance();
context.put(CTX_YEAR_CURRENT, String.valueOf(now.get(Calendar.YEAR)));
}
/**
* Remove all context keys
* @param context Map of context
*/
private void removeContextKeys(Map context) {
context.remove(CTX_ALGORITHM_VERSION);
context.remove(CTX_YEAR_CURRENT);
}
}