ml-modules.root.data-hub.5.impl.flow.mjs Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of marklogic-data-hub Show documentation
Show all versions of marklogic-data-hub Show documentation
Library for Creating an Operational Data Hub on MarkLogic
/**
Copyright (c) 2021 MarkLogic Corporation
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
*/
'use strict';
import Artifacts from "/data-hub/5/artifacts/core.mjs";
import Batch from "/data-hub/5/flow/batch.mjs";
import defaultConfig from "/com.marklogic.hub/config.mjs";
import provLib from "/data-hub/5/impl/prov.mjs";
import flowRunner from "/data-hub/5/flow/flowRunner.mjs";
import hubUtils from "/data-hub/5/impl/hub-utils.mjs";
import jobs from "/data-hub/5/impl/jobs.mjs";
import StepExecutionContext from "/data-hub/5/flow/stepExecutionContext.mjs";
import StepDefinition from "/data-hub/5/impl/stepDefinition.mjs";
import WriteQueue from "/data-hub/5/flow/writeQueue.mjs";
// Starting in 5.5, this is needed for backwards compatibility so that scaffolded modules can still
// refer to datahub.flow.flowUtils .
import consts from "/data-hub/5/impl/consts.mjs";
import flowUtils from "/data-hub/5/impl/flow-utils.mjs";
// define constants for caching expensive operations
const cachedFlows = {};
function releaseUnusedDocuments(contentArray, outputContentArray) {
const outputNodeSet = new Set(outputContentArray.map(content => content.value));
for (const contentObject of contentArray) {
if (!outputNodeSet.has(contentObject.value)) {
hubUtils.releaseDatabaseNodeFromContentObject(contentObject);
}
}
}
export default class Flow {
constructor(config) {
this.consts = consts;
this.config = config;
this.stepDefinition = new StepDefinition(config);
this.flowUtils = flowUtils;
}
getFlow(name) {
if (cachedFlows[name] === undefined) {
cachedFlows[name] = Artifacts.getFullFlow(name);
}
return cachedFlows[name];
}
/**
* Find records that match a query based on the given inputs. Each matching record is wrapped in a
* "content descriptor" object that is guaranteed to have at least a "uri" property.
*
* @param flowName
* @param stepNumber
* @param options This isn't just the runtime options provided by the user; because it is expected
* to be called by the processBatch endpoint, which is invoked by the Java ScriptStepRunner class,
* this is likely already the set of combined options. But it's not guaranteed to be the same,
* and so combinedOptions is still constructed within this function.
* @return {*}
*/
findMatchingContent(flowName, stepNumber, options) {
// getFlow will throw an error if the flow cannot be found
const flow = this.getFlow(flowName);
const flowStep = flow.steps[stepNumber];
if (!flowStep) {
httpUtils.throwBadRequest(`Could not find step '${stepNumber}' in flow '${flowName}'`);
}
const stepDefinition = this.stepDefinition.getStepDefinitionByNameAndType(flowStep.stepDefinitionName, flowStep.stepDefinitionType);
if (!stepDefinition) {
httpUtils.throwBadRequest(`Could not find a step definition with name '${flowStep.stepDefinitionName}' and type '${flowStep.stepDefinitionType}' for step '${stepNumber}' in flow '${flowName}'`);
}
const combinedOptions = flowUtils.makeCombinedOptions(flow, stepDefinition, stepNumber, options);
let query;
let uris = null;
if (options.uris) {
uris = hubUtils.normalizeToArray(options.uris);
if (combinedOptions.excludeAlreadyProcessed === true || combinedOptions.excludeAlreadyProcessed === "true") {
const stepId = flowStep.stepId ? flowStep.stepId : flowStep.name + "-" + flowStep.stepDefinitionType;
const filteredItems = this.filterItemsAlreadyProcessedByStep(uris, flowName, stepId);
if (filteredItems.length != uris.length) {
xdmp.trace(consts.TRACE_FLOW, 'excludeAlreadyProcessed filtered out some items; previous count: ' + uris.length + '; new count: ' + filteredItems.length);
}
uris = filteredItems;
}
if (combinedOptions.sourceQueryIsScript || (combinedOptions.sourceQueryIsModule && combinedOptions.isUpgradedLegacyFlow)) {
// When the source query is a script, map each item to a content object with the "uri" property containing the item value
return uris.map(uri => { return {uri}; });
}
query = cts.documentQuery(uris);
} else {
let sourceQuery = combinedOptions.sourceQuery || flow.sourceQuery;
query = sourceQuery ? xdmp.eval(sourceQuery) : null;
}
let sourceDatabase = combinedOptions.sourceDatabase || defaultConfig.STAGINGDATABASE;
return hubUtils.queryToContentDescriptorArray(query, combinedOptions, sourceDatabase);
}
/**
* Filters out each item from the items array that was already processed by the given flowName and stepId.
* The determination is based on Batch documents in the jobs database.
*
* @param items the array of items to process; this is referred to as "uris" outside the context of this function,
* but since these values are not necessarily URIs starting in 5.3.0, the term "items" is used here instead
* @param flowName
* @param stepId
* @returns an array of items that should still be processed
*/
filterItemsAlreadyProcessedByStep(items, flowName, stepId) {
// TODO This duplicates knowledge of hash construction with jobs.sjs. Will clean this up in 5.5 when we can create
// a better "utils" library that is not a class with a bunch of public functions.
const prefixAux = flowName + "|" + fn.lowerCase(stepId) + "|finished|";
let prefix = prefixAux.replace(`'`, '');
// A map is used in this script to avoid N calls to array.includes
return fn.head(xdmp.invokeFunction(() => {
const itemHashes = items.map(item => xdmp.hash64(`${prefix}${item}`));
const processedItemHashesMap = cts.values(cts.jsonPropertyReference('processedItemHashes'), null, ["map", "concurrent", "score-zero"], cts.andQuery([
cts.collectionQuery('Batch'),
cts.jsonPropertyRangeQuery('processedItemHashes', '=', itemHashes)
]));
return items.filter(item => !processedItemHashesMap[xdmp.hash64(`${prefix}${item}`)]);
},
{database: xdmp.database(this.config.JOBDATABASE)}
));
}
/**
* It's unlikely that this actually runs a "flow", unless the flow consists of one step and only one transaction is
* needed to run the step. More likely, this is really "process a batch of items for a step".
*
* @param flowName Required name of the flow to run
* @param jobId Required ID of the job associated with the execution of the given step and flow
* @param contentArray Array of content "descriptors", where each descriptor is expected to at least have a "uri" property.
* The value of the "uri" property is not necessarily a URI; if sourceQueryIsScript is true for the step, then
* the value of the "uri" property can be any string.
* @param options Unfortunately this is not consistently defined; depending on the client, it's either the runtime
* options provided by the user, or it's the already-combined options
* @param stepNumber The number of the step within the given flow to run
* @param interceptors Additional interceptors to run. Used with the merge REST endpoint
*/
runFlow(flowName, jobId, contentArray, options, stepNumber, interceptors = []) {
contentArray = contentArray || [];
const theFlow = this.getFlow(flowName);
if (!theFlow) {
throw Error('The flow with the name '+flowName+' could not be found.');
}
let jobDoc = jobs.getJob(jobId);
if (!(jobDoc || options.disableJobOutput)) {
jobDoc = jobs.createJob(flowName, jobId);
}
if (jobDoc) {
if (jobDoc.job) {
jobDoc = jobDoc.job;
}
jobId = jobDoc.jobId;
}
if (!stepNumber) {
stepNumber = 1;
}
const flowStep = theFlow.steps[stepNumber];
if (!flowStep) {
throw Error('Step '+stepNumber+' for the flow: '+flowName+' could not be found.');
}
const stepDefinition = this.stepDefinition.getStepDefinitionByNameAndType(flowStep.stepDefinitionName, flowStep.stepDefinitionType);
const stepExecutionContext = new StepExecutionContext(theFlow, stepNumber, stepDefinition, jobId, options);
const combinedOptions = stepExecutionContext.combinedOptions;
// add any additional interceptors
stepExecutionContext.flowStep.interceptors = (stepExecutionContext.flowStep.interceptors || []).concat(interceptors);
const batchItems = contentArray.map(contentObject => contentObject.uri);
let outputContentArray;
const writeQueue = new WriteQueue();
hubUtils.hubTrace(consts.TRACE_FLOW, `Running ${stepExecutionContext.describe()}; content array length: ${batchItems.length}`);
try {
outputContentArray = flowRunner.runStepAgainstSourceDatabase(stepExecutionContext, contentArray, writeQueue);
releaseUnusedDocuments(contentArray, outputContentArray);
} catch (error) {
stepExecutionContext.addStepErrorForEntireBatch(error, batchItems);
this.writeBatchDocumentIfEnabled(stepExecutionContext, jobDoc, batchItems, {});
throw error;
}
let writeTransactionInfo = {};
if (stepExecutionContext.stepOutputShouldBeWritten()) {
try {
const contentToWrite = writeQueue.getContentArray(stepExecutionContext.getTargetDatabase());
writeTransactionInfo = flowUtils.writeContentArray(contentToWrite, stepExecutionContext.getTargetDatabase(), provLib.getProvenanceWriteQueue());
} catch (e) {
this.handleWriteError(writeQueue, stepExecutionContext, e);
}
}
// This handles committing provenance to the JOBS db prior to 5.7
if (stepExecutionContext.provenanceIsEnabled()) {
provLib.commit();
}
this.writeBatchDocumentIfEnabled(stepExecutionContext, jobDoc, batchItems, writeTransactionInfo);
// This maps to the Java ResponseHolder class
const responseHolder = {
"jobId": stepExecutionContext.jobId,
// using failed/completed items length instead of content length since a step can create more or less documents than were passed to the step
"totalCount": stepExecutionContext.failedItems.length + stepExecutionContext.completedItems.length,
"errorCount": stepExecutionContext.failedItems.length,
"completedItems": stepExecutionContext.completedItems,
"failedItems": stepExecutionContext.failedItems,
"errors": stepExecutionContext.stepErrors
};
if (combinedOptions.fullOutput) {
responseHolder.documents = outputContentArray;
}
return responseHolder;
}
/**
* If batch output is enabled, a Batch document will be written to the jobs database.
*
* @param stepExecutionContext
* @param jobDoc {object} the job document for this step execution; can be null, in which case a Batch doc is not written
* @param batchItems {array} the set of items being processed in this batch
* @param writeTransactionInfo {object} info about the transaction(s) that wrote content to a database
*/
writeBatchDocumentIfEnabled(stepExecutionContext, jobDoc, batchItems, writeTransactionInfo) {
if (stepExecutionContext.batchOutputIsEnabled()) {
if (jobDoc != null) {
const batch = new Batch(stepExecutionContext.jobId, stepExecutionContext.flow.name);
batch.addSingleStepResult(stepExecutionContext, batchItems, writeTransactionInfo);
batch.persist();
} else {
hubUtils.hubTrace(consts.TRACE_FLOW,
"Batch document insertion is enabled, but job document is null, so unable to insert a batch document");
}
} else if (xdmp.traceEnabled(consts.TRACE_FLOW)) {
hubUtils.hubTrace(consts.TRACE_FLOW, `Batch document insertion is disabled`);
}
}
handleWriteError(writeQueue, stepExecutionContext, error) {
stepExecutionContext.clearCompletedItems();
// Note that in this scenario, and this scenario only, DHF 5 changes the definition of "failed items" to be
// "URIs written" (or "attempted to be written") instead of "items processed". This makes a difference when
// a step returns multiple content objects per item that it receives.
const failedItems = writeQueue.getContentUris(stepExecutionContext.getTargetDatabase());
stepExecutionContext.setFailedItems(failedItems);
const operation = error.stackFrames && error.stackFrames[0] && error.stackFrames[0].operation;
let uri;
// see if we can determine a uri based off the operation in the stackFrames
if (operation) {
uri = failedItems.find(item => operation.includes(`"${item}"`));
}
// Directly add this to avoid the failedItems count from being incremented
stepExecutionContext.stepErrors.push(Object.assign(error, {"uri": uri}));
}
}
© 2015 - 2024 Weber Informatics LLC | Privacy Policy