org.pentaho.di.job.entries.talendjobexec.JobEntryTalendJobExec Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of kettle-engine Show documentation
Show all versions of kettle-engine Show documentation
Container pom for Pentaho Data Integration modules
The newest version!
/*! ******************************************************************************
*
* Pentaho Data Integration
*
* Copyright (C) 2002-2018 by Hitachi Vantara : http://www.pentaho.com
*
*******************************************************************************
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*
******************************************************************************/
package org.pentaho.di.job.entries.talendjobexec;
import org.pentaho.di.job.entry.validator.AndValidator;
import org.pentaho.di.job.entry.validator.JobEntryValidatorUtils;
import java.io.File;
import java.lang.reflect.Method;
import java.net.URL;
import java.util.ArrayList;
import java.util.List;
import java.util.Map;
import java.util.concurrent.ConcurrentHashMap;
import org.apache.commons.vfs2.AllFileSelector;
import org.apache.commons.vfs2.FileObject;
import org.pentaho.di.cluster.SlaveServer;
import org.pentaho.di.core.CheckResultInterface;
import org.pentaho.di.core.Const;
import org.pentaho.di.core.util.Utils;
import org.pentaho.di.core.Result;
import org.pentaho.di.core.database.DatabaseMeta;
import org.pentaho.di.core.exception.KettleDatabaseException;
import org.pentaho.di.core.exception.KettleException;
import org.pentaho.di.core.exception.KettleXMLException;
import org.pentaho.di.core.fileinput.FileInputList;
import org.pentaho.di.core.plugins.KettleURLClassLoader;
import org.pentaho.di.core.variables.VariableSpace;
import org.pentaho.di.core.vfs.KettleVFS;
import org.pentaho.di.core.xml.XMLHandler;
import org.pentaho.di.i18n.BaseMessages;
import org.pentaho.di.job.JobMeta;
import org.pentaho.di.job.entry.JobEntryBase;
import org.pentaho.di.job.entry.JobEntryInterface;
import org.pentaho.di.repository.ObjectId;
import org.pentaho.di.repository.Repository;
import org.pentaho.di.resource.ResourceDefinition;
import org.pentaho.di.resource.ResourceEntry;
import org.pentaho.di.resource.ResourceEntry.ResourceType;
import org.pentaho.di.resource.ResourceNamingInterface;
import org.pentaho.di.resource.ResourceReference;
import org.pentaho.metastore.api.IMetaStore;
import org.w3c.dom.Node;
/**
* This executes an exported Talend Job.
*
* @author Matt
* @since 1-04-2011
*
*/
public class JobEntryTalendJobExec extends JobEntryBase implements Cloneable, JobEntryInterface {
private static Class> PKG = JobEntryTalendJobExec.class; // for i18n
private static Map classLoaderCache = new ConcurrentHashMap<>();
private String filename;
private String className;
public JobEntryTalendJobExec( String n ) {
super( n, "" );
filename = null;
}
public JobEntryTalendJobExec() {
this( "" );
}
public Object clone() {
JobEntryTalendJobExec je = (JobEntryTalendJobExec) super.clone();
return je;
}
public String getXML() {
StringBuilder retval = new StringBuilder( 50 );
retval.append( super.getXML() );
retval.append( " " ).append( XMLHandler.addTagValue( "filename", filename ) );
retval.append( " " ).append( XMLHandler.addTagValue( "class_name", className ) );
return retval.toString();
}
public void loadXML( Node entrynode, List databases, List slaveServers,
Repository rep, IMetaStore metaStore ) throws KettleXMLException {
try {
super.loadXML( entrynode, databases, slaveServers );
filename = XMLHandler.getTagValue( entrynode, "filename" );
className = XMLHandler.getTagValue( entrynode, "class_name" );
} catch ( KettleXMLException xe ) {
throw new KettleXMLException( BaseMessages.getString(
PKG, "JobEntryTalendJobExec.ERROR_0001_Cannot_Load_Job_Entry_From_Xml_Node" ), xe );
}
}
public void loadRep( Repository rep, IMetaStore metaStore, ObjectId id_jobentry, List databases,
List slaveServers ) throws KettleException {
try {
filename = rep.getJobEntryAttributeString( id_jobentry, "filename" );
className = rep.getJobEntryAttributeString( id_jobentry, "class_name" );
} catch ( KettleException dbe ) {
throw new KettleException( BaseMessages.getString(
PKG, "JobEntryTalendJobExec.ERROR_0002_Cannot_Load_Job_From_Repository", id_jobentry ), dbe );
}
}
public void saveRep( Repository rep, IMetaStore metaStore, ObjectId id_job ) throws KettleException {
try {
rep.saveJobEntryAttribute( id_job, getObjectId(), "filename", filename );
rep.saveJobEntryAttribute( id_job, getObjectId(), "class_name", className );
} catch ( KettleDatabaseException dbe ) {
throw new KettleException( BaseMessages.getString(
PKG, "JobEntryTalendJobExec.ERROR_0003_Cannot_Save_Job_Entry", id_job ), dbe );
}
}
public void setFilename( String filename ) {
this.filename = filename;
}
public String getFilename() {
return filename;
}
public String getRealFilename() {
return environmentSubstitute( getFilename() );
}
public Result execute( Result previousResult, int nr ) {
Result result = previousResult;
result.setResult( false );
if ( filename != null ) {
String realFilename = getRealFilename();
try {
FileObject file = KettleVFS.getFileObject( realFilename, this );
if ( file.exists() && file.isReadable() ) {
result = executeTalenJob( file, result, nr );
} else {
logDetailed( BaseMessages.getString( PKG, "JobEntryTalendJobExec.File_Does_Not_Exist", realFilename ) );
}
} catch ( Exception e ) {
result.setNrErrors( 1 );
logError(
BaseMessages.getString( PKG, "JobEntryTalendJobExec.ERROR_0004_IO_Exception", e.getMessage() ), e );
}
} else {
result.setNrErrors( 1 );
logError( BaseMessages.getString( PKG, "JobEntryTalendJobExec.ERROR_0005_No_Filename_Defined" ) );
}
return result;
}
private Result executeTalenJob( FileObject file, Result result, int nr ) throws Exception {
ClassLoader classLoader = null;
try {
classLoader = classLoaderCache.get( file.toString() );
if ( classLoader == null ) {
// Find the jar files in the ZIP file...
//
final URL[] jarFiles = prepareJarFiles( file );
// Create a new class loader with the extracted jar files.
//
classLoader = new KettleURLClassLoader( jarFiles, getClass().getClassLoader() );
Runtime.getRuntime().addShutdownHook( new Thread() {
@Override
public void run() {
try {
cleanupJarFiles( jarFiles );
} catch ( Exception e ) {
System.err.println( "Error cleaning up temporary Talend jar file extracts: "
+ Const.getStackTracker( e ) );
}
}
} );
classLoaderCache.put( file.toString(), classLoader );
}
Class> clazz = classLoader.loadClass( environmentSubstitute( getClassName() ) );
Object jobObject = clazz.newInstance();
Method runJob = clazz.getMethod( "runJobInTOS", String[].class );
// TODO: consider passing something of some sort in this next method:
// variables, arguments...
//
int returnCode = (Integer) runJob.invoke( jobObject, (Object) new String[] {} );
result.setExitStatus( returnCode );
result.setResult( true );
result.setNrErrors( 0 );
} catch ( Exception e ) {
throw new KettleException( BaseMessages.getString(
PKG, "JobEntryTalendJobExec.ERROR_0006_ExceptionExecutingTalenJob" ), e );
}
return result;
}
private URL[] prepareJarFiles( FileObject zipFile ) throws Exception {
// zip:file:///tmp/foo.zip
FileInputList fileList = FileInputList.createFileList( this, new String[] { "zip:" + zipFile.toString(), },
new String[] { ".*\\.jar$", }, // Include mask: only jar files
new String[] { ".*classpath\\.jar$", }, // Exclude mask: only jar files
new String[] { "Y", }, // File required
new boolean[] { true, } ); // Search sub-directories
List files = new ArrayList();
// Copy the jar files in the temp folder...
//
for ( FileObject file : fileList.getFiles() ) {
FileObject jarfilecopy =
KettleVFS.createTempFile(
file.getName().getBaseName(), ".jar", environmentSubstitute( "${java.io.tmpdir}" ) );
jarfilecopy.copyFrom( file, new AllFileSelector() );
files.add( jarfilecopy.getURL() );
}
return files.toArray( new URL[files.size()] );
}
private void cleanupJarFiles( URL[] jarFiles ) throws Exception {
if ( jarFiles == null ) {
return;
}
for ( URL jarFile : jarFiles ) {
File file = new File( jarFile.toURI() );
file.delete();
}
}
public boolean evaluates() {
return true;
}
public List getResourceDependencies( JobMeta jobMeta ) {
List references = super.getResourceDependencies( jobMeta );
if ( !Utils.isEmpty( filename ) ) {
String realFileName = jobMeta.environmentSubstitute( filename );
ResourceReference reference = new ResourceReference( this );
reference.getEntries().add( new ResourceEntry( realFileName, ResourceType.FILE ) );
references.add( reference );
}
return references;
}
@Override
public void check( List remarks, JobMeta jobMeta, VariableSpace space,
Repository repository, IMetaStore metaStore ) {
JobEntryValidatorUtils.andValidator().validate( this, "filename", remarks,
AndValidator.putValidators( JobEntryValidatorUtils.notBlankValidator() ) );
}
/**
* Since the exported job that runs this will reside in a ZIP file, we can't reference files relatively. So what this
* does is turn the name of files into absolute paths OR it simply includes the resource in the ZIP file. For now,
* we'll simply turn it into an absolute path and pray that the file is on a shared drive or something like that.
* TODO: create options to configure this behavior
*
* Exports the object to a flat-file system, adding content with filename keys to a set of definitions. The supplied
* resource naming interface allows the object to name appropriately without worrying about those parts of the
* implementation specific details.
*
* @param space
* The variable space to resolve (environment) variables with.
* @param definitions
* The map containing the filenames and content
* @param namingInterface
* The resource naming interface allows the object to be named appropriately
* @param repository
* The repository to load resources from
* @param metaStore
* the metaStore to load external metadata from
*
* @return The filename for this object. (also contained in the definitions map)
* @throws KettleException
* in case something goes wrong during the export
*/
public String exportResources( VariableSpace space, Map definitions,
ResourceNamingInterface namingInterface, Repository repository, IMetaStore metaStore ) throws KettleException {
try {
// The object that we're modifying here is a copy of the original!
// So let's change the filename from relative to absolute by grabbing the
// file object...
// In case the name of the file comes from previous steps, forget about
// this!
//
if ( !Utils.isEmpty( filename ) ) {
// From : ${FOLDER}/../foo/bar.csv
// To : /home/matt/test/files/foo/bar.csv
//
FileObject fileObject = KettleVFS.getFileObject( space.environmentSubstitute( filename ), space );
// If the file doesn't exist, forget about this effort too!
//
if ( fileObject.exists() ) {
// Convert to an absolute path...
//
filename = namingInterface.nameResource( fileObject, space, true );
return filename;
}
}
return null;
} catch ( Exception e ) {
throw new KettleException( e );
}
}
/**
* @return the className
*/
public String getClassName() {
return className;
}
/**
* @param className
* the className to set
*/
public void setClassName( String className ) {
this.className = className;
}
}