/*! ****************************************************************************** * * Pentaho Data Integration * * Copyright (C) 2002-2016 by Pentaho : http://www.pentaho.com * ******************************************************************************* * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. * ******************************************************************************/ package org.pentaho.di.job.entries.talendjobexec; import org.pentaho.di.job.entry.validator.AndValidator; import org.pentaho.di.job.entry.validator.JobEntryValidatorUtils; import java.io.File; import java.lang.reflect.Method; import java.net.URL; import java.util.ArrayList; import java.util.HashMap; import java.util.List; import java.util.Map; import org.apache.commons.vfs2.AllFileSelector; import org.apache.commons.vfs2.FileObject; import org.pentaho.di.cluster.SlaveServer; import org.pentaho.di.core.CheckResultInterface; import org.pentaho.di.core.Const; import org.pentaho.di.core.util.Utils; import org.pentaho.di.core.Result; import org.pentaho.di.core.database.DatabaseMeta; import org.pentaho.di.core.exception.KettleDatabaseException; import org.pentaho.di.core.exception.KettleException; import org.pentaho.di.core.exception.KettleXMLException; import org.pentaho.di.core.fileinput.FileInputList; import org.pentaho.di.core.plugins.KettleURLClassLoader; import org.pentaho.di.core.variables.VariableSpace; import org.pentaho.di.core.vfs.KettleVFS; import org.pentaho.di.core.xml.XMLHandler; import org.pentaho.di.i18n.BaseMessages; import org.pentaho.di.job.JobMeta; import org.pentaho.di.job.entry.JobEntryBase; import org.pentaho.di.job.entry.JobEntryInterface; import org.pentaho.di.repository.ObjectId; import org.pentaho.di.repository.Repository; import org.pentaho.di.resource.ResourceDefinition; import org.pentaho.di.resource.ResourceEntry; import org.pentaho.di.resource.ResourceEntry.ResourceType; import org.pentaho.di.resource.ResourceNamingInterface; import org.pentaho.di.resource.ResourceReference; import org.pentaho.metastore.api.IMetaStore; import org.w3c.dom.Node; /** * This executes an exported Talend Job. * * @author Matt * @since 1-04-2011 * */ public class JobEntryTalendJobExec extends JobEntryBase implements Cloneable, JobEntryInterface { private static Class<?> PKG = JobEntryTalendJobExec.class; // for i18n private static Map<String, ClassLoader> classLoaderCache = new HashMap<String, ClassLoader>(); private String filename; private String className; public JobEntryTalendJobExec( String n ) { super( n, "" ); filename = null; } public JobEntryTalendJobExec() { this( "" ); } public Object clone() { JobEntryTalendJobExec je = (JobEntryTalendJobExec) super.clone(); return je; } public String getXML() { StringBuilder retval = new StringBuilder( 50 ); retval.append( super.getXML() ); retval.append( " " ).append( XMLHandler.addTagValue( "filename", filename ) ); retval.append( " " ).append( XMLHandler.addTagValue( "class_name", className ) ); return retval.toString(); } public void loadXML( Node entrynode, List<DatabaseMeta> databases, List<SlaveServer> slaveServers, Repository rep, IMetaStore metaStore ) throws KettleXMLException { try { super.loadXML( entrynode, databases, slaveServers ); filename = XMLHandler.getTagValue( entrynode, "filename" ); className = XMLHandler.getTagValue( entrynode, "class_name" ); } catch ( KettleXMLException xe ) { throw new KettleXMLException( BaseMessages.getString( PKG, "JobEntryTalendJobExec.ERROR_0001_Cannot_Load_Job_Entry_From_Xml_Node" ), xe ); } } public void loadRep( Repository rep, IMetaStore metaStore, ObjectId id_jobentry, List<DatabaseMeta> databases, List<SlaveServer> slaveServers ) throws KettleException { try { filename = rep.getJobEntryAttributeString( id_jobentry, "filename" ); className = rep.getJobEntryAttributeString( id_jobentry, "class_name" ); } catch ( KettleException dbe ) { throw new KettleException( BaseMessages.getString( PKG, "JobEntryTalendJobExec.ERROR_0002_Cannot_Load_Job_From_Repository", id_jobentry ), dbe ); } } public void saveRep( Repository rep, IMetaStore metaStore, ObjectId id_job ) throws KettleException { try { rep.saveJobEntryAttribute( id_job, getObjectId(), "filename", filename ); rep.saveJobEntryAttribute( id_job, getObjectId(), "class_name", className ); } catch ( KettleDatabaseException dbe ) { throw new KettleException( BaseMessages.getString( PKG, "JobEntryTalendJobExec.ERROR_0003_Cannot_Save_Job_Entry", id_job ), dbe ); } } public void setFilename( String filename ) { this.filename = filename; } public String getFilename() { return filename; } public String getRealFilename() { return environmentSubstitute( getFilename() ); } public Result execute( Result previousResult, int nr ) { Result result = previousResult; result.setResult( false ); if ( filename != null ) { String realFilename = getRealFilename(); try { FileObject file = KettleVFS.getFileObject( realFilename, this ); if ( file.exists() && file.isReadable() ) { result = executeTalenJob( file, result, nr ); } else { logDetailed( BaseMessages.getString( PKG, "JobEntryTalendJobExec.File_Does_Not_Exist", realFilename ) ); } } catch ( Exception e ) { result.setNrErrors( 1 ); logError( BaseMessages.getString( PKG, "JobEntryTalendJobExec.ERROR_0004_IO_Exception", e.getMessage() ), e ); } } else { result.setNrErrors( 1 ); logError( BaseMessages.getString( PKG, "JobEntryTalendJobExec.ERROR_0005_No_Filename_Defined" ) ); } return result; } private Result executeTalenJob( FileObject file, Result result, int nr ) throws Exception { ClassLoader classLoader = null; try { classLoader = classLoaderCache.get( file.toString() ); if ( classLoader == null ) { // Find the jar files in the ZIP file... // final URL[] jarFiles = prepareJarFiles( file ); // Create a new class loader with the extracted jar files. // classLoader = new KettleURLClassLoader( jarFiles, getClass().getClassLoader() ); Runtime.getRuntime().addShutdownHook( new Thread() { @Override public void run() { try { cleanupJarFiles( jarFiles ); } catch ( Exception e ) { System.err.println( "Error cleaning up temporary Talend jar file extracts: " + Const.getStackTracker( e ) ); } } } ); classLoaderCache.put( file.toString(), classLoader ); } Class<?> clazz = classLoader.loadClass( environmentSubstitute( getClassName() ) ); Object jobObject = clazz.newInstance(); Method runJob = clazz.getMethod( "runJobInTOS", String[].class ); // TODO: consider passing something of some sort in this next method: // variables, arguments... // int returnCode = (Integer) runJob.invoke( jobObject, (Object) new String[] {} ); result.setExitStatus( returnCode ); result.setResult( true ); result.setNrErrors( 0 ); } catch ( Exception e ) { throw new KettleException( BaseMessages.getString( PKG, "JobEntryTalendJobExec.ERROR_0006_ExceptionExecutingTalenJob" ), e ); } return result; } private URL[] prepareJarFiles( FileObject zipFile ) throws Exception { // zip:file:///tmp/foo.zip FileInputList fileList = FileInputList.createFileList( this, new String[] { "zip:" + zipFile.toString(), }, new String[] { ".*\\.jar$", }, // Include mask: only jar files new String[] { ".*classpath\\.jar$", }, // Exclude mask: only jar files new String[] { "Y", }, // File required new boolean[] { true, } ); // Search sub-directories List<URL> files = new ArrayList<URL>(); // Copy the jar files in the temp folder... // for ( FileObject file : fileList.getFiles() ) { FileObject jarfilecopy = KettleVFS.createTempFile( file.getName().getBaseName(), ".jar", environmentSubstitute( "${java.io.tmpdir}" ) ); jarfilecopy.copyFrom( file, new AllFileSelector() ); files.add( jarfilecopy.getURL() ); } return files.toArray( new URL[files.size()] ); } private void cleanupJarFiles( URL[] jarFiles ) throws Exception { if ( jarFiles == null ) { return; } for ( URL jarFile : jarFiles ) { File file = new File( jarFile.toURI() ); file.delete(); } } public boolean evaluates() { return true; } public List<ResourceReference> getResourceDependencies( JobMeta jobMeta ) { List<ResourceReference> references = super.getResourceDependencies( jobMeta ); if ( !Utils.isEmpty( filename ) ) { String realFileName = jobMeta.environmentSubstitute( filename ); ResourceReference reference = new ResourceReference( this ); reference.getEntries().add( new ResourceEntry( realFileName, ResourceType.FILE ) ); references.add( reference ); } return references; } @Override public void check( List<CheckResultInterface> remarks, JobMeta jobMeta, VariableSpace space, Repository repository, IMetaStore metaStore ) { JobEntryValidatorUtils.andValidator().validate( this, "filename", remarks, AndValidator.putValidators( JobEntryValidatorUtils.notBlankValidator() ) ); } /** * Since the exported job that runs this will reside in a ZIP file, we can't reference files relatively. So what this * does is turn the name of files into absolute paths OR it simply includes the resource in the ZIP file. For now, * we'll simply turn it into an absolute path and pray that the file is on a shared drive or something like that. * TODO: create options to configure this behavior * * Exports the object to a flat-file system, adding content with filename keys to a set of definitions. The supplied * resource naming interface allows the object to name appropriately without worrying about those parts of the * implementation specific details. * * @param space * The variable space to resolve (environment) variables with. * @param definitions * The map containing the filenames and content * @param namingInterface * The resource naming interface allows the object to be named appropriately * @param repository * The repository to load resources from * @param metaStore * the metaStore to load external metadata from * * @return The filename for this object. (also contained in the definitions map) * @throws KettleException * in case something goes wrong during the export */ public String exportResources( VariableSpace space, Map<String, ResourceDefinition> definitions, ResourceNamingInterface namingInterface, Repository repository, IMetaStore metaStore ) throws KettleException { try { // The object that we're modifying here is a copy of the original! // So let's change the filename from relative to absolute by grabbing the // file object... // In case the name of the file comes from previous steps, forget about // this! // if ( !Utils.isEmpty( filename ) ) { // From : ${FOLDER}/../foo/bar.csv // To : /home/matt/test/files/foo/bar.csv // FileObject fileObject = KettleVFS.getFileObject( space.environmentSubstitute( filename ), space ); // If the file doesn't exist, forget about this effort too! // if ( fileObject.exists() ) { // Convert to an absolute path... // filename = namingInterface.nameResource( fileObject, space, true ); return filename; } } return null; } catch ( Exception e ) { throw new KettleException( e ); } } /** * @return the className */ public String getClassName() { return className; } /** * @param className * the className to set */ public void setClassName( String className ) { this.className = className; } }