All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.uima.ducc.sampleapps.DuccJobTextCR Maven / Gradle / Ivy

There is a newer version: 3.0.0
Show newest version
/*
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 * 
 *   http://www.apache.org/licenses/LICENSE-2.0
 * 
 * Unless required by applicable law or agreed to in writing,
 * software distributed under the License is distributed on an
 * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
 * KIND, either express or implied.  See the License for the
 * specific language governing permissions and limitations
 * under the License.
 */

package org.apache.uima.ducc.sampleapps;

import java.io.File;
import java.io.IOException;
import java.util.ArrayList;

import org.apache.uima.cas.CAS;
import org.apache.uima.cas.CASException;
import org.apache.uima.collection.CollectionException;
import org.apache.uima.collection.CollectionReader_ImplBase;
import org.apache.uima.ducc.Workitem;
import org.apache.uima.jcas.JCas;
import org.apache.uima.resource.ResourceConfigurationException;
import org.apache.uima.resource.ResourceInitializationException;
import org.apache.uima.util.Level;
import org.apache.uima.util.Logger;
import org.apache.uima.util.Progress;
import org.apache.uima.util.ProgressImpl;

/**
 * A DUCC Job collection reader that segments a collection text files from a directory 
 * in the filesystem into separate Work Items for a DUCC job.
 * In order to facilitate processing scale out, an optional blocksize parameter can be 
 * specified that will be used to break larger files into multiple Work Items.
 * 
 * By default any previously completed output files found in the output directory are preserved,
 * but a configuration parameter is available to ignore previous output. 

 * This CR can be configured with the following parameters:
 * 
    *
  • InputDirectory - path to directory containing input files
  • *
  • OutputDirectory - path to directory for output files
  • *
  • IgnorePreviousOutput (optional) - flag to ignore previous output files
  • *
  • Encoding (optional) - character encoding of the input files
  • *
  • Language (optional) - language of the input documents
  • *
  • BlockSize (optional) - Block size used to process input files
  • *
  • SendToLast (optional) - flag to route WorkItem CAS to last pipeline component. Only used for jobs with initial CM.
  • *
  • SendToAll (optional) - flag to route WorkItem CAS to all pipeline components. Only used for jobs with initial CM.
  • *
* */ public class DuccJobTextCR extends CollectionReader_ImplBase { /** * Name of configuration parameter that must be set to the path of a directory containing input * files. */ public static final String PARAM_INPUTDIR = "InputDirectory"; /** * Name of configuration parameter that must be set to the path of the base directory * where output files will be created. */ public static final String PARAM_OUTPUTDIR = "OutputDirectory"; /** * Name of configuration parameter that indicates if previous output should be ignored. */ public static final String PARAM_IGNOREPREVIOUS = "IgnorePreviousOutput"; /** * Name of configuration parameter that contains the character encoding used by the input files. * If not specified, the default system encoding will be used. */ public static final String PARAM_ENCODING = "Encoding"; /** * Name of optional configuration parameter that contains the language of the documents in the * input directory. If specified this information will be added to the CAS. */ public static final String PARAM_LANGUAGE = "Language"; /** * Name of configuration parameter specifying the block size used to break input files into work-items. * Output files will correspond to the input data found in each block. * If not specified, the entire file will be processed as a single work-item. */ public static final String PARAM_BLOCKSIZE = "BlockSize"; /** * Flag to route WorkItem CAS to last pipeline component. Used to flush any output data. * This string is ignored if the Job has a CM component. */ public static final String PARAM_SENDTOLAST = "SendToLast"; /** * Flag to route WorkItem CAS to all pipeline components. * If the Job has no CM component the WI CAS is already sent to AE and CC. */ public static final String PARAM_SENDTOALL = "SendToAll"; class WorkItem { public WorkItem(String absolutePathIn, String absolutePathOut, int i, long len, long off, boolean end) { filename=absolutePathIn; outname=absolutePathOut; index=i; length=(int)len; offset=(int)off; last=end; } String filename; String outname; int index; int offset; int length; boolean last; } private ArrayList mWorkList; private String mInputdirectory; private String mOutputdirectory; private Boolean mIgnorePrevious; private String mEncoding; private String mLanguage; private int mBlocksize; private int mCurrentIndex; private Boolean mSendToLast; private Boolean mSendToAll; private int mPreviouslyDone; private Logger logger; /** * @see org.apache.uima.collection.CollectionReader_ImplBase#initialize() */ public void initialize() throws ResourceInitializationException { logger = getUimaContext().getLogger(); mInputdirectory = ((String) getConfigParameterValue(PARAM_INPUTDIR)).trim(); mOutputdirectory = ((String) getConfigParameterValue(PARAM_OUTPUTDIR)).trim(); mIgnorePrevious = (Boolean) getConfigParameterValue(PARAM_IGNOREPREVIOUS); mEncoding = (String) getConfigParameterValue(PARAM_ENCODING); mLanguage = (String) getConfigParameterValue(PARAM_LANGUAGE); mSendToLast = (Boolean) getConfigParameterValue(PARAM_SENDTOLAST); mSendToAll = (Boolean) getConfigParameterValue(PARAM_SENDTOALL); if (null == mIgnorePrevious) { mIgnorePrevious = Boolean.FALSE; } if (null == mSendToLast) { mSendToLast = Boolean.FALSE; } if (null == mSendToAll) { mSendToAll = Boolean.FALSE; } mCurrentIndex = 0; mPreviouslyDone = 0; // if input directory does not exist or is not a directory, throw exception File inDirectory = new File(mInputdirectory); if (!inDirectory.exists() || !inDirectory.isDirectory()) { throw new ResourceInitializationException(ResourceConfigurationException.DIRECTORY_NOT_FOUND, new Object[] { PARAM_INPUTDIR, this.getMetaData().getName(), inDirectory.getPath() }); } // if output directory is a file throw exception File outDirectory = new File(mOutputdirectory); if (outDirectory.exists() && !outDirectory.isDirectory()) { throw new ResourceInitializationException(new RuntimeException("Specified output directory "+mOutputdirectory+" is a file")); } mBlocksize = 0; logger.log(Level.INFO, "Processing input files from "+mInputdirectory); if (null != getConfigParameterValue(PARAM_BLOCKSIZE)) { mBlocksize = (Integer) getConfigParameterValue(PARAM_BLOCKSIZE); logger.log(Level.INFO, "Using blocksize "+ mBlocksize); } if (null != mIgnorePrevious && mIgnorePrevious) { logger.log(Level.INFO, "Overwriting previous outfiles"); } // get list of files or file-parts in the specified directory mWorkList = new ArrayList(); addFilesFromDir(inDirectory); if (0 < mPreviouslyDone) { logger.log(Level.INFO, "Preserving "+mPreviouslyDone+" output files in "+mOutputdirectory); } logger.log(Level.INFO, "Processing "+mWorkList.size()+" output files in "+mOutputdirectory); } /** * This method adds files or file-chunks in the input directory, * if the respective output file does not exist, * or if mIgnorePrevious = true. * * @param dir */ private void addFilesFromDir(File dir) { File[] files = dir.listFiles(); for (int i = 0; files != null && i < files.length; i++) { if (!files[i].isDirectory()) { String outfilename = files[i].getAbsolutePath(); outfilename = outfilename.substring(mInputdirectory.length()); outfilename = mOutputdirectory+outfilename; if (mBlocksize == 0) { File outFile = new File(outfilename+"_processed.zip"); if (!mIgnorePrevious && outFile.exists()) { mPreviouslyDone++; } if (mIgnorePrevious || !outFile.exists()) { mWorkList.add(new WorkItem(files[i].getAbsolutePath(),outfilename+"_processed.zip",0,files[i].length(),0,true)); logger.log(Level.FINE, "adding "+outfilename); } } // use blocksize else { long fsize = files[i].length(); long offset=0; int j=0; while (fsize > 0) { String outfilechunk = outfilename+"_"+j; long length = (fsize < mBlocksize) ? fsize : mBlocksize; File outFile = new File(outfilechunk+"_processed.zip"); if (!mIgnorePrevious && outFile.exists()) { mPreviouslyDone++; } if (mIgnorePrevious || !outFile.exists()) { mWorkList.add(new WorkItem(files[i].getAbsolutePath(),outfilechunk+"_processed.zip",j,length,offset,fsize==length)); logger.log(Level.FINE, "adding "+outfilechunk); } j++; fsize -= length; offset += length; } } } } } /** * @see org.apache.uima.collection.CollectionReader#hasNext() */ public boolean hasNext() { return mCurrentIndex < mWorkList.size(); } /** * @see org.apache.uima.collection.CollectionReader#getNext(org.apache.uima.cas.CAS) */ public void getNext(CAS aCAS) throws IOException, CollectionException { JCas jcas; try { jcas = aCAS.getJCas(); Workitem wi = new Workitem(jcas); wi.setInputspec(mWorkList.get(mCurrentIndex).filename); wi.setOutputspec(mWorkList.get(mCurrentIndex).outname); wi.setBlockindex(mWorkList.get(mCurrentIndex).index); wi.setBlocksize(mBlocksize); wi.setBytelength(mWorkList.get(mCurrentIndex).length); if (null != mEncoding) { wi.setEncoding(mEncoding); } if (null != mLanguage) { wi.setLanguage(mLanguage); } wi.setSendToLast(mSendToLast); wi.setSendToAll(mSendToAll); wi.addToIndexes(); wi.setLastBlock(mWorkList.get(mCurrentIndex).last); logger.log(Level.INFO, "Sending "+wi.getInputspec()+" index="+wi.getBlockindex()+" last="+wi.getLastBlock()+" length="+wi.getBytelength()); mCurrentIndex++; jcas.setDocumentText(wi.getInputspec()+" index="+wi.getBlockindex()+" length="+wi.getBytelength()); } catch (CASException e) { throw new CollectionException(e); } //create WorkItem info structure } /** * @see org.apache.uima.collection.base_cpm.BaseCollectionReader#close() */ public void close() throws IOException { } /** * @see org.apache.uima.collection.base_cpm.BaseCollectionReader#getProgress() */ public Progress[] getProgress() { return new Progress[] { new ProgressImpl(mCurrentIndex, mWorkList.size(), Progress.ENTITIES) }; } /** * Gets the total number of documents that will be returned by this collection reader. This is not * part of the general collection reader interface. * * @return the number of documents in the collection */ public int getNumberOfDocuments() { return mWorkList.size(); } }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy