blob: 66ac8f521cc9cd7c80eea61e89b1c946695a1231 [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package org.apache.uima.ducc.sampleapps;
import java.io.File;
import java.io.IOException;
import java.util.ArrayList;
import org.apache.uima.cas.CAS;
import org.apache.uima.cas.CASException;
import org.apache.uima.collection.CollectionException;
import org.apache.uima.collection.CollectionReader_ImplBase;
import org.apache.uima.ducc.Workitem;
import org.apache.uima.jcas.JCas;
import org.apache.uima.resource.ResourceConfigurationException;
import org.apache.uima.resource.ResourceInitializationException;
import org.apache.uima.util.Level;
import org.apache.uima.util.Logger;
import org.apache.uima.util.Progress;
import org.apache.uima.util.ProgressImpl;
/**
* A DUCC Job collection reader that segments a collection text files from a directory
* in the filesystem into separate Work Items for a DUCC job.
* In order to facilitate processing scale out, an optional blocksize parameter can be
* specified that will be used to break larger files into multiple Work Items.
*
* By default any previously completed output files found in the output directory are preserved,
* but a configuration parameter is available to ignore previous output.
* This CR can be configured with the following parameters:
* <ul>
* <li><code>InputDirectory</code> - path to directory containing input files</li>
* <li><code>OutputDirectory</code> - path to directory for output files</li>
* <li><code>IgnorePreviousOutput</code> (optional) - flag to ignore previous output files</li>
* <li><code>Encoding</code> (optional) - character encoding of the input files</li>
* <li><code>Language</code> (optional) - language of the input documents</li>
* <li><code>BlockSize</code> (optional) - Block size used to process input files</li>
* <li><code>SendToLast</code> (optional) - flag to route WorkItem CAS to last pipeline component. Only used for jobs with initial CM.</li>
* <li><code>SendToAll</code> (optional) - flag to route WorkItem CAS to all pipeline components. Only used for jobs with initial CM.</li>
* </ul>
*
*/
public class DuccJobTextCR extends CollectionReader_ImplBase {
/**
* Name of configuration parameter that must be set to the path of a directory containing input
* files.
*/
public static final String PARAM_INPUTDIR = "InputDirectory";
/**
* Name of configuration parameter that must be set to the path of the base directory
* where output files will be created.
*/
public static final String PARAM_OUTPUTDIR = "OutputDirectory";
/**
* Name of configuration parameter that indicates if previous output should be ignored.
*/
public static final String PARAM_IGNOREPREVIOUS = "IgnorePreviousOutput";
/**
* Name of configuration parameter that contains the character encoding used by the input files.
* If not specified, the default system encoding will be used.
*/
public static final String PARAM_ENCODING = "Encoding";
/**
* Name of optional configuration parameter that contains the language of the documents in the
* input directory. If specified this information will be added to the CAS.
*/
public static final String PARAM_LANGUAGE = "Language";
/**
* Name of configuration parameter specifying the block size used to break input files into work-items.
* Output files will correspond to the input data found in each block.
* If not specified, the entire file will be processed as a single work-item.
*/
public static final String PARAM_BLOCKSIZE = "BlockSize";
/**
* Flag to route WorkItem CAS to last pipeline component. Used to flush any output data.
* This string is ignored if the Job has a CM component.
*/
public static final String PARAM_SENDTOLAST = "SendToLast";
/**
* Flag to route WorkItem CAS to all pipeline components.
* If the Job has no CM component the WI CAS is already sent to AE and CC.
*/
public static final String PARAM_SENDTOALL = "SendToAll";
class WorkItem {
public WorkItem(String absolutePathIn, String absolutePathOut, int i, long len, long off, boolean end) {
filename=absolutePathIn;
outname=absolutePathOut;
index=i;
length=(int)len;
offset=(int)off;
last=end;
}
String filename;
String outname;
int index;
int offset;
int length;
boolean last;
}
private ArrayList<WorkItem> mWorkList;
private String mInputdirectory;
private String mOutputdirectory;
private Boolean mIgnorePrevious;
private String mEncoding;
private String mLanguage;
private int mBlocksize;
private int mCurrentIndex;
private Boolean mSendToLast;
private Boolean mSendToAll;
private int mPreviouslyDone;
private Logger logger;
/**
* @see org.apache.uima.collection.CollectionReader_ImplBase#initialize()
*/
public void initialize() throws ResourceInitializationException {
logger = getUimaContext().getLogger();
mInputdirectory = ((String) getConfigParameterValue(PARAM_INPUTDIR)).trim();
mOutputdirectory = ((String) getConfigParameterValue(PARAM_OUTPUTDIR)).trim();
mIgnorePrevious = (Boolean) getConfigParameterValue(PARAM_IGNOREPREVIOUS);
mEncoding = (String) getConfigParameterValue(PARAM_ENCODING);
mLanguage = (String) getConfigParameterValue(PARAM_LANGUAGE);
mSendToLast = (Boolean) getConfigParameterValue(PARAM_SENDTOLAST);
mSendToAll = (Boolean) getConfigParameterValue(PARAM_SENDTOALL);
if (null == mIgnorePrevious) {
mIgnorePrevious = Boolean.FALSE;
}
if (null == mSendToLast) {
mSendToLast = Boolean.FALSE;
}
if (null == mSendToAll) {
mSendToAll = Boolean.FALSE;
}
mCurrentIndex = 0;
mPreviouslyDone = 0;
// if input directory does not exist or is not a directory, throw exception
File inDirectory = new File(mInputdirectory);
if (!inDirectory.exists() || !inDirectory.isDirectory()) {
throw new ResourceInitializationException(ResourceConfigurationException.DIRECTORY_NOT_FOUND,
new Object[] { PARAM_INPUTDIR, this.getMetaData().getName(), inDirectory.getPath() });
}
// if output directory is a file throw exception
File outDirectory = new File(mOutputdirectory);
if (outDirectory.exists() && !outDirectory.isDirectory()) {
throw new ResourceInitializationException(new RuntimeException("Specified output directory "+mOutputdirectory+" is a file"));
}
mBlocksize = 0;
logger.log(Level.INFO, "Processing input files from "+mInputdirectory);
if (null != getConfigParameterValue(PARAM_BLOCKSIZE)) {
mBlocksize = (Integer) getConfigParameterValue(PARAM_BLOCKSIZE);
logger.log(Level.INFO, "Using blocksize "+ mBlocksize);
}
if (null != mIgnorePrevious && mIgnorePrevious) {
logger.log(Level.INFO, "Overwriting previous outfiles");
}
// get list of files or file-parts in the specified directory
mWorkList = new ArrayList<WorkItem>();
addFilesFromDir(inDirectory);
if (0 < mPreviouslyDone) {
logger.log(Level.INFO, "Preserving "+mPreviouslyDone+" output files in "+mOutputdirectory);
}
logger.log(Level.INFO, "Processing "+mWorkList.size()+" output files in "+mOutputdirectory);
}
/**
* This method adds files or file-chunks in the input directory,
* if the respective output file does not exist,
* or if mIgnorePrevious = true.
*
* @param dir
*/
private void addFilesFromDir(File dir) {
File[] files = dir.listFiles();
for (int i = 0; files != null && i < files.length; i++) {
if (!files[i].isDirectory()) {
String outfilename = files[i].getAbsolutePath();
outfilename = outfilename.substring(mInputdirectory.length());
outfilename = mOutputdirectory+outfilename;
if (mBlocksize == 0) {
File outFile = new File(outfilename+"_processed.zip");
if (!mIgnorePrevious && outFile.exists()) {
mPreviouslyDone++;
}
if (mIgnorePrevious || !outFile.exists()) {
mWorkList.add(new WorkItem(files[i].getAbsolutePath(),outfilename+"_processed.zip",0,files[i].length(),0,true));
logger.log(Level.FINE, "adding "+outfilename);
}
}
// use blocksize
else {
long fsize = files[i].length();
long offset=0;
int j=0;
while (fsize > 0) {
String outfilechunk = outfilename+"_"+j;
long length = (fsize < mBlocksize) ? fsize : mBlocksize;
File outFile = new File(outfilechunk+"_processed.zip");
if (!mIgnorePrevious && outFile.exists()) {
mPreviouslyDone++;
}
if (mIgnorePrevious || !outFile.exists()) {
mWorkList.add(new WorkItem(files[i].getAbsolutePath(),outfilechunk+"_processed.zip",j,length,offset,fsize==length));
logger.log(Level.FINE, "adding "+outfilechunk);
}
j++;
fsize -= length;
offset += length;
}
}
}
}
}
/**
* @see org.apache.uima.collection.CollectionReader#hasNext()
*/
public boolean hasNext() {
return mCurrentIndex < mWorkList.size();
}
/**
* @see org.apache.uima.collection.CollectionReader#getNext(org.apache.uima.cas.CAS)
*/
public void getNext(CAS aCAS) throws IOException, CollectionException {
JCas jcas;
try {
jcas = aCAS.getJCas();
Workitem wi = new Workitem(jcas);
wi.setInputspec(mWorkList.get(mCurrentIndex).filename);
wi.setOutputspec(mWorkList.get(mCurrentIndex).outname);
wi.setBlockindex(mWorkList.get(mCurrentIndex).index);
wi.setBlocksize(mBlocksize);
wi.setBytelength(mWorkList.get(mCurrentIndex).length);
if (null != mEncoding) {
wi.setEncoding(mEncoding);
}
if (null != mLanguage) {
wi.setLanguage(mLanguage);
}
wi.setSendToLast(mSendToLast);
wi.setSendToAll(mSendToAll);
wi.addToIndexes();
wi.setLastBlock(mWorkList.get(mCurrentIndex).last);
logger.log(Level.INFO, "Sending "+wi.getInputspec()+" index="+wi.getBlockindex()+" last="+wi.getLastBlock()+" length="+wi.getBytelength());
mCurrentIndex++;
jcas.setDocumentText(wi.getInputspec()+" index="+wi.getBlockindex()+" length="+wi.getBytelength());
} catch (CASException e) {
throw new CollectionException(e);
}
//create WorkItem info structure
}
/**
* @see org.apache.uima.collection.base_cpm.BaseCollectionReader#close()
*/
public void close() throws IOException {
}
/**
* @see org.apache.uima.collection.base_cpm.BaseCollectionReader#getProgress()
*/
public Progress[] getProgress() {
return new Progress[] { new ProgressImpl(mCurrentIndex, mWorkList.size(), Progress.ENTITIES) };
}
/**
* Gets the total number of documents that will be returned by this collection reader. This is not
* part of the general collection reader interface.
*
* @return the number of documents in the collection
*/
public int getNumberOfDocuments() {
return mWorkList.size();
}
}