blob: e981e25db8c32b39ff47baf7b4aa2e6c1fb0cddf [file] [log] [blame]
// Lucene version compatibility level 8.2.0
using Lucene.Net.Analysis.OpenNlp.Tools;
using Lucene.Net.Analysis.TokenAttributes;
using Lucene.Net.Util;
using System.Collections.Generic;
namespace Lucene.Net.Analysis.OpenNlp
{
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/// <summary>
/// Runs OpenNLP dictionary-based and/or MaxEnt lemmatizers.
/// <para/>
/// Both a dictionary-based lemmatizer and a MaxEnt lemmatizer are supported,
/// via the "dictionary" and "lemmatizerModel" params, respectively.
/// If both are configured, the dictionary-based lemmatizer is tried first,
/// and then the MaxEnt lemmatizer is consulted for out-of-vocabulary tokens.
/// <para/>
/// The dictionary file must be encoded as UTF-8, with one entry per line,
/// in the form <c>word[tab]lemma[tab]part-of-speech</c>
/// </summary>
public class OpenNLPLemmatizerFilter : TokenFilter
{
private readonly NLPLemmatizerOp lemmatizerOp;
private readonly ICharTermAttribute termAtt;
private readonly ITypeAttribute typeAtt;
private readonly IKeywordAttribute keywordAtt;
private readonly IFlagsAttribute flagsAtt;
private IList<AttributeSource> sentenceTokenAttrs = new List<AttributeSource>();
private IEnumerator<AttributeSource> sentenceTokenAttrsIter = null;
private bool moreTokensAvailable = true;
private string[] sentenceTokens = null; // non-keyword tokens
private string[] sentenceTokenTypes = null; // types for non-keyword tokens
private string[] lemmas = null; // lemmas for non-keyword tokens
private int lemmaNum = 0; // lemma counter
public OpenNLPLemmatizerFilter(TokenStream input, NLPLemmatizerOp lemmatizerOp)
: base(input)
{
this.lemmatizerOp = lemmatizerOp;
this.termAtt = AddAttribute<ICharTermAttribute>();
this.typeAtt = AddAttribute<ITypeAttribute>();
this.keywordAtt = AddAttribute<IKeywordAttribute>();
this.flagsAtt = AddAttribute<IFlagsAttribute>();
}
public override sealed bool IncrementToken()
{
if (!moreTokensAvailable)
{
Clear();
return false;
}
if (sentenceTokenAttrsIter == null || !sentenceTokenAttrsIter.MoveNext())
{
NextSentence();
if (sentenceTokens == null)
{ // zero non-keyword tokens
Clear();
return false;
}
lemmas = lemmatizerOp.Lemmatize(sentenceTokens, sentenceTokenTypes);
lemmaNum = 0;
sentenceTokenAttrsIter = sentenceTokenAttrs.GetEnumerator();
sentenceTokenAttrsIter.MoveNext();
}
ClearAttributes();
sentenceTokenAttrsIter.Current.CopyTo(this);
if (!keywordAtt.IsKeyword)
{
termAtt.SetEmpty().Append(lemmas[lemmaNum++]);
}
return true;
}
private void NextSentence()
{
var tokenList = new List<string>();
var typeList = new List<string>();
sentenceTokenAttrs.Clear();
bool endOfSentence = false;
while (!endOfSentence && (moreTokensAvailable = m_input.IncrementToken()))
{
if (!keywordAtt.IsKeyword)
{
tokenList.Add(termAtt.ToString());
typeList.Add(typeAtt.Type);
}
endOfSentence = 0 != (flagsAtt.Flags & OpenNLPTokenizer.EOS_FLAG_BIT);
sentenceTokenAttrs.Add(m_input.CloneAttributes());
}
sentenceTokens = tokenList.Count > 0 ? tokenList.ToArray() : null;
sentenceTokenTypes = typeList.Count > 0 ? typeList.ToArray() : null;
}
public override void Reset()
{
base.Reset();
moreTokensAvailable = true;
Clear();
}
private void Clear()
{
sentenceTokenAttrs.Clear();
sentenceTokenAttrsIter?.Dispose();
sentenceTokenAttrsIter = null;
sentenceTokens = null;
sentenceTokenTypes = null;
lemmas = null;
lemmaNum = 0;
}
}
}