| package org.apache.lucene.analysis.ngram; |
| |
| /* |
| * Licensed to the Apache Software Foundation (ASF) under one or more |
| * contributor license agreements. See the NOTICE file distributed with |
| * this work for additional information regarding copyright ownership. |
| * The ASF licenses this file to You under the Apache License, Version 2.0 |
| * (the "License"); you may not use this file except in compliance with |
| * the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| import java.io.IOException; |
| import java.io.Reader; |
| import java.io.StringReader; |
| import java.util.Random; |
| |
| import org.apache.lucene.analysis.Analyzer; |
| import org.apache.lucene.analysis.BaseTokenStreamTestCase; |
| import org.apache.lucene.analysis.MockTokenizer; |
| import org.apache.lucene.analysis.TokenFilter; |
| import org.apache.lucene.analysis.TokenStream; |
| import org.apache.lucene.analysis.Tokenizer; |
| import org.apache.lucene.analysis.core.KeywordTokenizer; |
| import org.apache.lucene.analysis.core.LetterTokenizer; |
| import org.apache.lucene.analysis.core.WhitespaceTokenizer; |
| import org.apache.lucene.analysis.shingle.ShingleFilter; |
| import org.apache.lucene.analysis.tokenattributes.PositionIncrementAttribute; |
| import org.apache.lucene.util.Version; |
| |
| /** |
| * Tests {@link EdgeNGramTokenFilter} for correctness. |
| */ |
| public class EdgeNGramTokenFilterTest extends BaseTokenStreamTestCase { |
| private TokenStream input; |
| |
| @Override |
| public void setUp() throws Exception { |
| super.setUp(); |
| input = new MockTokenizer(new StringReader("abcde"), MockTokenizer.WHITESPACE, false); |
| } |
| |
| public void testInvalidInput() throws Exception { |
| boolean gotException = false; |
| try { |
| new EdgeNGramTokenFilter(TEST_VERSION_CURRENT, input, 0, 0); |
| } catch (IllegalArgumentException e) { |
| gotException = true; |
| } |
| assertTrue(gotException); |
| } |
| |
| public void testInvalidInput2() throws Exception { |
| boolean gotException = false; |
| try { |
| new EdgeNGramTokenFilter(TEST_VERSION_CURRENT, input, 2, 1); |
| } catch (IllegalArgumentException e) { |
| gotException = true; |
| } |
| assertTrue(gotException); |
| } |
| |
| public void testInvalidInput3() throws Exception { |
| boolean gotException = false; |
| try { |
| new EdgeNGramTokenFilter(TEST_VERSION_CURRENT, input, -1, 2); |
| } catch (IllegalArgumentException e) { |
| gotException = true; |
| } |
| assertTrue(gotException); |
| } |
| |
| public void testFrontUnigram() throws Exception { |
| EdgeNGramTokenFilter tokenizer = new EdgeNGramTokenFilter(TEST_VERSION_CURRENT, input, 1, 1); |
| assertTokenStreamContents(tokenizer, new String[]{"a"}, new int[]{0}, new int[]{5}); |
| } |
| |
| public void testOversizedNgrams() throws Exception { |
| EdgeNGramTokenFilter tokenizer = new EdgeNGramTokenFilter(TEST_VERSION_CURRENT, input, 6, 6); |
| assertTokenStreamContents(tokenizer, new String[0], new int[0], new int[0]); |
| } |
| |
| public void testFrontRangeOfNgrams() throws Exception { |
| EdgeNGramTokenFilter tokenizer = new EdgeNGramTokenFilter(TEST_VERSION_CURRENT, input, 1, 3); |
| assertTokenStreamContents(tokenizer, new String[]{"a","ab","abc"}, new int[]{0,0,0}, new int[]{5,5,5}); |
| } |
| |
| public void testFilterPositions() throws Exception { |
| TokenStream ts = new MockTokenizer(new StringReader("abcde vwxyz"), MockTokenizer.WHITESPACE, false); |
| EdgeNGramTokenFilter tokenizer = new EdgeNGramTokenFilter(TEST_VERSION_CURRENT, ts, 1, 3); |
| assertTokenStreamContents(tokenizer, |
| new String[]{"a","ab","abc","v","vw","vwx"}, |
| new int[]{0,0,0,6,6,6}, |
| new int[]{5,5,5,11,11,11}, |
| null, |
| new int[]{1,0,0,1,0,0}, |
| null, |
| null, |
| false); |
| } |
| |
| private static class PositionFilter extends TokenFilter { |
| |
| private final PositionIncrementAttribute posIncrAtt = addAttribute(PositionIncrementAttribute.class); |
| private boolean started; |
| |
| PositionFilter(final TokenStream input) { |
| super(input); |
| } |
| |
| @Override |
| public final boolean incrementToken() throws IOException { |
| if (input.incrementToken()) { |
| if (started) { |
| posIncrAtt.setPositionIncrement(0); |
| } else { |
| started = true; |
| } |
| return true; |
| } else { |
| return false; |
| } |
| } |
| |
| @Override |
| public void reset() throws IOException { |
| super.reset(); |
| started = false; |
| } |
| } |
| |
| public void testFirstTokenPositionIncrement() throws Exception { |
| TokenStream ts = new MockTokenizer(new StringReader("a abc"), MockTokenizer.WHITESPACE, false); |
| ts = new PositionFilter(ts); // All but first token will get 0 position increment |
| EdgeNGramTokenFilter filter = new EdgeNGramTokenFilter(TEST_VERSION_CURRENT, ts, 2, 3); |
| // The first token "a" will not be output, since it's smaller than the mingram size of 2. |
| // The second token on input to EdgeNGramTokenFilter will have position increment of 0, |
| // which should be increased to 1, since this is the first output token in the stream. |
| assertTokenStreamContents(filter, |
| new String[] { "ab", "abc" }, |
| new int[] { 2, 2 }, |
| new int[] { 5, 5 }, |
| new int[] { 1, 0 } |
| ); |
| } |
| |
| public void testSmallTokenInStream() throws Exception { |
| input = new MockTokenizer(new StringReader("abc de fgh"), MockTokenizer.WHITESPACE, false); |
| EdgeNGramTokenFilter tokenizer = new EdgeNGramTokenFilter(TEST_VERSION_CURRENT, input, 3, 3); |
| assertTokenStreamContents(tokenizer, new String[]{"abc","fgh"}, new int[]{0,7}, new int[]{3,10}); |
| } |
| |
| public void testReset() throws Exception { |
| WhitespaceTokenizer tokenizer = new WhitespaceTokenizer(TEST_VERSION_CURRENT, new StringReader("abcde")); |
| EdgeNGramTokenFilter filter = new EdgeNGramTokenFilter(TEST_VERSION_CURRENT, tokenizer, 1, 3); |
| assertTokenStreamContents(filter, new String[]{"a","ab","abc"}, new int[]{0,0,0}, new int[]{5,5,5}); |
| tokenizer.setReader(new StringReader("abcde")); |
| assertTokenStreamContents(filter, new String[]{"a","ab","abc"}, new int[]{0,0,0}, new int[]{5,5,5}); |
| } |
| |
| /** blast some random strings through the analyzer */ |
| public void testRandomStrings() throws Exception { |
| Analyzer a = new Analyzer() { |
| @Override |
| protected TokenStreamComponents createComponents(String fieldName, Reader reader) { |
| Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); |
| return new TokenStreamComponents(tokenizer, |
| new EdgeNGramTokenFilter(TEST_VERSION_CURRENT, tokenizer, 2, 4)); |
| } |
| }; |
| checkRandomData(random(), a, 1000*RANDOM_MULTIPLIER); |
| } |
| |
| public void testEmptyTerm() throws Exception { |
| Random random = random(); |
| Analyzer a = new Analyzer() { |
| @Override |
| protected TokenStreamComponents createComponents(String fieldName, Reader reader) { |
| Tokenizer tokenizer = new KeywordTokenizer(reader); |
| return new TokenStreamComponents(tokenizer, |
| new EdgeNGramTokenFilter(TEST_VERSION_CURRENT, tokenizer, 2, 15)); |
| } |
| }; |
| checkAnalysisConsistency(random, a, random.nextBoolean(), ""); |
| } |
| |
| public void testGraphs() throws IOException { |
| TokenStream tk = new LetterTokenizer(Version.LUCENE_44, new StringReader("abc d efgh ij klmno p q")); |
| tk = new ShingleFilter(tk); |
| tk = new EdgeNGramTokenFilter(Version.LUCENE_44, tk, 7, 10); |
| tk.reset(); |
| assertTokenStreamContents(tk, |
| new String[] { "efgh ij", "ij klmn", "ij klmno", "klmno p" }, |
| new int[] { 6,11,11,14 }, |
| new int[] { 13,19,19,21 }, |
| new int[] { 3,1,0,1 }, |
| new int[] { 2,2,2,2 }, |
| 23 |
| ); |
| } |
| } |