| /* |
| * Licensed to the Apache Software Foundation (ASF) under one or more |
| * contributor license agreements. See the NOTICE file distributed with |
| * this work for additional information regarding copyright ownership. |
| * The ASF licenses this file to You under the Apache License, Version 2.0 |
| * (the "License"); you may not use this file except in compliance with |
| * the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| package org.apache.lucene.analysis.miscellaneous; |
| |
| |
| import java.io.IOException; |
| |
| import org.apache.lucene.analysis.Analyzer; |
| import org.apache.lucene.analysis.BaseTokenStreamTestCase; |
| import org.apache.lucene.analysis.MockAnalyzer; |
| import org.apache.lucene.document.Document; |
| import org.apache.lucene.document.Field; |
| import org.apache.lucene.index.DirectoryReader; |
| import org.apache.lucene.index.IndexReader; |
| import org.apache.lucene.index.IndexWriter; |
| import org.apache.lucene.index.IndexWriterConfig; |
| import org.apache.lucene.index.Term; |
| import org.apache.lucene.store.Directory; |
| import org.apache.lucene.util.TestUtil; |
| |
| public class TestLimitTokenCountAnalyzer extends BaseTokenStreamTestCase { |
| |
| public void testLimitTokenCountAnalyzer() throws IOException { |
| for (boolean consumeAll : new boolean[] { true, false }) { |
| MockAnalyzer mock = new MockAnalyzer(random()); |
| |
| // if we are consuming all tokens, we can use the checks, |
| // otherwise we can't |
| mock.setEnableChecks(consumeAll); |
| Analyzer a = new LimitTokenCountAnalyzer(mock, 2, consumeAll); |
| |
| // dont use assertAnalyzesTo here, as the end offset is not the end of the string (unless consumeAll is true, in which case it's correct)! |
| assertTokenStreamContents(a.tokenStream("dummy", "1 2 3 4 5"), new String[] { "1", "2" }, new int[] { 0, 3 }, new int[] { 1, 4 }, consumeAll ? 16 : null); |
| assertTokenStreamContents(a.tokenStream("dummy", "1 2 3 4 5"), new String[] { "1", "2" }, new int[] { 0, 2 }, new int[] { 1, 3 }, consumeAll ? 9 : null); |
| |
| // less than the limit, ensure we behave correctly |
| assertTokenStreamContents(a.tokenStream("dummy", "1 "), new String[] { "1" }, new int[] { 0 }, new int[] { 1 }, consumeAll ? 3 : null); |
| |
| // equal to limit |
| assertTokenStreamContents(a.tokenStream("dummy", "1 2 "), new String[] { "1", "2" }, new int[] { 0, 3 }, new int[] { 1, 4 }, consumeAll ? 6 : null); |
| a.close(); |
| } |
| } |
| |
| public void testLimitTokenCountIndexWriter() throws IOException { |
| |
| for (boolean consumeAll : new boolean[] { true, false }) { |
| Directory dir = newDirectory(); |
| int limit = TestUtil.nextInt(random(), 50, 101000); |
| MockAnalyzer mock = new MockAnalyzer(random()); |
| |
| // if we are consuming all tokens, we can use the checks, |
| // otherwise we can't |
| mock.setEnableChecks(consumeAll); |
| Analyzer a = new LimitTokenCountAnalyzer(mock, limit, consumeAll); |
| |
| IndexWriter writer = new IndexWriter(dir, new IndexWriterConfig(a)); |
| |
| Document doc = new Document(); |
| StringBuilder b = new StringBuilder(); |
| for(int i=1;i<limit;i++) |
| b.append(" a"); |
| b.append(" x"); |
| b.append(" z"); |
| doc.add(newTextField("field", b.toString(), Field.Store.NO)); |
| writer.addDocument(doc); |
| writer.close(); |
| |
| IndexReader reader = DirectoryReader.open(dir); |
| Term t = new Term("field", "x"); |
| assertEquals(1, reader.docFreq(t)); |
| t = new Term("field", "z"); |
| assertEquals(0, reader.docFreq(t)); |
| reader.close(); |
| dir.close(); |
| a.close(); |
| } |
| } |
| |
| } |