| /* |
| * |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, |
| * software distributed under the License is distributed on an |
| * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| * KIND, either express or implied. See the License for the |
| * specific language governing permissions and limitations |
| * under the License. |
| * |
| */ |
| |
| using Lucene.Net.Analysis; |
| using Lucene.Net.Analysis.TokenAttributes; |
| using Lucene.Net.Documents; |
| using Lucene.Net.Index; |
| using Lucene.Net.QueryParsers.Classic; |
| using Lucene.Net.QueryParsers.Flexible.Standard; |
| using Lucene.Net.Search; |
| using Lucene.Net.Store; |
| using Lucene.Net.Support; |
| using Lucene.Net.Util; |
| using Lucene.Net.Util.Automaton; |
| using NUnit.Framework; |
| using System; |
| using System.Globalization; |
| using Console = Lucene.Net.Util.SystemConsole; |
| |
| namespace Lucene.Net.QueryParsers.Util |
| { |
| /// <summary> |
| /// In .NET the abstact members were moved to AbstractQueryParserTestBase |
| /// because the Visual Studio test runner does not find or run tests in |
| /// abstract classes. |
| /// </summary> |
| [TestFixture] |
| public abstract class QueryParserTestBase : LuceneTestCase |
| { |
| public static Analyzer qpAnalyzer; |
| |
| [OneTimeSetUp] |
| public override void BeforeClass() |
| { |
| base.BeforeClass(); |
| qpAnalyzer = new QPTestAnalyzer(); |
| } |
| |
| [OneTimeTearDown] |
| public override void AfterClass() |
| { |
| qpAnalyzer = null; |
| base.AfterClass(); |
| } |
| |
| public sealed class QPTestFilter : TokenFilter |
| { |
| ICharTermAttribute termAtt; |
| IOffsetAttribute offsetAtt; |
| |
| /** |
| * Filter which discards the token 'stop' and which expands the |
| * token 'phrase' into 'phrase1 phrase2' |
| */ |
| public QPTestFilter(TokenStream @in) |
| : base(@in) |
| { |
| termAtt = AddAttribute<ICharTermAttribute>(); |
| offsetAtt = AddAttribute<IOffsetAttribute>(); |
| } |
| |
| bool inPhrase = false; |
| int savedStart = 0, savedEnd = 0; |
| |
| public override sealed bool IncrementToken() |
| { |
| if (inPhrase) |
| { |
| inPhrase = false; |
| ClearAttributes(); |
| termAtt.Append("phrase2"); |
| offsetAtt.SetOffset(savedStart, savedEnd); |
| return true; |
| } |
| else |
| while (m_input.IncrementToken()) |
| { |
| if (termAtt.toString().Equals("phrase", StringComparison.Ordinal)) |
| { |
| inPhrase = true; |
| savedStart = offsetAtt.StartOffset; |
| savedEnd = offsetAtt.EndOffset; |
| termAtt.SetEmpty().Append("phrase1"); |
| offsetAtt.SetOffset(savedStart, savedEnd); |
| return true; |
| } |
| else if (!termAtt.toString().Equals("stop", StringComparison.Ordinal)) |
| return true; |
| } |
| return false; |
| } |
| } |
| |
| public sealed class QPTestAnalyzer : Analyzer |
| { |
| /// <summary> |
| /// Filters MockTokenizer with StopFilter. |
| /// </summary> |
| protected internal override TokenStreamComponents CreateComponents(string fieldName, System.IO.TextReader reader) |
| { |
| Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true); |
| return new TokenStreamComponents(tokenizer, new QPTestFilter(tokenizer)); |
| } |
| } |
| |
| private int originalMaxClauses; |
| |
| private string defaultField = "field"; |
| public string DefaultField |
| { |
| get => defaultField; |
| set => defaultField = value; |
| } |
| |
| public override void SetUp() |
| { |
| base.SetUp(); |
| originalMaxClauses = BooleanQuery.MaxClauseCount; |
| } |
| |
| public abstract ICommonQueryParserConfiguration GetParserConfig(Analyzer a); |
| |
| public abstract void SetDefaultOperatorOR(ICommonQueryParserConfiguration cqpC); |
| |
| public abstract void SetDefaultOperatorAND(ICommonQueryParserConfiguration cqpC); |
| |
| public abstract void SetAnalyzeRangeTerms(ICommonQueryParserConfiguration cqpC, bool value); |
| |
| public abstract void SetAutoGeneratePhraseQueries(ICommonQueryParserConfiguration cqpC, bool value); |
| |
| public abstract void SetDateResolution(ICommonQueryParserConfiguration cqpC, string field, DateTools.Resolution value); |
| |
| public abstract Query GetQuery(string query, ICommonQueryParserConfiguration cqpC); |
| |
| public abstract Query GetQuery(string query, Analyzer a); |
| |
| public abstract bool IsQueryParserException(Exception exception); |
| |
| public Query GetQuery(string query) |
| { |
| return GetQuery(query, (Analyzer)null); |
| } |
| |
| public void AssertQueryEquals(string query, Analyzer a, string result) |
| { |
| Query q = GetQuery(query, a); |
| string s = q.ToString("field"); |
| if (!s.Equals(result, StringComparison.Ordinal)) |
| { |
| fail("Query /" + query + "/ yielded /" + s |
| + "/, expecting /" + result + "/"); |
| } |
| } |
| |
| public void AssertQueryEquals(ICommonQueryParserConfiguration cqpC, string field, string query, string result) |
| { |
| Query q = GetQuery(query, cqpC); |
| string s = q.ToString(field); |
| if (!s.Equals(result, StringComparison.Ordinal)) |
| { |
| fail("Query /" + query + "/ yielded /" + s |
| + "/, expecting /" + result + "/"); |
| } |
| } |
| |
| public void AssertEscapedQueryEquals(string query, Analyzer a, string result) |
| { |
| string escapedQuery = QueryParserBase.Escape(query); |
| if (!escapedQuery.Equals(result, StringComparison.Ordinal)) |
| { |
| fail("Query /" + query + "/ yielded /" + escapedQuery |
| + "/, expecting /" + result + "/"); |
| } |
| } |
| |
| public void AssertWildcardQueryEquals(string query, bool lowercase, string result, bool allowLeadingWildcard) |
| { |
| ICommonQueryParserConfiguration cqpC = GetParserConfig(null); |
| cqpC.LowercaseExpandedTerms = lowercase; |
| cqpC.AllowLeadingWildcard = allowLeadingWildcard; |
| Query q = GetQuery(query, cqpC); |
| string s = q.ToString("field"); |
| if (!s.Equals(result, StringComparison.Ordinal)) |
| { |
| fail("WildcardQuery /" + query + "/ yielded /" + s |
| + "/, expecting /" + result + "/"); |
| } |
| } |
| |
| public void AssertWildcardQueryEquals(string query, bool lowercase, string result) |
| { |
| AssertWildcardQueryEquals(query, lowercase, result, false); |
| } |
| |
| public void AssertWildcardQueryEquals(string query, string result) |
| { |
| Query q = GetQuery(query); |
| string s = q.ToString("field"); |
| if (!s.Equals(result, StringComparison.Ordinal)) |
| { |
| fail("WildcardQuery /" + query + "/ yielded /" + s + "/, expecting /" |
| + result + "/"); |
| } |
| } |
| |
| public Query GetQueryDOA(string query, Analyzer a) |
| { |
| if (a == null) |
| a = new MockAnalyzer(Random, MockTokenizer.SIMPLE, true); |
| ICommonQueryParserConfiguration qp = GetParserConfig(a); |
| SetDefaultOperatorAND(qp); |
| return GetQuery(query, qp); |
| } |
| |
| public void AssertQueryEqualsDOA(string query, Analyzer a, string result) |
| { |
| Query q = GetQueryDOA(query, a); |
| string s = q.ToString("field"); |
| if (!s.Equals(result, StringComparison.Ordinal)) |
| { |
| fail("Query /" + query + "/ yielded /" + s |
| + "/, expecting /" + result + "/"); |
| } |
| } |
| |
| [Test] |
| public virtual void TestCJK() |
| { |
| // Test Ideographic Space - As wide as a CJK character cell (fullwidth) |
| // used google to translate the word "term" to japanese -> 用語 |
| AssertQueryEquals("term\u3000term\u3000term", null, "term\u0020term\u0020term"); |
| AssertQueryEquals("用語\u3000用語\u3000用語", null, "用語\u0020用語\u0020用語"); |
| } |
| |
| protected class SimpleCJKTokenizer : Tokenizer |
| { |
| private ICharTermAttribute termAtt; |
| |
| public SimpleCJKTokenizer(System.IO.TextReader input) |
| : base(input) |
| { |
| termAtt = AddAttribute<ICharTermAttribute>(); |
| } |
| |
| public override sealed bool IncrementToken() |
| { |
| int ch = m_input.Read(); |
| if (ch < 0) |
| return false; |
| ClearAttributes(); |
| termAtt.SetEmpty().Append((char)ch); |
| return true; |
| } |
| } |
| |
| private class SimpleCJKAnalyzer : Analyzer |
| { |
| protected internal override TokenStreamComponents CreateComponents(string fieldName, System.IO.TextReader reader) |
| { |
| return new TokenStreamComponents(new SimpleCJKTokenizer(reader)); |
| } |
| } |
| |
| [Test] |
| public virtual void TestCJKTerm() |
| { |
| // individual CJK chars as terms |
| SimpleCJKAnalyzer analyzer = new SimpleCJKAnalyzer(); |
| |
| BooleanQuery expected = new BooleanQuery(); |
| expected.Add(new TermQuery(new Term("field", "ä¸")), Occur.SHOULD); |
| expected.Add(new TermQuery(new Term("field", "国")), Occur.SHOULD); |
| |
| assertEquals(expected, GetQuery("ä¸å›½", analyzer)); |
| } |
| |
| [Test] |
| public virtual void TestCJKBoostedTerm() |
| { |
| // individual CJK chars as terms |
| SimpleCJKAnalyzer analyzer = new SimpleCJKAnalyzer(); |
| |
| BooleanQuery expected = new BooleanQuery(); |
| expected.Boost = (0.5f); |
| expected.Add(new TermQuery(new Term("field", "ä¸")), Occur.SHOULD); |
| expected.Add(new TermQuery(new Term("field", "国")), Occur.SHOULD); |
| |
| assertEquals(expected, GetQuery("ä¸å›½^0.5", analyzer)); |
| } |
| |
| [Test] |
| public virtual void TestCJKPhrase() |
| { |
| // individual CJK chars as terms |
| SimpleCJKAnalyzer analyzer = new SimpleCJKAnalyzer(); |
| |
| PhraseQuery expected = new PhraseQuery(); |
| expected.Add(new Term("field", "ä¸")); |
| expected.Add(new Term("field", "国")); |
| |
| assertEquals(expected, GetQuery("\"ä¸å›½\"", analyzer)); |
| } |
| |
| [Test] |
| public virtual void TestCJKBoostedPhrase() |
| { |
| // individual CJK chars as terms |
| SimpleCJKAnalyzer analyzer = new SimpleCJKAnalyzer(); |
| |
| PhraseQuery expected = new PhraseQuery(); |
| expected.Boost = (0.5f); |
| expected.Add(new Term("field", "ä¸")); |
| expected.Add(new Term("field", "国")); |
| |
| assertEquals(expected, GetQuery("\"ä¸å›½\"^0.5", analyzer)); |
| } |
| |
| [Test] |
| public virtual void TestCJKSloppyPhrase() |
| { |
| // individual CJK chars as terms |
| SimpleCJKAnalyzer analyzer = new SimpleCJKAnalyzer(); |
| |
| PhraseQuery expected = new PhraseQuery(); |
| expected.Slop = (3); |
| expected.Add(new Term("field", "ä¸")); |
| expected.Add(new Term("field", "国")); |
| |
| assertEquals(expected, GetQuery("\"ä¸å›½\"~3", analyzer)); |
| } |
| |
| [Test] |
| public virtual void TestAutoGeneratePhraseQueriesOn() |
| { |
| // individual CJK chars as terms |
| SimpleCJKAnalyzer analyzer = new SimpleCJKAnalyzer(); |
| |
| PhraseQuery expected = new PhraseQuery(); |
| expected.Add(new Term("field", "ä¸")); |
| expected.Add(new Term("field", "国")); |
| ICommonQueryParserConfiguration qp = GetParserConfig(analyzer); |
| SetAutoGeneratePhraseQueries(qp, true); |
| assertEquals(expected, GetQuery("ä¸å›½", qp)); |
| } |
| |
| [Test] |
| public virtual void TestSimple() |
| { |
| AssertQueryEquals("term term term", null, "term term term"); |
| AssertQueryEquals("türm term term", new MockAnalyzer(Random), "türm term term"); |
| AssertQueryEquals("ümlaut", new MockAnalyzer(Random), "ümlaut"); |
| |
| // FIXME: enhance MockAnalyzer to be able to support this |
| // it must no longer extend CharTokenizer |
| //AssertQueryEquals("\"\"", new KeywordAnalyzer(), ""); |
| //AssertQueryEquals("foo:\"\"", new KeywordAnalyzer(), "foo:"); |
| |
| AssertQueryEquals("a AND b", null, "+a +b"); |
| AssertQueryEquals("(a AND b)", null, "+a +b"); |
| AssertQueryEquals("c OR (a AND b)", null, "c (+a +b)"); |
| AssertQueryEquals("a AND NOT b", null, "+a -b"); |
| AssertQueryEquals("a AND -b", null, "+a -b"); |
| AssertQueryEquals("a AND !b", null, "+a -b"); |
| AssertQueryEquals("a && b", null, "+a +b"); |
| // AssertQueryEquals("a && ! b", null, "+a -b"); |
| |
| AssertQueryEquals("a OR b", null, "a b"); |
| AssertQueryEquals("a || b", null, "a b"); |
| AssertQueryEquals("a OR !b", null, "a -b"); |
| // AssertQueryEquals("a OR ! b", null, "a -b"); |
| AssertQueryEquals("a OR -b", null, "a -b"); |
| |
| AssertQueryEquals("+term -term term", null, "+term -term term"); |
| AssertQueryEquals("foo:term AND field:anotherTerm", null, |
| "+foo:term +anotherterm"); |
| AssertQueryEquals("term AND \"phrase phrase\"", null, |
| "+term +\"phrase phrase\""); |
| AssertQueryEquals("\"hello there\"", null, "\"hello there\""); |
| assertTrue(GetQuery("a AND b") is BooleanQuery); |
| assertTrue(GetQuery("hello") is TermQuery); |
| assertTrue(GetQuery("\"hello there\"") is PhraseQuery); |
| |
| AssertQueryEquals("germ term^2.0", null, "germ term^2.0"); |
| AssertQueryEquals("(term)^2.0", null, "term^2.0"); |
| AssertQueryEquals("(germ term)^2.0", null, "(germ term)^2.0"); |
| AssertQueryEquals("term^2.0", null, "term^2.0"); |
| AssertQueryEquals("term^2", null, "term^2.0"); |
| AssertQueryEquals("\"germ term\"^2.0", null, "\"germ term\"^2.0"); |
| AssertQueryEquals("\"term germ\"^2", null, "\"term germ\"^2.0"); |
| |
| AssertQueryEquals("(foo OR bar) AND (baz OR boo)", null, |
| "+(foo bar) +(baz boo)"); |
| AssertQueryEquals("((a OR b) AND NOT c) OR d", null, |
| "(+(a b) -c) d"); |
| AssertQueryEquals("+(apple \"steve jobs\") -(foo bar baz)", null, |
| "+(apple \"steve jobs\") -(foo bar baz)"); |
| AssertQueryEquals("+title:(dog OR cat) -author:\"bob dole\"", null, |
| "+(title:dog title:cat) -author:\"bob dole\""); |
| |
| } |
| |
| public abstract void TestDefaultOperator(); |
| |
| private class OperatorVsWhitespaceAnalyzer : Analyzer |
| { |
| protected internal override TokenStreamComponents CreateComponents(string fieldName, System.IO.TextReader reader) |
| { |
| return new TokenStreamComponents(new MockTokenizer(reader, MockTokenizer.WHITESPACE, false)); |
| } |
| } |
| |
| [Test] |
| public virtual void TestOperatorVsWhitespace() |
| { //LUCENE-2566 |
| // +,-,! should be directly adjacent to operand (i.e. not separated by whitespace) to be treated as an operator |
| Analyzer a = new OperatorVsWhitespaceAnalyzer(); |
| AssertQueryEquals("a - b", a, "a - b"); |
| AssertQueryEquals("a + b", a, "a + b"); |
| AssertQueryEquals("a ! b", a, "a ! b"); |
| } |
| |
| [Test] |
| public virtual void TestPunct() |
| { |
| Analyzer a = new MockAnalyzer(Random, MockTokenizer.WHITESPACE, false); |
| AssertQueryEquals("a&b", a, "a&b"); |
| AssertQueryEquals("a&&b", a, "a&&b"); |
| AssertQueryEquals(".NET", a, ".NET"); |
| } |
| |
| [Test] |
| public virtual void TestSlop() |
| { |
| AssertQueryEquals("\"term germ\"~2", null, "\"term germ\"~2"); |
| AssertQueryEquals("\"term germ\"~2 flork", null, "\"term germ\"~2 flork"); |
| AssertQueryEquals("\"term\"~2", null, "term"); |
| AssertQueryEquals("\" \"~2 germ", null, "germ"); |
| AssertQueryEquals("\"term germ\"~2^2", null, "\"term germ\"~2^2.0"); |
| } |
| |
| [Test] |
| public virtual void TestNumber() |
| { |
| // The numbers go away because SimpleAnalzyer ignores them |
| AssertQueryEquals("3", null, ""); |
| AssertQueryEquals("term 1.0 1 2", null, "term"); |
| AssertQueryEquals("term term1 term2", null, "term term term"); |
| |
| Analyzer a = new MockAnalyzer(Random, MockTokenizer.WHITESPACE, true); |
| AssertQueryEquals("3", a, "3"); |
| AssertQueryEquals("term 1.0 1 2", a, "term 1.0 1 2"); |
| AssertQueryEquals("term term1 term2", a, "term term1 term2"); |
| } |
| |
| [Test] |
| public virtual void TestWildcard() |
| { |
| AssertQueryEquals("term*", null, "term*"); |
| AssertQueryEquals("term*^2", null, "term*^2.0"); |
| AssertQueryEquals("term~", null, "term~2"); |
| AssertQueryEquals("term~1", null, "term~1"); |
| AssertQueryEquals("term~0.7", null, "term~1"); |
| AssertQueryEquals("term~^3", null, "term~2^3.0"); |
| AssertQueryEquals("term^3~", null, "term~2^3.0"); |
| AssertQueryEquals("term*germ", null, "term*germ"); |
| AssertQueryEquals("term*germ^3", null, "term*germ^3.0"); |
| |
| assertTrue(GetQuery("term*") is PrefixQuery); |
| assertTrue(GetQuery("term*^2") is PrefixQuery); |
| assertTrue(GetQuery("term~") is FuzzyQuery); |
| assertTrue(GetQuery("term~0.7") is FuzzyQuery); |
| FuzzyQuery fq = (FuzzyQuery)GetQuery("term~0.7"); |
| assertEquals(1, fq.MaxEdits); |
| assertEquals(FuzzyQuery.DefaultPrefixLength, fq.PrefixLength); |
| fq = (FuzzyQuery)GetQuery("term~"); |
| assertEquals(2, fq.MaxEdits); |
| assertEquals(FuzzyQuery.DefaultPrefixLength, fq.PrefixLength); |
| |
| AssertParseException("term~1.1"); // value > 1, throws exception |
| |
| assertTrue(GetQuery("term*germ") is WildcardQuery); |
| |
| /* Tests to see that wild card terms are (or are not) properly |
| * lower-cased with propery parser configuration |
| */ |
| // First prefix queries: |
| // by default, convert to lowercase: |
| AssertWildcardQueryEquals("Term*", true, "term*"); |
| // explicitly set lowercase: |
| AssertWildcardQueryEquals("term*", true, "term*"); |
| AssertWildcardQueryEquals("Term*", true, "term*"); |
| AssertWildcardQueryEquals("TERM*", true, "term*"); |
| // explicitly disable lowercase conversion: |
| AssertWildcardQueryEquals("term*", false, "term*"); |
| AssertWildcardQueryEquals("Term*", false, "Term*"); |
| AssertWildcardQueryEquals("TERM*", false, "TERM*"); |
| // Then 'full' wildcard queries: |
| // by default, convert to lowercase: |
| AssertWildcardQueryEquals("Te?m", "te?m"); |
| // explicitly set lowercase: |
| AssertWildcardQueryEquals("te?m", true, "te?m"); |
| AssertWildcardQueryEquals("Te?m", true, "te?m"); |
| AssertWildcardQueryEquals("TE?M", true, "te?m"); |
| AssertWildcardQueryEquals("Te?m*gerM", true, "te?m*germ"); |
| // explicitly disable lowercase conversion: |
| AssertWildcardQueryEquals("te?m", false, "te?m"); |
| AssertWildcardQueryEquals("Te?m", false, "Te?m"); |
| AssertWildcardQueryEquals("TE?M", false, "TE?M"); |
| AssertWildcardQueryEquals("Te?m*gerM", false, "Te?m*gerM"); |
| // Fuzzy queries: |
| AssertWildcardQueryEquals("Term~", "term~2"); |
| AssertWildcardQueryEquals("Term~", true, "term~2"); |
| AssertWildcardQueryEquals("Term~", false, "Term~2"); |
| // Range queries: |
| AssertWildcardQueryEquals("[A TO C]", "[a TO c]"); |
| AssertWildcardQueryEquals("[A TO C]", true, "[a TO c]"); |
| AssertWildcardQueryEquals("[A TO C]", false, "[A TO C]"); |
| // Test suffix queries: first disallow |
| try |
| { |
| AssertWildcardQueryEquals("*Term", true, "*term"); |
| } |
| catch (Exception pe) |
| { |
| // expected exception |
| if (!IsQueryParserException(pe)) |
| { |
| fail(); |
| } |
| } |
| try |
| { |
| AssertWildcardQueryEquals("?Term", true, "?term"); |
| fail(); |
| } |
| catch (Exception pe) |
| { |
| // expected exception |
| if (!IsQueryParserException(pe)) |
| { |
| fail(); |
| } |
| } |
| // Test suffix queries: then allow |
| AssertWildcardQueryEquals("*Term", true, "*term", true); |
| AssertWildcardQueryEquals("?Term", true, "?term", true); |
| } |
| |
| [Test] |
| public virtual void TestLeadingWildcardType() |
| { |
| ICommonQueryParserConfiguration cqpC = GetParserConfig(null); |
| cqpC.AllowLeadingWildcard = (true); |
| assertEquals(typeof(WildcardQuery), GetQuery("t*erm*", cqpC).GetType()); |
| assertEquals(typeof(WildcardQuery), GetQuery("?term*", cqpC).GetType()); |
| assertEquals(typeof(WildcardQuery), GetQuery("*term*", cqpC).GetType()); |
| } |
| |
| [Test] |
| public virtual void TestQPA() |
| { |
| AssertQueryEquals("term term^3.0 term", qpAnalyzer, "term term^3.0 term"); |
| AssertQueryEquals("term stop^3.0 term", qpAnalyzer, "term term"); |
| |
| AssertQueryEquals("term term term", qpAnalyzer, "term term term"); |
| AssertQueryEquals("term +stop term", qpAnalyzer, "term term"); |
| AssertQueryEquals("term -stop term", qpAnalyzer, "term term"); |
| |
| AssertQueryEquals("drop AND (stop) AND roll", qpAnalyzer, "+drop +roll"); |
| AssertQueryEquals("term +(stop) term", qpAnalyzer, "term term"); |
| AssertQueryEquals("term -(stop) term", qpAnalyzer, "term term"); |
| |
| AssertQueryEquals("drop AND stop AND roll", qpAnalyzer, "+drop +roll"); |
| AssertQueryEquals("term phrase term", qpAnalyzer, |
| "term (phrase1 phrase2) term"); |
| AssertQueryEquals("term AND NOT phrase term", qpAnalyzer, |
| "+term -(phrase1 phrase2) term"); |
| AssertQueryEquals("stop^3", qpAnalyzer, ""); |
| AssertQueryEquals("stop", qpAnalyzer, ""); |
| AssertQueryEquals("(stop)^3", qpAnalyzer, ""); |
| AssertQueryEquals("((stop))^3", qpAnalyzer, ""); |
| AssertQueryEquals("(stop^3)", qpAnalyzer, ""); |
| AssertQueryEquals("((stop)^3)", qpAnalyzer, ""); |
| AssertQueryEquals("(stop)", qpAnalyzer, ""); |
| AssertQueryEquals("((stop))", qpAnalyzer, ""); |
| assertTrue(GetQuery("term term term", qpAnalyzer) is BooleanQuery); |
| assertTrue(GetQuery("term +stop", qpAnalyzer) is TermQuery); |
| |
| ICommonQueryParserConfiguration cqpc = GetParserConfig(qpAnalyzer); |
| SetDefaultOperatorAND(cqpc); |
| AssertQueryEquals(cqpc, "field", "term phrase term", |
| "+term +(+phrase1 +phrase2) +term"); |
| AssertQueryEquals(cqpc, "field", "phrase", |
| "+phrase1 +phrase2"); |
| } |
| |
| [Test] |
| public virtual void TestRange() |
| { |
| AssertQueryEquals("[ a TO z]", null, "[a TO z]"); |
| AssertQueryEquals("[ a TO z}", null, "[a TO z}"); |
| AssertQueryEquals("{ a TO z]", null, "{a TO z]"); |
| |
| assertEquals(MultiTermQuery.CONSTANT_SCORE_AUTO_REWRITE_DEFAULT, ((TermRangeQuery)GetQuery("[ a TO z]")).MultiTermRewriteMethod); |
| |
| ICommonQueryParserConfiguration qp = GetParserConfig(new MockAnalyzer(Random, MockTokenizer.SIMPLE, true)); |
| |
| qp.MultiTermRewriteMethod=(MultiTermQuery.SCORING_BOOLEAN_QUERY_REWRITE); |
| assertEquals(MultiTermQuery.SCORING_BOOLEAN_QUERY_REWRITE, ((TermRangeQuery)GetQuery("[ a TO z]", qp)).MultiTermRewriteMethod); |
| |
| // test open ranges |
| AssertQueryEquals("[ a TO * ]", null, "[a TO *]"); |
| AssertQueryEquals("[ * TO z ]", null, "[* TO z]"); |
| AssertQueryEquals("[ * TO * ]", null, "[* TO *]"); |
| |
| // mixing exclude and include bounds |
| AssertQueryEquals("{ a TO z ]", null, "{a TO z]"); |
| AssertQueryEquals("[ a TO z }", null, "[a TO z}"); |
| AssertQueryEquals("{ a TO * ]", null, "{a TO *]"); |
| AssertQueryEquals("[ * TO z }", null, "[* TO z}"); |
| |
| AssertQueryEquals("[ a TO z ]", null, "[a TO z]"); |
| AssertQueryEquals("{ a TO z}", null, "{a TO z}"); |
| AssertQueryEquals("{ a TO z }", null, "{a TO z}"); |
| AssertQueryEquals("{ a TO z }^2.0", null, "{a TO z}^2.0"); |
| AssertQueryEquals("[ a TO z] OR bar", null, "[a TO z] bar"); |
| AssertQueryEquals("[ a TO z] AND bar", null, "+[a TO z] +bar"); |
| AssertQueryEquals("( bar blar { a TO z}) ", null, "bar blar {a TO z}"); |
| AssertQueryEquals("gack ( bar blar { a TO z}) ", null, "gack (bar blar {a TO z})"); |
| |
| AssertQueryEquals("[* TO Z]", null, "[* TO z]"); |
| AssertQueryEquals("[A TO *]", null, "[a TO *]"); |
| AssertQueryEquals("[* TO *]", null, "[* TO *]"); |
| } |
| |
| [Test] |
| public virtual void TestRangeWithPhrase() |
| { |
| AssertQueryEquals("[\\* TO \"*\"]", null, "[\\* TO \\*]"); |
| AssertQueryEquals("[\"*\" TO *]", null, "[\\* TO *]"); |
| } |
| |
| private string EscapeDateString(string s) |
| { |
| if (s.IndexOf(' ') > -1) |
| { |
| return "\"" + s + "\""; |
| } |
| else |
| { |
| return s; |
| } |
| } |
| |
| /// <summary>for testing DateTools support</summary> |
| private string GetDate(string s, DateTools.Resolution resolution) |
| { |
| // we use the default Locale since LuceneTestCase randomizes it |
| DateTime d = DateTime.ParseExact(s, CultureInfo.CurrentCulture.DateTimeFormat.ShortDatePattern, CultureInfo.CurrentCulture); |
| return GetDate(d, resolution); |
| } |
| |
| /// <summary>for testing DateTools support</summary> |
| private string GetDate(DateTime d, DateTools.Resolution resolution) |
| { |
| return DateTools.DateToString(d, resolution); |
| } |
| |
| private string GetLocalizedDate(int year, int month, int day) |
| { |
| // we use the default Locale/TZ since LuceneTestCase randomizes it |
| DateTime date = new GregorianCalendar().ToDateTime(year, month, day, 23, 59, 59, 999); |
| date = TimeZoneInfo.ConvertTime(date, TimeZoneInfo.Local); |
| return date.ToString("d"); |
| |
| //// we use the default Locale/TZ since LuceneTestCase randomizes it |
| //DateFormat df = DateFormat.getDateInstance(DateFormat.SHORT, Locale.getDefault()); |
| //Calendar calendar = new GregorianCalendar(TimeZone.getDefault(), Locale.getDefault()); |
| //calendar.clear(); |
| //calendar.set(year, month, day); |
| //calendar.set(Calendar.HOUR_OF_DAY, 23); |
| //calendar.set(Calendar.MINUTE, 59); |
| //calendar.set(Calendar.SECOND, 59); |
| //calendar.set(Calendar.MILLISECOND, 999); |
| //return df.format(calendar.getTime()); |
| } |
| |
| [Test] |
| public virtual void TestDateRange() |
| { |
| string startDate = GetLocalizedDate(2002, 1, 1); |
| string endDate = GetLocalizedDate(2002, 1, 4); |
| |
| // we use the default Locale/TZ since LuceneTestCase randomizes it |
| //Calendar endDateExpected = new GregorianCalendar(TimeZone.getDefault(), Locale.getDefault()); |
| //endDateExpected.clear(); |
| //endDateExpected.set(2002, 1, 4, 23, 59, 59); |
| //endDateExpected.set(Calendar.MILLISECOND, 999); |
| |
| // we use the default Locale/TZ since LuceneTestCase randomizes it |
| DateTime endDateExpected = new GregorianCalendar().ToDateTime(2002, 1, 4, 23, 59, 59, 999); |
| string defaultField = "default"; |
| string monthField = "month"; |
| string hourField = "hour"; |
| Analyzer a = new MockAnalyzer(Random, MockTokenizer.SIMPLE, true); |
| ICommonQueryParserConfiguration qp = GetParserConfig(a); |
| |
| // set a field specific date resolution |
| SetDateResolution(qp, monthField, DateTools.Resolution.MONTH); |
| |
| // set default date resolution to MILLISECOND |
| qp.SetDateResolution(DateTools.Resolution.MILLISECOND); |
| |
| // set second field specific date resolution |
| SetDateResolution(qp, hourField, DateTools.Resolution.HOUR); |
| |
| // for this field no field specific date resolution has been set, |
| // so verify if the default resolution is used |
| AssertDateRangeQueryEquals(qp, defaultField, startDate, endDate, |
| endDateExpected /*.getTime()*/, DateTools.Resolution.MILLISECOND); |
| |
| // verify if field specific date resolutions are used for these two fields |
| AssertDateRangeQueryEquals(qp, monthField, startDate, endDate, |
| endDateExpected /*.getTime()*/, DateTools.Resolution.MONTH); |
| |
| AssertDateRangeQueryEquals(qp, hourField, startDate, endDate, |
| endDateExpected /*.getTime()*/, DateTools.Resolution.HOUR); |
| } |
| |
| public void AssertDateRangeQueryEquals(ICommonQueryParserConfiguration cqpC, string field, string startDate, string endDate, |
| DateTime endDateInclusive, DateTools.Resolution resolution) |
| { |
| AssertQueryEquals(cqpC, field, field + ":[" + EscapeDateString(startDate) + " TO " + EscapeDateString(endDate) + "]", |
| "[" + GetDate(startDate, resolution) + " TO " + GetDate(endDateInclusive, resolution) + "]"); |
| AssertQueryEquals(cqpC, field, field + ":{" + EscapeDateString(startDate) + " TO " + EscapeDateString(endDate) + "}", |
| "{" + GetDate(startDate, resolution) + " TO " + GetDate(endDate, resolution) + "}"); |
| } |
| |
| [Test] |
| public virtual void TestEscaped() |
| { |
| Analyzer a = new MockAnalyzer(Random, MockTokenizer.WHITESPACE, false); |
| |
| /*AssertQueryEquals("\\[brackets", a, "\\[brackets"); |
| AssertQueryEquals("\\[brackets", null, "brackets"); |
| AssertQueryEquals("\\\\", a, "\\\\"); |
| AssertQueryEquals("\\+blah", a, "\\+blah"); |
| AssertQueryEquals("\\(blah", a, "\\(blah"); |
| |
| AssertQueryEquals("\\-blah", a, "\\-blah"); |
| AssertQueryEquals("\\!blah", a, "\\!blah"); |
| AssertQueryEquals("\\{blah", a, "\\{blah"); |
| AssertQueryEquals("\\}blah", a, "\\}blah"); |
| AssertQueryEquals("\\:blah", a, "\\:blah"); |
| AssertQueryEquals("\\^blah", a, "\\^blah"); |
| AssertQueryEquals("\\[blah", a, "\\[blah"); |
| AssertQueryEquals("\\]blah", a, "\\]blah"); |
| AssertQueryEquals("\\\"blah", a, "\\\"blah"); |
| AssertQueryEquals("\\(blah", a, "\\(blah"); |
| AssertQueryEquals("\\)blah", a, "\\)blah"); |
| AssertQueryEquals("\\~blah", a, "\\~blah"); |
| AssertQueryEquals("\\*blah", a, "\\*blah"); |
| AssertQueryEquals("\\?blah", a, "\\?blah"); |
| //AssertQueryEquals("foo \\&\\& bar", a, "foo \\&\\& bar"); |
| //AssertQueryEquals("foo \\|| bar", a, "foo \\|| bar"); |
| //AssertQueryEquals("foo \\AND bar", a, "foo \\AND bar");*/ |
| |
| AssertQueryEquals("\\a", a, "a"); |
| |
| AssertQueryEquals("a\\-b:c", a, "a-b:c"); |
| AssertQueryEquals("a\\+b:c", a, "a+b:c"); |
| AssertQueryEquals("a\\:b:c", a, "a:b:c"); |
| AssertQueryEquals("a\\\\b:c", a, "a\\b:c"); |
| |
| AssertQueryEquals("a:b\\-c", a, "a:b-c"); |
| AssertQueryEquals("a:b\\+c", a, "a:b+c"); |
| AssertQueryEquals("a:b\\:c", a, "a:b:c"); |
| AssertQueryEquals("a:b\\\\c", a, "a:b\\c"); |
| |
| AssertQueryEquals("a:b\\-c*", a, "a:b-c*"); |
| AssertQueryEquals("a:b\\+c*", a, "a:b+c*"); |
| AssertQueryEquals("a:b\\:c*", a, "a:b:c*"); |
| |
| AssertQueryEquals("a:b\\\\c*", a, "a:b\\c*"); |
| |
| AssertQueryEquals("a:b\\-c~", a, "a:b-c~2"); |
| AssertQueryEquals("a:b\\+c~", a, "a:b+c~2"); |
| AssertQueryEquals("a:b\\:c~", a, "a:b:c~2"); |
| AssertQueryEquals("a:b\\\\c~", a, "a:b\\c~2"); |
| |
| AssertQueryEquals("[ a\\- TO a\\+ ]", null, "[a- TO a+]"); |
| AssertQueryEquals("[ a\\: TO a\\~ ]", null, "[a: TO a~]"); |
| AssertQueryEquals("[ a\\\\ TO a\\* ]", null, "[a\\ TO a*]"); |
| |
| AssertQueryEquals("[\"c\\:\\\\temp\\\\\\~foo0.txt\" TO \"c\\:\\\\temp\\\\\\~foo9.txt\"]", a, |
| "[c:\\temp\\~foo0.txt TO c:\\temp\\~foo9.txt]"); |
| |
| AssertQueryEquals("a\\\\\\+b", a, "a\\+b"); |
| |
| AssertQueryEquals("a \\\"b c\\\" d", a, "a \"b c\" d"); |
| AssertQueryEquals("\"a \\\"b c\\\" d\"", a, "\"a \"b c\" d\""); |
| AssertQueryEquals("\"a \\+b c d\"", a, "\"a +b c d\""); |
| |
| AssertQueryEquals("c\\:\\\\temp\\\\\\~foo.txt", a, "c:\\temp\\~foo.txt"); |
| |
| AssertParseException("XY\\"); // there must be a character after the escape char |
| |
| // test unicode escaping |
| AssertQueryEquals("a\\u0062c", a, "abc"); |
| AssertQueryEquals("XY\\u005a", a, "XYZ"); |
| AssertQueryEquals("XY\\u005A", a, "XYZ"); |
| AssertQueryEquals("\"a \\\\\\u0028\\u0062\\\" c\"", a, "\"a \\(b\" c\""); |
| |
| AssertParseException("XY\\u005G"); // test non-hex character in escaped unicode sequence |
| AssertParseException("XY\\u005"); // test incomplete escaped unicode sequence |
| |
| // Tests bug LUCENE-800 |
| AssertQueryEquals("(item:\\\\ item:ABCD\\\\)", a, "item:\\ item:ABCD\\"); |
| AssertParseException("(item:\\\\ item:ABCD\\\\))"); // unmatched closing paranthesis |
| AssertQueryEquals("\\*", a, "*"); |
| AssertQueryEquals("\\\\", a, "\\"); // escaped backslash |
| |
| AssertParseException("\\"); // a backslash must always be escaped |
| |
| // LUCENE-1189 |
| AssertQueryEquals("(\"a\\\\\") or (\"b\")", a, "a\\ or b"); |
| } |
| |
| [Test] |
| public virtual void TestEscapedVsQuestionMarkAsWildcard() |
| { |
| Analyzer a = new MockAnalyzer(Random, MockTokenizer.WHITESPACE, false); |
| AssertQueryEquals("a:b\\-?c", a, "a:b\\-?c"); |
| AssertQueryEquals("a:b\\+?c", a, "a:b\\+?c"); |
| AssertQueryEquals("a:b\\:?c", a, "a:b\\:?c"); |
| |
| AssertQueryEquals("a:b\\\\?c", a, "a:b\\\\?c"); |
| } |
| |
| [Test] |
| public virtual void TestQueryStringEscaping() |
| { |
| Analyzer a = new MockAnalyzer(Random, MockTokenizer.WHITESPACE, false); |
| |
| AssertEscapedQueryEquals("a-b:c", a, "a\\-b\\:c"); |
| AssertEscapedQueryEquals("a+b:c", a, "a\\+b\\:c"); |
| AssertEscapedQueryEquals("a:b:c", a, "a\\:b\\:c"); |
| AssertEscapedQueryEquals("a\\b:c", a, "a\\\\b\\:c"); |
| |
| AssertEscapedQueryEquals("a:b-c", a, "a\\:b\\-c"); |
| AssertEscapedQueryEquals("a:b+c", a, "a\\:b\\+c"); |
| AssertEscapedQueryEquals("a:b:c", a, "a\\:b\\:c"); |
| AssertEscapedQueryEquals("a:b\\c", a, "a\\:b\\\\c"); |
| |
| AssertEscapedQueryEquals("a:b-c*", a, "a\\:b\\-c\\*"); |
| AssertEscapedQueryEquals("a:b+c*", a, "a\\:b\\+c\\*"); |
| AssertEscapedQueryEquals("a:b:c*", a, "a\\:b\\:c\\*"); |
| |
| AssertEscapedQueryEquals("a:b\\\\c*", a, "a\\:b\\\\\\\\c\\*"); |
| |
| AssertEscapedQueryEquals("a:b-?c", a, "a\\:b\\-\\?c"); |
| AssertEscapedQueryEquals("a:b+?c", a, "a\\:b\\+\\?c"); |
| AssertEscapedQueryEquals("a:b:?c", a, "a\\:b\\:\\?c"); |
| |
| AssertEscapedQueryEquals("a:b?c", a, "a\\:b\\?c"); |
| |
| AssertEscapedQueryEquals("a:b-c~", a, "a\\:b\\-c\\~"); |
| AssertEscapedQueryEquals("a:b+c~", a, "a\\:b\\+c\\~"); |
| AssertEscapedQueryEquals("a:b:c~", a, "a\\:b\\:c\\~"); |
| AssertEscapedQueryEquals("a:b\\c~", a, "a\\:b\\\\c\\~"); |
| |
| AssertEscapedQueryEquals("[ a - TO a+ ]", null, "\\[ a \\- TO a\\+ \\]"); |
| AssertEscapedQueryEquals("[ a : TO a~ ]", null, "\\[ a \\: TO a\\~ \\]"); |
| AssertEscapedQueryEquals("[ a\\ TO a* ]", null, "\\[ a\\\\ TO a\\* \\]"); |
| |
| // LUCENE-881 |
| AssertEscapedQueryEquals("|| abc ||", a, "\\|\\| abc \\|\\|"); |
| AssertEscapedQueryEquals("&& abc &&", a, "\\&\\& abc \\&\\&"); |
| } |
| |
| [Test] |
| public virtual void TestTabNewlineCarriageReturn() |
| { |
| AssertQueryEqualsDOA("+weltbank +worlbank", null, |
| "+weltbank +worlbank"); |
| |
| AssertQueryEqualsDOA("+weltbank\n+worlbank", null, |
| "+weltbank +worlbank"); |
| AssertQueryEqualsDOA("weltbank \n+worlbank", null, |
| "+weltbank +worlbank"); |
| AssertQueryEqualsDOA("weltbank \n +worlbank", null, |
| "+weltbank +worlbank"); |
| |
| AssertQueryEqualsDOA("+weltbank\r+worlbank", null, |
| "+weltbank +worlbank"); |
| AssertQueryEqualsDOA("weltbank \r+worlbank", null, |
| "+weltbank +worlbank"); |
| AssertQueryEqualsDOA("weltbank \r +worlbank", null, |
| "+weltbank +worlbank"); |
| |
| AssertQueryEqualsDOA("+weltbank\r\n+worlbank", null, |
| "+weltbank +worlbank"); |
| AssertQueryEqualsDOA("weltbank \r\n+worlbank", null, |
| "+weltbank +worlbank"); |
| AssertQueryEqualsDOA("weltbank \r\n +worlbank", null, |
| "+weltbank +worlbank"); |
| AssertQueryEqualsDOA("weltbank \r \n +worlbank", null, |
| "+weltbank +worlbank"); |
| |
| AssertQueryEqualsDOA("+weltbank\t+worlbank", null, |
| "+weltbank +worlbank"); |
| AssertQueryEqualsDOA("weltbank \t+worlbank", null, |
| "+weltbank +worlbank"); |
| AssertQueryEqualsDOA("weltbank \t +worlbank", null, |
| "+weltbank +worlbank"); |
| } |
| |
| [Test] |
| public virtual void TestSimpleDAO() |
| { |
| AssertQueryEqualsDOA("term term term", null, "+term +term +term"); |
| AssertQueryEqualsDOA("term +term term", null, "+term +term +term"); |
| AssertQueryEqualsDOA("term term +term", null, "+term +term +term"); |
| AssertQueryEqualsDOA("term +term +term", null, "+term +term +term"); |
| AssertQueryEqualsDOA("-term term term", null, "-term +term +term"); |
| } |
| |
| [Test] |
| public virtual void TestBoost() |
| { |
| CharacterRunAutomaton stopWords = new CharacterRunAutomaton(BasicAutomata.MakeString("on")); |
| Analyzer oneStopAnalyzer = new MockAnalyzer(Random, MockTokenizer.SIMPLE, true, stopWords); |
| ICommonQueryParserConfiguration qp = GetParserConfig(oneStopAnalyzer); |
| Query q = GetQuery("on^1.0", qp); |
| assertNotNull(q); |
| q = GetQuery("\"hello\"^2.0", qp); |
| assertNotNull(q); |
| assertEquals(q.Boost, (float)2.0, (float)0.5); |
| q = GetQuery("hello^2.0", qp); |
| assertNotNull(q); |
| assertEquals(q.Boost, (float)2.0, (float)0.5); |
| q = GetQuery("\"on\"^1.0", qp); |
| assertNotNull(q); |
| |
| Analyzer a2 = new MockAnalyzer(Random, MockTokenizer.SIMPLE, true, MockTokenFilter.ENGLISH_STOPSET); |
| ICommonQueryParserConfiguration qp2 = GetParserConfig(a2); |
| q = GetQuery("the^3", qp2); |
| // "the" is a stop word so the result is an empty query: |
| assertNotNull(q); |
| assertEquals("", q.toString()); |
| assertEquals(1.0f, q.Boost, 0.01f); |
| } |
| |
| public void AssertParseException(string queryString) |
| { |
| try |
| { |
| GetQuery(queryString); |
| } |
| catch (Exception expected) |
| { |
| if (IsQueryParserException(expected)) |
| { |
| return; |
| } |
| } |
| fail("ParseException expected, not thrown"); |
| } |
| |
| public void AssertParseException(string queryString, Analyzer a) |
| { |
| try |
| { |
| GetQuery(queryString, a); |
| } |
| catch (Exception expected) |
| { |
| if (IsQueryParserException(expected)) |
| { |
| return; |
| } |
| } |
| fail("ParseException expected, not thrown"); |
| } |
| |
| [Test] |
| public virtual void TestException() |
| { |
| AssertParseException("\"some phrase"); |
| AssertParseException("(foo bar"); |
| AssertParseException("foo bar))"); |
| AssertParseException("field:term:with:colon some more terms"); |
| AssertParseException("(sub query)^5.0^2.0 plus more"); |
| AssertParseException("secret AND illegal) AND access:confidential"); |
| } |
| |
| [Test] |
| public virtual void TestBooleanQuery() |
| { |
| BooleanQuery.MaxClauseCount = (2); |
| Analyzer purWhitespaceAnalyzer = new MockAnalyzer(Random, MockTokenizer.WHITESPACE, false); |
| AssertParseException("one two three", purWhitespaceAnalyzer); |
| } |
| |
| [Test] |
| public virtual void TestPrecedence() |
| { |
| ICommonQueryParserConfiguration qp = GetParserConfig(new MockAnalyzer(Random, MockTokenizer.WHITESPACE, false)); |
| Query query1 = GetQuery("A AND B OR C AND D", qp); |
| Query query2 = GetQuery("+A +B +C +D", qp); |
| assertEquals(query1, query2); |
| } |
| |
| // LUCENETODO: convert this from DateField to DateUtil |
| // public void TestLocalDateFormat() throws IOException, ParseException { |
| // Directory ramDir = newDirectory(); |
| // IndexWriter iw = new IndexWriter(ramDir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random, MockTokenizer.WHITESPACE, false))); |
| // addDateDoc("a", 2005, 12, 2, 10, 15, 33, iw); |
| // addDateDoc("b", 2005, 12, 4, 22, 15, 00, iw); |
| // iw.close(); |
| // IndexSearcher is = new IndexSearcher(ramDir, true); |
| // assertHits(1, "[12/1/2005 TO 12/3/2005]", is); |
| // assertHits(2, "[12/1/2005 TO 12/4/2005]", is); |
| // assertHits(1, "[12/3/2005 TO 12/4/2005]", is); |
| // assertHits(1, "{12/1/2005 TO 12/3/2005}", is); |
| // assertHits(1, "{12/1/2005 TO 12/4/2005}", is); |
| // assertHits(0, "{12/3/2005 TO 12/4/2005}", is); |
| // is.close(); |
| // ramDir.close(); |
| // } |
| // |
| // private void addDateDoc(String content, int year, int month, |
| // int day, int hour, int minute, int second, IndexWriter iw) throws IOException { |
| // Document d = new Document(); |
| // d.add(newField("f", content, Field.Store.YES, Field.Index.ANALYZED)); |
| // Calendar cal = Calendar.getInstance(Locale.ENGLISH); |
| // cal.set(year, month - 1, day, hour, minute, second); |
| // d.add(newField("date", DateField.dateToString(cal.getTime()), Field.Store.YES, Field.Index.NOT_ANALYZED)); |
| // iw.addDocument(d); |
| // } |
| |
| public abstract void TestStarParsing(); |
| |
| [Test] |
| public virtual void TestEscapedWildcard() |
| { |
| ICommonQueryParserConfiguration qp = GetParserConfig(new MockAnalyzer(Random, MockTokenizer.WHITESPACE, false)); |
| WildcardQuery q = new WildcardQuery(new Term("field", "foo\\?ba?r")); |
| assertEquals(q, GetQuery("foo\\?ba?r", qp)); |
| } |
| |
| [Test] |
| public virtual void TestRegexps() |
| { |
| ICommonQueryParserConfiguration qp = GetParserConfig(new MockAnalyzer(Random, MockTokenizer.WHITESPACE, false)); |
| RegexpQuery q = new RegexpQuery(new Term("field", "[a-z][123]")); |
| assertEquals(q, GetQuery("/[a-z][123]/", qp)); |
| qp.LowercaseExpandedTerms = (true); |
| assertEquals(q, GetQuery("/[A-Z][123]/", qp)); |
| q.Boost = (0.5f); |
| assertEquals(q, GetQuery("/[A-Z][123]/^0.5", qp)); |
| qp.MultiTermRewriteMethod = MultiTermQuery.SCORING_BOOLEAN_QUERY_REWRITE; |
| q.MultiTermRewriteMethod = MultiTermQuery.SCORING_BOOLEAN_QUERY_REWRITE; |
| assertTrue(GetQuery("/[A-Z][123]/^0.5", qp) is RegexpQuery); |
| assertEquals(MultiTermQuery.SCORING_BOOLEAN_QUERY_REWRITE, ((RegexpQuery)GetQuery("/[A-Z][123]/^0.5", qp)).MultiTermRewriteMethod); |
| assertEquals(q, GetQuery("/[A-Z][123]/^0.5", qp)); |
| qp.MultiTermRewriteMethod=(MultiTermQuery.CONSTANT_SCORE_AUTO_REWRITE_DEFAULT); |
| |
| Query escaped = new RegexpQuery(new Term("field", "[a-z]\\/[123]")); |
| assertEquals(escaped, GetQuery("/[a-z]\\/[123]/", qp)); |
| Query escaped2 = new RegexpQuery(new Term("field", "[a-z]\\*[123]")); |
| assertEquals(escaped2, GetQuery("/[a-z]\\*[123]/", qp)); |
| |
| BooleanQuery complex = new BooleanQuery(); |
| complex.Add(new RegexpQuery(new Term("field", "[a-z]\\/[123]")), Occur.MUST); |
| complex.Add(new TermQuery(new Term("path", "/etc/init.d/")), Occur.MUST); |
| complex.Add(new TermQuery(new Term("field", "/etc/init[.]d/lucene/")), Occur.SHOULD); |
| assertEquals(complex, GetQuery("/[a-z]\\/[123]/ AND path:\"/etc/init.d/\" OR \"/etc\\/init\\[.\\]d/lucene/\" ", qp)); |
| |
| Query re = new RegexpQuery(new Term("field", "http.*")); |
| assertEquals(re, GetQuery("field:/http.*/", qp)); |
| assertEquals(re, GetQuery("/http.*/", qp)); |
| |
| re = new RegexpQuery(new Term("field", "http~0.5")); |
| assertEquals(re, GetQuery("field:/http~0.5/", qp)); |
| assertEquals(re, GetQuery("/http~0.5/", qp)); |
| |
| re = new RegexpQuery(new Term("field", "boo")); |
| assertEquals(re, GetQuery("field:/boo/", qp)); |
| assertEquals(re, GetQuery("/boo/", qp)); |
| |
| assertEquals(new TermQuery(new Term("field", "/boo/")), GetQuery("\"/boo/\"", qp)); |
| assertEquals(new TermQuery(new Term("field", "/boo/")), GetQuery("\\/boo\\/", qp)); |
| |
| BooleanQuery two = new BooleanQuery(); |
| two.Add(new RegexpQuery(new Term("field", "foo")), Occur.SHOULD); |
| two.Add(new RegexpQuery(new Term("field", "bar")), Occur.SHOULD); |
| assertEquals(two, GetQuery("field:/foo/ field:/bar/", qp)); |
| assertEquals(two, GetQuery("/foo/ /bar/", qp)); |
| } |
| |
| [Test] |
| public virtual void TestStopwords() |
| { |
| CharacterRunAutomaton stopSet = new CharacterRunAutomaton(new RegExp("the|foo").ToAutomaton()); |
| ICommonQueryParserConfiguration qp = GetParserConfig(new MockAnalyzer(Random, MockTokenizer.SIMPLE, true, stopSet)); |
| Query result = GetQuery("field:the OR field:foo", qp); |
| assertNotNull("result is null and it shouldn't be", result); |
| assertTrue("result is not a BooleanQuery", result is BooleanQuery); |
| assertTrue(((BooleanQuery)result).Clauses.Count + " does not equal: " + 0, ((BooleanQuery)result).Clauses.Count == 0); |
| result = GetQuery("field:woo OR field:the", qp); |
| assertNotNull("result is null and it shouldn't be", result); |
| assertTrue("result is not a TermQuery", result is TermQuery); |
| result = GetQuery("(fieldX:xxxxx OR fieldy:xxxxxxxx)^2 AND (fieldx:the OR fieldy:foo)", qp); |
| assertNotNull("result is null and it shouldn't be", result); |
| assertTrue("result is not a BooleanQuery", result is BooleanQuery); |
| if (Verbose) Console.WriteLine("Result: " + result); |
| assertTrue(((BooleanQuery)result).Clauses.Count + " does not equal: " + 2, ((BooleanQuery)result).Clauses.Count == 2); |
| } |
| |
| [Test] |
| public virtual void TestPositionIncrement() |
| { |
| ICommonQueryParserConfiguration qp = GetParserConfig(new MockAnalyzer(Random, MockTokenizer.SIMPLE, true, MockTokenFilter.ENGLISH_STOPSET)); |
| qp.EnablePositionIncrements = (true); |
| String qtxt = "\"the words in poisitions pos02578 are stopped in this phrasequery\""; |
| // 0 2 5 7 8 |
| int[] expectedPositions = { 1, 3, 4, 6, 9 }; |
| PhraseQuery pq = (PhraseQuery)GetQuery(qtxt, qp); |
| //System.out.println("Query text: "+qtxt); |
| //System.out.println("Result: "+pq); |
| Term[] t = pq.GetTerms(); |
| int[] pos = pq.GetPositions(); |
| for (int i = 0; i < t.Length; i++) |
| { |
| //System.out.println(i+". "+t[i]+" pos: "+pos[i]); |
| assertEquals("term " + i + " = " + t[i] + " has wrong term-position!", expectedPositions[i], pos[i]); |
| } |
| } |
| |
| [Test] |
| public virtual void TestMatchAllDocs() |
| { |
| ICommonQueryParserConfiguration qp = GetParserConfig(new MockAnalyzer(Random, MockTokenizer.WHITESPACE, false)); |
| assertEquals(new MatchAllDocsQuery(), GetQuery("*:*", qp)); |
| assertEquals(new MatchAllDocsQuery(), GetQuery("(*:*)", qp)); |
| BooleanQuery bq = (BooleanQuery)GetQuery("+*:* -*:*", qp); |
| assertTrue(bq.GetClauses()[0].Query is MatchAllDocsQuery); |
| assertTrue(bq.GetClauses()[1].Query is MatchAllDocsQuery); |
| } |
| |
| private void AssertHits(int expected, String query, IndexSearcher @is) |
| { |
| string oldDefaultField = DefaultField; |
| DefaultField = "date"; |
| ICommonQueryParserConfiguration qp = GetParserConfig(new MockAnalyzer(Random, MockTokenizer.WHITESPACE, false)); |
| qp.Locale = new CultureInfo("en"); |
| Query q = GetQuery(query, qp); |
| ScoreDoc[] hits = @is.Search(q, null, 1000).ScoreDocs; |
| assertEquals(expected, hits.Length); |
| DefaultField = oldDefaultField; |
| } |
| |
| public override void TearDown() |
| { |
| BooleanQuery.MaxClauseCount = originalMaxClauses; |
| base.TearDown(); |
| } |
| |
| // LUCENE-2002: make sure defaults for StandardAnalyzer's |
| // enableStopPositionIncr & QueryParser's enablePosIncr |
| // "match" |
| [Test] |
| public virtual void TestPositionIncrements() |
| { |
| using (Directory dir = NewDirectory()) |
| { |
| Analyzer a = new MockAnalyzer(Random, MockTokenizer.SIMPLE, true, MockTokenFilter.ENGLISH_STOPSET); |
| using (IndexWriter w = new IndexWriter(dir, NewIndexWriterConfig(TEST_VERSION_CURRENT, a))) |
| { |
| Document doc = new Document(); |
| doc.Add(NewTextField("field", "the wizard of ozzy", Field.Store.NO)); |
| w.AddDocument(doc); |
| using (IndexReader r = DirectoryReader.Open(w, true)) |
| { |
| IndexSearcher s = NewSearcher(r); |
| |
| Query q = GetQuery("\"wizard of ozzy\"", a); |
| assertEquals(1, s.Search(q, 1).TotalHits); |
| } |
| } |
| } |
| } |
| |
| /// <summary> |
| /// adds synonym of "dog" for "dogs". |
| /// </summary> |
| protected class MockSynonymFilter : TokenFilter |
| { |
| ICharTermAttribute termAtt; |
| IPositionIncrementAttribute posIncAtt; |
| bool addSynonym = false; |
| |
| public MockSynonymFilter(TokenStream input) |
| : base(input) |
| { |
| termAtt = AddAttribute<ICharTermAttribute>(); |
| posIncAtt = AddAttribute<IPositionIncrementAttribute>(); |
| } |
| |
| public override sealed bool IncrementToken() |
| { |
| if (addSynonym) |
| { // inject our synonym |
| ClearAttributes(); |
| termAtt.SetEmpty().Append("dog"); |
| posIncAtt.PositionIncrement = (0); |
| addSynonym = false; |
| return true; |
| } |
| |
| if (m_input.IncrementToken()) |
| { |
| addSynonym = termAtt.toString().Equals("dogs", StringComparison.Ordinal); |
| return true; |
| } |
| else |
| { |
| return false; |
| } |
| } |
| } |
| |
| /// <summary> |
| /// whitespace+lowercase analyzer without synonyms |
| /// </summary> |
| protected class Analyzer1 : Analyzer |
| { |
| public Analyzer1() |
| { } |
| |
| protected internal override TokenStreamComponents CreateComponents(string fieldName, System.IO.TextReader reader) |
| { |
| Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, true); |
| return new TokenStreamComponents(tokenizer, new MockSynonymFilter(tokenizer)); |
| } |
| } |
| |
| /// <summary> |
| /// whitespace+lowercase analyzer without synonyms |
| /// </summary> |
| protected class Analyzer2 : Analyzer |
| { |
| public Analyzer2() |
| { } |
| |
| protected internal override TokenStreamComponents CreateComponents(string fieldName, System.IO.TextReader reader) |
| { |
| return new TokenStreamComponents(new MockTokenizer(reader, MockTokenizer.WHITESPACE, true)); |
| } |
| } |
| |
| public abstract void TestNewFieldQuery(); |
| |
| /// <summary> |
| /// Mock collation analyzer: indexes terms as "collated" + term |
| /// </summary> |
| private class MockCollationFilter : TokenFilter |
| { |
| private ICharTermAttribute termAtt; |
| |
| public MockCollationFilter(TokenStream input) |
| : base(input) |
| { |
| termAtt = AddAttribute<ICharTermAttribute>(); |
| } |
| |
| public override sealed bool IncrementToken() |
| { |
| if (m_input.IncrementToken()) |
| { |
| string term = termAtt.toString(); |
| termAtt.SetEmpty().Append("collated").Append(term); |
| return true; |
| } |
| else |
| { |
| return false; |
| } |
| } |
| } |
| |
| private class MockCollationAnalyzer : Analyzer |
| { |
| protected internal override TokenStreamComponents CreateComponents(string fieldName, System.IO.TextReader reader) |
| { |
| Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, true); |
| return new TokenStreamComponents(tokenizer, new MockCollationFilter(tokenizer)); |
| } |
| } |
| |
| [Test] |
| public virtual void TestCollatedRange() |
| { |
| ICommonQueryParserConfiguration qp = GetParserConfig(new MockCollationAnalyzer()); |
| SetAnalyzeRangeTerms(qp, true); |
| Query expected = TermRangeQuery.NewStringRange(DefaultField, "collatedabc", "collateddef", true, true); |
| Query actual = GetQuery("[abc TO def]", qp); |
| assertEquals(expected, actual); |
| } |
| |
| [Test] |
| public virtual void TestDistanceAsEditsParsing() |
| { |
| FuzzyQuery q = (FuzzyQuery)GetQuery("foobar~2", new MockAnalyzer(Random)); |
| assertEquals(2, q.MaxEdits); |
| } |
| |
| [Test] |
| public virtual void TestPhraseQueryToString() |
| { |
| Analyzer analyzer = new MockAnalyzer(Random, MockTokenizer.SIMPLE, true, MockTokenFilter.ENGLISH_STOPSET); |
| ICommonQueryParserConfiguration qp = GetParserConfig(analyzer); |
| qp.EnablePositionIncrements = (true); |
| PhraseQuery q = (PhraseQuery)GetQuery("\"this hi this is a test is\"", qp); |
| assertEquals("field:\"? hi ? ? ? test\"", q.toString()); |
| } |
| |
| [Test] |
| public virtual void TestParseWildcardAndPhraseQueries() |
| { |
| string field = "content"; |
| string oldDefaultField = DefaultField; |
| DefaultField = (field); |
| ICommonQueryParserConfiguration qp = GetParserConfig(new MockAnalyzer(Random)); |
| qp.AllowLeadingWildcard=(true); |
| |
| string[][] prefixQueries = new string[3][] { |
| new string[] {"a*", "ab*", "abc*",}, |
| new string[] {"h*", "hi*", "hij*", "\\\\7*"}, |
| new string[] {"o*", "op*", "opq*", "\\\\\\\\*"}, |
| }; |
| |
| string[][] wildcardQueries = new string[3][] { |
| new string[] {"*a*", "*ab*", "*abc**", "ab*e*", "*g?", "*f?1", "abc**"}, |
| new string[] {"*h*", "*hi*", "*hij**", "hi*k*", "*n?", "*m?1", "hij**"}, |
| new string[] {"*o*", "*op*", "*opq**", "op*q*", "*u?", "*t?1", "opq**"}, |
| }; |
| |
| // test queries that must be prefix queries |
| for (int i = 0; i < prefixQueries.Length; i++) |
| { |
| for (int j = 0; j < prefixQueries[i].Length; j++) |
| { |
| string queryString = prefixQueries[i][j]; |
| Query q = GetQuery(queryString, qp); |
| assertEquals(typeof(PrefixQuery), q.GetType()); |
| } |
| } |
| |
| // test queries that must be wildcard queries |
| for (int i = 0; i < wildcardQueries.Length; i++) |
| { |
| for (int j = 0; j < wildcardQueries[i].Length; j++) |
| { |
| string qtxt = wildcardQueries[i][j]; |
| Query q = GetQuery(qtxt, qp); |
| assertEquals(typeof(WildcardQuery), q.GetType()); |
| } |
| } |
| DefaultField = (oldDefaultField); |
| } |
| |
| [Test] |
| public virtual void TestPhraseQueryPositionIncrements() |
| { |
| CharacterRunAutomaton stopStopList = |
| new CharacterRunAutomaton(new RegExp("[sS][tT][oO][pP]").ToAutomaton()); |
| |
| ICommonQueryParserConfiguration qp = GetParserConfig(new MockAnalyzer(Random, MockTokenizer.WHITESPACE, false, stopStopList)); |
| |
| qp = GetParserConfig( |
| new MockAnalyzer(Random, MockTokenizer.WHITESPACE, false, stopStopList)); |
| qp.EnablePositionIncrements=(true); |
| |
| PhraseQuery phraseQuery = new PhraseQuery(); |
| phraseQuery.Add(new Term("field", "1")); |
| phraseQuery.Add(new Term("field", "2"), 2); |
| assertEquals(phraseQuery, GetQuery("\"1 stop 2\"", qp)); |
| } |
| |
| [Test] |
| public virtual void TestMatchAllQueryParsing() |
| { |
| // test simple parsing of MatchAllDocsQuery |
| string oldDefaultField = DefaultField; |
| DefaultField = ("key"); |
| ICommonQueryParserConfiguration qp = GetParserConfig(new MockAnalyzer(Random)); |
| assertEquals(new MatchAllDocsQuery(), GetQuery(new MatchAllDocsQuery().toString(), qp)); |
| |
| // test parsing with non-default boost |
| MatchAllDocsQuery query = new MatchAllDocsQuery(); |
| query.Boost = (2.3f); |
| assertEquals(query, GetQuery(query.toString(), qp)); |
| DefaultField = (oldDefaultField); |
| } |
| |
| [Test] |
| public virtual void TestNestedAndClausesFoo() |
| { |
| string query = "(field1:[1 TO *] AND field1:[* TO 2]) AND field2:(z)"; |
| BooleanQuery q = new BooleanQuery(); |
| BooleanQuery bq = new BooleanQuery(); |
| bq.Add(TermRangeQuery.NewStringRange("field1", "1", null, true, true), Occur.MUST); |
| bq.Add(TermRangeQuery.NewStringRange("field1", null, "2", true, true), Occur.MUST); |
| q.Add(bq, Occur.MUST); |
| q.Add(new TermQuery(new Term("field2", "z")), Occur.MUST); |
| assertEquals(q, GetQuery(query, new MockAnalyzer(Random))); |
| } |
| } |
| } |