| using Lucene.Net.Analysis.TokenAttributes; |
| using Lucene.Net.Documents; |
| using Lucene.Net.Index.Extensions; |
| using Lucene.Net.Store; |
| using NUnit.Framework; |
| using System; |
| using Console = Lucene.Net.Util.SystemConsole; |
| |
| namespace Lucene.Net.Index |
| { |
| /* |
| * Licensed to the Apache Software Foundation (ASF) under one or more |
| * contributor license agreements. See the NOTICE file distributed with |
| * this work for additional information regarding copyright ownership. |
| * The ASF licenses this file to You under the Apache License, Version 2.0 |
| * (the "License"); you may not use this file except in compliance with |
| * the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| using BaseDirectoryWrapper = Lucene.Net.Store.BaseDirectoryWrapper; |
| using Document = Documents.Document; |
| using Field = Field; |
| using FieldType = FieldType; |
| using LuceneTestCase = Lucene.Net.Util.LuceneTestCase; |
| using MockAnalyzer = Lucene.Net.Analysis.MockAnalyzer; |
| using MockDirectoryWrapper = Lucene.Net.Store.MockDirectoryWrapper; |
| using TextField = TextField; |
| using TokenStream = Lucene.Net.Analysis.TokenStream; |
| |
| /// <summary> |
| /// Test indexes ~82M docs with 26 terms each, so you get > Integer.MAX_VALUE terms/docs pairs |
| /// @lucene.experimental |
| /// </summary> |
| [SuppressCodecs("SimpleText", "Memory", "Direct", "Compressing")] |
| [TestFixture] |
| public class Test2BPostings : LuceneTestCase |
| { |
| [Test] |
| [Nightly] |
| [Ignore("LUCENENET specific - takes too long to run on Azure DevOps")] |
| public virtual void Test([ValueSource(typeof(ConcurrentMergeSchedulerFactories), "Values")]Func<IConcurrentMergeScheduler> newScheduler) |
| { |
| BaseDirectoryWrapper dir = NewFSDirectory(CreateTempDir("2BPostings")); |
| if (dir is MockDirectoryWrapper) |
| { |
| ((MockDirectoryWrapper)dir).Throttling = Throttling.NEVER; |
| } |
| |
| var config = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(Random)) |
| .SetMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH) |
| .SetRAMBufferSizeMB(256.0) |
| .SetMergeScheduler(newScheduler()) |
| .SetMergePolicy(NewLogMergePolicy(false, 10)) |
| .SetOpenMode(OpenMode.CREATE); |
| |
| IndexWriter w = new IndexWriter(dir, config); |
| |
| MergePolicy mp = w.Config.MergePolicy; |
| if (mp is LogByteSizeMergePolicy) |
| { |
| // 1 petabyte: |
| ((LogByteSizeMergePolicy)mp).MaxMergeMB = 1024 * 1024 * 1024; |
| } |
| |
| Document doc = new Document(); |
| FieldType ft = new FieldType(TextField.TYPE_NOT_STORED); |
| ft.OmitNorms = true; |
| ft.IndexOptions = IndexOptions.DOCS_ONLY; |
| Field field = new Field("field", new MyTokenStream(), ft); |
| doc.Add(field); |
| |
| int numDocs = (int.MaxValue / 26) + 1; |
| for (int i = 0; i < numDocs; i++) |
| { |
| w.AddDocument(doc); |
| if (Verbose && i % 100000 == 0) |
| { |
| Console.WriteLine(i + " of " + numDocs + "..."); |
| } |
| } |
| w.ForceMerge(1); |
| w.Dispose(); |
| dir.Dispose(); |
| } |
| |
| public sealed class MyTokenStream : TokenStream |
| { |
| internal readonly ICharTermAttribute termAtt; |
| internal int index; |
| |
| public MyTokenStream() |
| { |
| termAtt = AddAttribute<ICharTermAttribute>(); |
| } |
| |
| public override bool IncrementToken() |
| { |
| if (index <= 'z') |
| { |
| ClearAttributes(); |
| termAtt.Length = 1; |
| termAtt.Buffer[0] = (char)index++; |
| return true; |
| } |
| return false; |
| } |
| |
| public override void Reset() |
| { |
| index = 'a'; |
| } |
| } |
| } |
| } |