| /* |
| * Licensed to the Apache Software Foundation (ASF) under one or more |
| * contributor license agreements. See the NOTICE file distributed with |
| * this work for additional information regarding copyright ownership. |
| * The ASF licenses this file to You under the Apache License, Version 2.0 |
| * (the "License"); you may not use this file except in compliance with |
| * the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| package org.apache.lucene.index; |
| |
| |
| import java.io.IOException; |
| import java.util.HashMap; |
| import java.util.Map; |
| |
| import org.apache.lucene.codecs.NormsProducer; |
| import org.apache.lucene.util.ByteBlockPool; |
| import org.apache.lucene.util.Counter; |
| import org.apache.lucene.util.IntBlockPool; |
| |
| /** This class is passed each token produced by the analyzer |
| * on each field during indexing, and it stores these |
| * tokens in a hash table, and allocates separate byte |
| * streams per token. Consumers of this class, eg {@link |
| * FreqProxTermsWriter} and {@link TermVectorsConsumer}, |
| * write their own byte streams under each term. */ |
| abstract class TermsHash { |
| |
| final TermsHash nextTermsHash; |
| |
| final IntBlockPool intPool; |
| final ByteBlockPool bytePool; |
| ByteBlockPool termBytePool; |
| final Counter bytesUsed; |
| |
| TermsHash(final IntBlockPool.Allocator intBlockAllocator, final ByteBlockPool.Allocator byteBlockAllocator, Counter bytesUsed, TermsHash nextTermsHash) { |
| this.nextTermsHash = nextTermsHash; |
| this.bytesUsed = bytesUsed; |
| intPool = new IntBlockPool(intBlockAllocator); |
| bytePool = new ByteBlockPool(byteBlockAllocator); |
| |
| if (nextTermsHash != null) { |
| // We are primary |
| termBytePool = bytePool; |
| nextTermsHash.termBytePool = bytePool; |
| } |
| } |
| |
| public void abort() { |
| try { |
| reset(); |
| } finally { |
| if (nextTermsHash != null) { |
| nextTermsHash.abort(); |
| } |
| } |
| } |
| |
| // Clear all state |
| void reset() { |
| // we don't reuse so we drop everything and don't fill with 0 |
| intPool.reset(false, false); |
| bytePool.reset(false, false); |
| } |
| |
| void flush(Map<String,TermsHashPerField> fieldsToFlush, final SegmentWriteState state, |
| Sorter.DocMap sortMap, NormsProducer norms) throws IOException { |
| if (nextTermsHash != null) { |
| Map<String,TermsHashPerField> nextChildFields = new HashMap<>(); |
| for (final Map.Entry<String,TermsHashPerField> entry : fieldsToFlush.entrySet()) { |
| nextChildFields.put(entry.getKey(), entry.getValue().getNextPerField()); |
| } |
| nextTermsHash.flush(nextChildFields, state, sortMap, norms); |
| } |
| } |
| |
| abstract TermsHashPerField addField(FieldInvertState fieldInvertState, FieldInfo fieldInfo); |
| |
| void finishDocument(int docID) throws IOException { |
| if (nextTermsHash != null) { |
| nextTermsHash.finishDocument(docID); |
| } |
| } |
| |
| void startDocument() throws IOException { |
| if (nextTermsHash != null) { |
| nextTermsHash.startDocument(); |
| } |
| } |
| } |