package org.apache.lucene.index; /** * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ import java.io.IOException; import org.apache.lucene.index.FieldInfo.IndexOptions; import org.apache.lucene.index.FreqProxTermsWriterPerField.FreqProxPostingsArray; // TODO FI: some of this is "generic" to TermsHash* so we // should factor it out so other consumers don't have to // duplicate this code /** Used by DocumentsWriter to merge the postings from * multiple ThreadStates when creating a segment */ final class FreqProxFieldMergeState { final FreqProxTermsWriterPerField field; final int numPostings; final CharBlockPool charPool; final int[] termIDs; final FreqProxPostingsArray postings; int currentTermID; char[] text; int textOffset; private int postingUpto = -1; final ByteSliceReader freq = new ByteSliceReader(); final ByteSliceReader prox = new ByteSliceReader(); int docID; int termFreq; public FreqProxFieldMergeState(FreqProxTermsWriterPerField field) { this.field = field; this.charPool = field.perThread.termsHashPerThread.charPool; this.numPostings = field.termsHashPerField.numPostings; this.termIDs = field.termsHashPerField.sortPostings(); this.postings = (FreqProxPostingsArray) field.termsHashPerField.postingsArray; } boolean nextTerm() throws IOException { postingUpto++; if (postingUpto == numPostings) return false; currentTermID = termIDs[postingUpto]; docID = 0; final int textStart = postings.textStarts[currentTermID]; text = charPool.buffers[textStart >> DocumentsWriter.CHAR_BLOCK_SHIFT]; textOffset = textStart & DocumentsWriter.CHAR_BLOCK_MASK; field.termsHashPerField.initReader(freq, currentTermID, 0); if (field.fieldInfo.indexOptions == IndexOptions.DOCS_AND_FREQS_AND_POSITIONS) field.termsHashPerField.initReader(prox, currentTermID, 1); // Should always be true boolean result = nextDoc(); assert result; return true; } public String termText() { int upto = textOffset; while(text[upto] != 0xffff) { upto++; } return new String(text, textOffset, upto-textOffset); } public boolean nextDoc() throws IOException { if (freq.eof()) { if (postings.lastDocCodes[currentTermID] != -1) { // Return last doc docID = postings.lastDocIDs[currentTermID]; if (field.indexOptions != IndexOptions.DOCS_ONLY) termFreq = postings.docFreqs[currentTermID]; postings.lastDocCodes[currentTermID] = -1; return true; } else // EOF return false; } final int code = freq.readVInt(); if (field.indexOptions == IndexOptions.DOCS_ONLY) docID += code; else { docID += code >>> 1; if ((code & 1) != 0) termFreq = 1; else termFreq = freq.readVInt(); } assert docID != postings.lastDocIDs[currentTermID]; return true; } }