package org.apache.lucene.index;
/**
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
import java.io.IOException;
import java.util.Comparator;
import org.apache.lucene.util.BytesRef;
import org.apache.lucene.index.FreqProxTermsWriterPerField.FreqProxPostingsArray;
// TODO FI: some of this is "generic" to TermsHash* so we
// should factor it out so other consumers don't have to
// duplicate this code
/** Used by DocumentsWriter to merge the postings from
* multiple ThreadStates when creating a segment */
final class FreqProxFieldMergeState {
final FreqProxTermsWriterPerField field;
final int numPostings;
private final ByteBlockPool bytePool;
final int[] termIDs;
final FreqProxPostingsArray postings;
int currentTermID;
final BytesRef text = new BytesRef();
private int postingUpto = -1;
final ByteSliceReader freq = new ByteSliceReader();
final ByteSliceReader prox = new ByteSliceReader();
int docID;
int termFreq;
public FreqProxFieldMergeState(FreqProxTermsWriterPerField field, Comparator<BytesRef> termComp) {
this.field = field;
this.numPostings = field.termsHashPerField.numPostings;
this.bytePool = field.perThread.termsHashPerThread.bytePool;
this.termIDs = field.termsHashPerField.sortPostings(termComp);
this.postings = (FreqProxPostingsArray) field.termsHashPerField.postingsArray;
}
boolean nextTerm() throws IOException {
postingUpto++;
if (postingUpto == numPostings) {
return false;
}
currentTermID = termIDs[postingUpto];
docID = 0;
// Get BytesRef
final int textStart = postings.textStarts[currentTermID];
bytePool.setBytesRef(text, textStart);
field.termsHashPerField.initReader(freq, currentTermID, 0);
if (!field.fieldInfo.omitTermFreqAndPositions) {
field.termsHashPerField.initReader(prox, currentTermID, 1);
}
// Should always be true
boolean result = nextDoc();
assert result;
return true;
}
public boolean nextDoc() throws IOException {
if (freq.eof()) {
if (postings.lastDocCodes[currentTermID] != -1) {
// Return last doc
docID = postings.lastDocIDs[currentTermID];
if (!field.omitTermFreqAndPositions)
termFreq = postings.docFreqs[currentTermID];
postings.lastDocCodes[currentTermID] = -1;
return true;
} else
// EOF
return false;
}
final int code = freq.readVInt();
if (field.omitTermFreqAndPositions)
docID += code;
else {
docID += code >>> 1;
if ((code & 1) != 0)
termFreq = 1;
else
termFreq = freq.readVInt();
}
assert docID != postings.lastDocIDs[currentTermID];
return true;
}
}