package org.apache.lucene.index; /** * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ import java.io.IOException; import java.util.Collection; import java.util.HashMap; import java.util.HashSet; import java.util.Iterator; import java.util.Map; import org.apache.lucene.util.AttributeSource; /** This is a DocFieldConsumer that inverts each field, * separately, from a Document, and accepts a * InvertedTermsConsumer to process those terms. */ final class DocInverter extends DocFieldConsumer { final InvertedDocConsumer consumer; final InvertedDocEndConsumer endConsumer; public DocInverter(InvertedDocConsumer consumer, InvertedDocEndConsumer endConsumer) { this.consumer = consumer; this.endConsumer = endConsumer; } void setFieldInfos(FieldInfos fieldInfos) { super.setFieldInfos(fieldInfos); consumer.setFieldInfos(fieldInfos); endConsumer.setFieldInfos(fieldInfos); } void flush(Map threadsAndFields, SegmentWriteState state) throws IOException { Map childThreadsAndFields = new HashMap(); Map endChildThreadsAndFields = new HashMap(); Iterator it = threadsAndFields.entrySet().iterator(); while(it.hasNext()) { Map.Entry entry = (Map.Entry) it.next(); DocInverterPerThread perThread = (DocInverterPerThread) entry.getKey(); Collection fields = (Collection) entry.getValue(); Iterator fieldsIt = fields.iterator(); Collection childFields = new HashSet(); Collection endChildFields = new HashSet(); while(fieldsIt.hasNext()) { DocInverterPerField perField = (DocInverterPerField) fieldsIt.next(); childFields.add(perField.consumer); endChildFields.add(perField.endConsumer); } childThreadsAndFields.put(perThread.consumer, childFields); endChildThreadsAndFields.put(perThread.endConsumer, endChildFields); } consumer.flush(childThreadsAndFields, state); endConsumer.flush(endChildThreadsAndFields, state); } public void closeDocStore(SegmentWriteState state) throws IOException { consumer.closeDocStore(state); endConsumer.closeDocStore(state); } void abort() { consumer.abort(); endConsumer.abort(); } public boolean freeRAM() { return consumer.freeRAM(); } public DocFieldConsumerPerThread addThread(DocFieldProcessorPerThread docFieldProcessorPerThread) { return new DocInverterPerThread(docFieldProcessorPerThread, this); } }