/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.lucene.codecs.lucene60; import java.io.Closeable; import java.io.IOException; import java.util.ArrayList; import java.util.Collection; import java.util.Collections; import java.util.HashMap; import java.util.List; import java.util.Map; import org.apache.lucene.codecs.CodecUtil; import org.apache.lucene.codecs.PointsReader; import org.apache.lucene.index.FieldInfo; import org.apache.lucene.index.IndexFileNames; import org.apache.lucene.index.PointValues; import org.apache.lucene.index.SegmentReadState; import org.apache.lucene.store.ChecksumIndexInput; import org.apache.lucene.store.IndexInput; import org.apache.lucene.util.Accountable; import org.apache.lucene.util.Accountables; import org.apache.lucene.util.IOUtils; import org.apache.lucene.util.bkd.BKDReader; /** Reads point values previously written with {@link Lucene60PointsWriter} */ public class Lucene60PointsReader extends PointsReader implements Closeable { final IndexInput dataIn; final SegmentReadState readState; final Map<Integer,BKDReader> readers = new HashMap<>(); /** Sole constructor */ public Lucene60PointsReader(SegmentReadState readState) throws IOException { this.readState = readState; String indexFileName = IndexFileNames.segmentFileName(readState.segmentInfo.name, readState.segmentSuffix, Lucene60PointsFormat.INDEX_EXTENSION); Map<Integer,Long> fieldToFileOffset = new HashMap<>(); // Read index file try (ChecksumIndexInput indexIn = readState.directory.openChecksumInput(indexFileName, readState.context)) { Throwable priorE = null; try { CodecUtil.checkIndexHeader(indexIn, Lucene60PointsFormat.META_CODEC_NAME, Lucene60PointsFormat.INDEX_VERSION_START, Lucene60PointsFormat.INDEX_VERSION_START, readState.segmentInfo.getId(), readState.segmentSuffix); int count = indexIn.readVInt(); for(int i=0;i<count;i++) { int fieldNumber = indexIn.readVInt(); long fp = indexIn.readVLong(); fieldToFileOffset.put(fieldNumber, fp); } } catch (Throwable t) { priorE = t; } finally { CodecUtil.checkFooter(indexIn, priorE); } } String dataFileName = IndexFileNames.segmentFileName(readState.segmentInfo.name, readState.segmentSuffix, Lucene60PointsFormat.DATA_EXTENSION); boolean success = false; dataIn = readState.directory.openInput(dataFileName, readState.context); try { CodecUtil.checkIndexHeader(dataIn, Lucene60PointsFormat.DATA_CODEC_NAME, Lucene60PointsFormat.DATA_VERSION_START, Lucene60PointsFormat.DATA_VERSION_START, readState.segmentInfo.getId(), readState.segmentSuffix); // NOTE: data file is too costly to verify checksum against all the bytes on open, // but for now we at least verify proper structure of the checksum footer: which looks // for FOOTER_MAGIC + algorithmID. This is cheap and can detect some forms of corruption // such as file truncation. CodecUtil.retrieveChecksum(dataIn); for(Map.Entry<Integer,Long> ent : fieldToFileOffset.entrySet()) { int fieldNumber = ent.getKey(); long fp = ent.getValue(); dataIn.seek(fp); BKDReader reader = new BKDReader(dataIn); readers.put(fieldNumber, reader); } success = true; } finally { if (success == false) { IOUtils.closeWhileHandlingException(this); } } } /** Returns the underlying {@link BKDReader}. * * @lucene.internal */ @Override public PointValues getValues(String fieldName) { FieldInfo fieldInfo = readState.fieldInfos.fieldInfo(fieldName); if (fieldInfo == null) { throw new IllegalArgumentException("field=\"" + fieldName + "\" is unrecognized"); } if (fieldInfo.getPointDimensionCount() == 0) { throw new IllegalArgumentException("field=\"" + fieldName + "\" did not index point values"); } return readers.get(fieldInfo.number); } @Override public long ramBytesUsed() { long sizeInBytes = 0; for(BKDReader reader : readers.values()) { sizeInBytes += reader.ramBytesUsed(); } return sizeInBytes; } @Override public Collection<Accountable> getChildResources() { List<Accountable> resources = new ArrayList<>(); for(Map.Entry<Integer,BKDReader> ent : readers.entrySet()) { resources.add(Accountables.namedAccountable(readState.fieldInfos.fieldInfo(ent.getKey()).name, ent.getValue())); } return Collections.unmodifiableList(resources); } @Override public void checkIntegrity() throws IOException { CodecUtil.checksumEntireFile(dataIn); } @Override public void close() throws IOException { dataIn.close(); // Free up heap: readers.clear(); } }