/* * Licensed to ElasticSearch and Shay Banon under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. ElasticSearch licenses this * file to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, * software distributed under the License is distributed on an * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. */ package org.elasticsearch.search.facet.range; import org.apache.lucene.index.IndexReader; import org.elasticsearch.index.cache.field.data.FieldDataCache; import org.elasticsearch.index.field.data.FieldDataType; import org.elasticsearch.index.field.data.NumericFieldData; import org.elasticsearch.index.mapper.MapperService; import org.elasticsearch.search.facet.AbstractFacetCollector; import org.elasticsearch.search.facet.Facet; import org.elasticsearch.search.facet.FacetPhaseExecutionException; import org.elasticsearch.search.internal.SearchContext; import java.io.IOException; /** * */ public class KeyValueRangeFacetCollector extends AbstractFacetCollector { private final String keyIndexFieldName; private final String valueIndexFieldName; private final FieldDataCache fieldDataCache; private final FieldDataType keyFieldDataType; private NumericFieldData keyFieldData; private final FieldDataType valueFieldDataType; private final RangeFacet.Entry[] entries; private final RangeProc rangeProc; public KeyValueRangeFacetCollector(String facetName, String keyFieldName, String valueFieldName, RangeFacet.Entry[] entries, SearchContext context) { super(facetName); this.entries = entries; this.fieldDataCache = context.fieldDataCache(); MapperService.SmartNameFieldMappers smartMappers = context.smartFieldMappers(keyFieldName); if (smartMappers == null || !smartMappers.hasMapper()) { throw new FacetPhaseExecutionException(facetName, "No mapping found for field [" + keyFieldName + "]"); } // add type filter if there is exact doc mapper associated with it if (smartMappers.explicitTypeInNameWithDocMapper()) { setFilter(context.filterCache().cache(smartMappers.docMapper().typeFilter())); } keyIndexFieldName = smartMappers.mapper().names().indexName(); keyFieldDataType = smartMappers.mapper().fieldDataType(); smartMappers = context.smartFieldMappers(valueFieldName); if (smartMappers == null || !smartMappers.hasMapper()) { throw new FacetPhaseExecutionException(facetName, "No mapping found for value_field [" + valueFieldName + "]"); } valueIndexFieldName = smartMappers.mapper().names().indexName(); valueFieldDataType = smartMappers.mapper().fieldDataType(); this.rangeProc = new RangeProc(entries); } @Override protected void doSetNextReader(IndexReader reader, int docBase) throws IOException { keyFieldData = (NumericFieldData) fieldDataCache.cache(keyFieldDataType, reader, keyIndexFieldName); rangeProc.valueFieldData = (NumericFieldData) fieldDataCache.cache(valueFieldDataType, reader, valueIndexFieldName); } @Override protected void doCollect(int doc) throws IOException { for (RangeFacet.Entry entry : entries) { entry.foundInDoc = false; } keyFieldData.forEachValueInDoc(doc, rangeProc); } @Override public Facet facet() { return new InternalRangeFacet(facetName, entries); } public static class RangeProc implements NumericFieldData.DoubleValueInDocProc { private final RangeFacet.Entry[] entries; NumericFieldData valueFieldData; public RangeProc(RangeFacet.Entry[] entries) { this.entries = entries; } @Override public void onValue(int docId, double value) { for (RangeFacet.Entry entry : entries) { if (entry.foundInDoc) { continue; } if (value >= entry.getFrom() && value < entry.getTo()) { entry.foundInDoc = true; entry.count++; if (valueFieldData.multiValued()) { double[] valuesValues = valueFieldData.doubleValues(docId); entry.totalCount += valuesValues.length; for (double valueValue : valuesValues) { entry.total += valueValue; if (valueValue < entry.min) { entry.min = valueValue; } if (valueValue > entry.max) { entry.max = valueValue; } } } else { double valueValue = valueFieldData.doubleValue(docId); entry.totalCount++; entry.total += valueValue; if (valueValue < entry.min) { entry.min = valueValue; } if (valueValue > entry.max) { entry.max = valueValue; } } } } } } }