/*
* Licensed to Elasticsearch under one or more contributor
* license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright
* ownership. Elasticsearch licenses this file to you under
* the Apache License, Version 2.0 (the "License"); you may
* not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package org.elasticsearch.search.aggregations.bucket.terms;
import org.elasticsearch.search.aggregations.pipeline.PipelineAggregator;
import org.elasticsearch.test.InternalAggregationTestCase;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
import java.util.Map.Entry;
import java.util.Optional;
import java.util.stream.Collectors;
import java.util.stream.Stream;
public abstract class InternalTermsTestCase extends InternalAggregationTestCase<InternalTerms<?,?>> {
@Override
protected InternalTerms<?, ?> createUnmappedInstance(
String name,
List<PipelineAggregator> pipelineAggregators,
Map<String, Object> metaData) {
InternalTerms<?, ?> testInstance = createTestInstance(name, pipelineAggregators, metaData);
return new UnmappedTerms(name, testInstance.order, testInstance.requiredSize, testInstance.minDocCount,
pipelineAggregators, metaData);
}
@Override
protected void assertReduced(InternalTerms<?, ?> reduced, List<InternalTerms<?, ?>> inputs) {
final int requiredSize = inputs.get(0).requiredSize;
Map<Object, Long> reducedCounts = toCounts(reduced.getBuckets().stream());
Map<Object, Long> totalCounts = toCounts(inputs.stream().map(Terms::getBuckets).flatMap(List::stream));
assertEquals(reducedCounts.size() == requiredSize,
totalCounts.size() >= requiredSize);
Map<Object, Long> expectedReducedCounts = new HashMap<>(totalCounts);
expectedReducedCounts.keySet().retainAll(reducedCounts.keySet());
assertEquals(expectedReducedCounts, reducedCounts);
final long minFinalcount = reduced.getBuckets().isEmpty()
? -1
: reduced.getBuckets().get(reduced.getBuckets().size() - 1).getDocCount();
Map<Object, Long> evictedTerms = new HashMap<>(totalCounts);
evictedTerms.keySet().removeAll(reducedCounts.keySet());
Optional<Entry<Object, Long>> missingTerm = evictedTerms.entrySet().stream()
.filter(e -> e.getValue() > minFinalcount).findAny();
if (missingTerm.isPresent()) {
fail("Missed term: " + missingTerm + " from " + reducedCounts);
}
final long reducedTotalDocCount = reduced.getSumOfOtherDocCounts()
+ reduced.getBuckets().stream().mapToLong(Terms.Bucket::getDocCount).sum();
final long expectedTotalDocCount = inputs.stream().map(Terms::getBuckets)
.flatMap(List::stream).mapToLong(Terms.Bucket::getDocCount).sum();
assertEquals(expectedTotalDocCount, reducedTotalDocCount);
}
private static Map<Object, Long> toCounts(Stream<? extends Terms.Bucket> buckets) {
return buckets.collect(Collectors.toMap(
Terms.Bucket::getKey,
Terms.Bucket::getDocCount,
Long::sum));
}
}