package org.apache.lucene.index; /** * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ import java.io.IOException; import java.util.Random; import org.apache.lucene.analysis.MockAnalyzer; import org.apache.lucene.document.Document; import org.apache.lucene.document.Field; import org.apache.lucene.index.IndexWriterConfig.OpenMode; import org.apache.lucene.store.Directory; import org.apache.lucene.util._TestUtil; import org.apache.lucene.util.LuceneTestCase; public class TestIndexWriterMergePolicy extends LuceneTestCase { private Random random; @Override public void setUp() throws Exception { super.setUp(); random = newRandom(); } // Test the normal case public void testNormalCase() throws IOException { Directory dir = newDirectory(random); IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(random, TEST_VERSION_CURRENT, new MockAnalyzer()) .setMaxBufferedDocs(10).setMergePolicy(new LogDocMergePolicy())); for (int i = 0; i < 100; i++) { addDoc(writer); checkInvariants(writer); } writer.close(); dir.close(); } // Test to see if there is over merge public void testNoOverMerge() throws IOException { Directory dir = newDirectory(random); IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(random, TEST_VERSION_CURRENT, new MockAnalyzer()) .setMaxBufferedDocs(10).setMergePolicy(new LogDocMergePolicy())); boolean noOverMerge = false; for (int i = 0; i < 100; i++) { addDoc(writer); checkInvariants(writer); if (writer.getNumBufferedDocuments() + writer.getSegmentCount() >= 18) { noOverMerge = true; } } assertTrue(noOverMerge); writer.close(); dir.close(); } // Test the case where flush is forced after every addDoc public void testForceFlush() throws IOException { Directory dir = newDirectory(random); LogDocMergePolicy mp = new LogDocMergePolicy(); mp.setMinMergeDocs(100); mp.setMergeFactor(10); IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(random, TEST_VERSION_CURRENT, new MockAnalyzer()) .setMaxBufferedDocs(10).setMergePolicy(mp)); for (int i = 0; i < 100; i++) { addDoc(writer); writer.close(); mp = new LogDocMergePolicy(); mp.setMergeFactor(10); writer = new IndexWriter(dir, newIndexWriterConfig(random, TEST_VERSION_CURRENT, new MockAnalyzer()).setOpenMode( OpenMode.APPEND).setMaxBufferedDocs(10).setMergePolicy(mp)); mp.setMinMergeDocs(100); checkInvariants(writer); } writer.close(); dir.close(); } // Test the case where mergeFactor changes public void testMergeFactorChange() throws IOException { Directory dir = newDirectory(random); IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(random, TEST_VERSION_CURRENT, new MockAnalyzer()) .setMaxBufferedDocs(10).setMergePolicy(new LogDocMergePolicy())); for (int i = 0; i < 250; i++) { addDoc(writer); checkInvariants(writer); } ((LogMergePolicy) writer.getConfig().getMergePolicy()).setMergeFactor(5); // merge policy only fixes segments on levels where merges // have been triggered, so check invariants after all adds for (int i = 0; i < 10; i++) { addDoc(writer); } checkInvariants(writer); writer.close(); dir.close(); } // Test the case where both mergeFactor and maxBufferedDocs change public void testMaxBufferedDocsChange() throws IOException { Directory dir = newDirectory(random); IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(random, TEST_VERSION_CURRENT, new MockAnalyzer()) .setMaxBufferedDocs(101).setMergePolicy(new LogDocMergePolicy())); // leftmost* segment has 1 doc // rightmost* segment has 100 docs for (int i = 1; i <= 100; i++) { for (int j = 0; j < i; j++) { addDoc(writer); checkInvariants(writer); } writer.close(); writer = new IndexWriter(dir, newIndexWriterConfig(random, TEST_VERSION_CURRENT, new MockAnalyzer()).setOpenMode( OpenMode.APPEND).setMaxBufferedDocs(101).setMergePolicy( new LogDocMergePolicy())); } writer.close(); LogDocMergePolicy ldmp = new LogDocMergePolicy(); ldmp.setMergeFactor(10); writer = new IndexWriter(dir, newIndexWriterConfig(random, TEST_VERSION_CURRENT, new MockAnalyzer()).setOpenMode( OpenMode.APPEND).setMaxBufferedDocs(10).setMergePolicy(ldmp).setMergeScheduler(new ConcurrentMergeScheduler())); // merge policy only fixes segments on levels where merges // have been triggered, so check invariants after all adds for (int i = 0; i < 100; i++) { addDoc(writer); } checkInvariants(writer); for (int i = 100; i < 1000; i++) { addDoc(writer); } writer.commit(); ((ConcurrentMergeScheduler) writer.getConfig().getMergeScheduler()).sync(); writer.commit(); checkInvariants(writer); writer.close(); dir.close(); } // Test the case where a merge results in no doc at all public void testMergeDocCount0() throws IOException { Directory dir = newDirectory(random); LogDocMergePolicy ldmp = new LogDocMergePolicy(); ldmp.setMergeFactor(100); IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(random, TEST_VERSION_CURRENT, new MockAnalyzer()) .setMaxBufferedDocs(10).setMergePolicy(ldmp)); for (int i = 0; i < 250; i++) { addDoc(writer); checkInvariants(writer); } writer.close(); IndexReader reader = IndexReader.open(dir, false); reader.deleteDocuments(new Term("content", "aaa")); reader.close(); ldmp = new LogDocMergePolicy(); ldmp.setMergeFactor(5); writer = new IndexWriter(dir, newIndexWriterConfig(random, TEST_VERSION_CURRENT, new MockAnalyzer()).setOpenMode( OpenMode.APPEND).setMaxBufferedDocs(10).setMergePolicy(ldmp).setMergeScheduler(new ConcurrentMergeScheduler())); // merge factor is changed, so check invariants after all adds for (int i = 0; i < 10; i++) { addDoc(writer); } writer.commit(); ((ConcurrentMergeScheduler) writer.getConfig().getMergeScheduler()).sync(); writer.commit(); checkInvariants(writer); assertEquals(10, writer.maxDoc()); writer.close(); dir.close(); } private void addDoc(IndexWriter writer) throws IOException { Document doc = new Document(); doc.add(new Field("content", "aaa", Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(doc); } private void checkInvariants(IndexWriter writer) throws IOException { _TestUtil.syncConcurrentMerges(writer); int maxBufferedDocs = writer.getConfig().getMaxBufferedDocs(); int mergeFactor = ((LogMergePolicy) writer.getConfig().getMergePolicy()).getMergeFactor(); int maxMergeDocs = ((LogMergePolicy) writer.getConfig().getMergePolicy()).getMaxMergeDocs(); int ramSegmentCount = writer.getNumBufferedDocuments(); assertTrue(ramSegmentCount < maxBufferedDocs); int lowerBound = -1; int upperBound = maxBufferedDocs; int numSegments = 0; int segmentCount = writer.getSegmentCount(); for (int i = segmentCount - 1; i >= 0; i--) { int docCount = writer.getDocCount(i); assertTrue(docCount > lowerBound); if (docCount <= upperBound) { numSegments++; } else { if (upperBound * mergeFactor <= maxMergeDocs) { assertTrue("maxMergeDocs=" + maxMergeDocs + "; numSegments=" + numSegments + "; upperBound=" + upperBound + "; mergeFactor=" + mergeFactor + "; segs=" + writer.segString(), numSegments < mergeFactor); } do { lowerBound = upperBound; upperBound *= mergeFactor; } while (docCount > upperBound); numSegments = 1; } } if (upperBound * mergeFactor <= maxMergeDocs) { assertTrue(numSegments < mergeFactor); } String[] files = writer.getDirectory().listAll(); int segmentCfsCount = 0; for (int i = 0; i < files.length; i++) { if (files[i].endsWith(".cfs")) { segmentCfsCount++; } } assertEquals(segmentCount, segmentCfsCount); } /* private void printSegmentDocCounts(IndexWriter writer) { int segmentCount = writer.getSegmentCount(); System.out.println("" + segmentCount + " segments total"); for (int i = 0; i < segmentCount; i++) { System.out.println(" segment " + i + " has " + writer.getDocCount(i) + " docs"); } } */ }