/* * Hibernate, Relational Persistence for Idiomatic Java * * Copyright (c) 2010, Red Hat, Inc. and/or its affiliates or third-party contributors as * indicated by the @author tags or express copyright attribution * statements applied by the authors. All third-party contributions are * distributed under license by Red Hat, Inc. * * This copyrighted material is made available to anyone wishing to use, modify, * copy, or redistribute it subject to the terms and conditions of the GNU * Lesser General Public License, as published by the Free Software Foundation. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY * or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public License * for more details. * * You should have received a copy of the GNU Lesser General Public License * along with this distribution; if not, write to: * Free Software Foundation, Inc. * 51 Franklin Street, Fifth Floor * Boston, MA 02110-1301 USA */ package org.hibernate.search.test.performance.reader; import java.io.File; import java.io.IOException; import java.util.concurrent.CountDownLatch; import java.util.concurrent.Executors; import java.util.concurrent.ThreadPoolExecutor; import java.util.concurrent.TimeUnit; import org.apache.lucene.analysis.SimpleAnalyzer; import org.apache.lucene.analysis.StopAnalyzer; import org.apache.lucene.index.CorruptIndexException; import org.apache.lucene.index.IndexWriter; import org.apache.lucene.store.FSDirectory; import org.apache.lucene.store.LockObtainFailedException; import org.hibernate.search.Environment; import org.hibernate.search.test.SearchTestCase; import org.hibernate.search.util.impl.FileHelper; /** * @author Sanne Grinovero */ public abstract class ReaderPerformance extends SearchTestCase { //more iterations for more reliable measures: private static final int TOTAL_WORK_BATCHES = 10; //the next 3 define the kind of workload mix to test on: private static final int SEARCHERS_PER_BATCH = 10; private static final int UPDATES_PER_BATCH = 2; private static final int INSERTIONS_PER_BATCH = 1; private static final int WORKER_THREADS = 20; private static final int WARMUP_CYCLES = 6; public void setUp() throws Exception { File baseIndexDir = getBaseIndexDir(); baseIndexDir.mkdir(); File[] files = baseIndexDir.listFiles(); for ( File file : files ) { FileHelper.delete( file ); } super.setUp(); } private void buildBigIndex() throws InterruptedException, CorruptIndexException, LockObtainFailedException, IOException { System.out.println( "Going to create fake index..." ); FSDirectory directory = FSDirectory.open( new File( getBaseIndexDir(), Detective.class.getCanonicalName() ) ); IndexWriter.MaxFieldLength fieldLength = new IndexWriter.MaxFieldLength( IndexWriter.DEFAULT_MAX_FIELD_LENGTH ); IndexWriter iw = new IndexWriter( directory, new SimpleAnalyzer(), true, fieldLength ); IndexFillRunnable filler = new IndexFillRunnable( iw ); ThreadPoolExecutor executor = ( ThreadPoolExecutor ) Executors.newFixedThreadPool( WORKER_THREADS ); for ( int batch = 0; batch <= 5000000; batch++ ) { executor.execute( filler ); } executor.shutdown(); executor.awaitTermination( 600, TimeUnit.SECONDS ); iw.commit(); iw.optimize(); iw.close(); System.out.println( "Index created." ); } @SuppressWarnings("unchecked") protected Class<?>[] getAnnotatedClasses() { return new Class[] { Detective.class, Suspect.class }; } public void tearDown() throws Exception { super.tearDown(); FileHelper.delete( getBaseIndexDir() ); } protected void configure(org.hibernate.cfg.Configuration cfg) { super.configure( cfg ); cfg.setProperty( "hibernate.search.default.directory_provider", "filesystem" ); cfg.setProperty( "hibernate.search.default.indexBase", getBaseIndexDir().getAbsolutePath() ); cfg.setProperty( "hibernate.search.default.optimizer.transaction_limit.max", "10" ); // workaround too many open files cfg.setProperty( "hibernate.search.default." + Environment.EXCLUSIVE_INDEX_USE, "true" ); cfg.setProperty( Environment.ANALYZER_CLASS, StopAnalyzer.class.getName() ); cfg.setProperty( Environment.READER_STRATEGY, getReaderStrategyName() ); } protected abstract String getReaderStrategyName(); public final void testPerformance() throws InterruptedException, CorruptIndexException, LockObtainFailedException, IOException { buildBigIndex(); for ( int i = 0; i < WARMUP_CYCLES; i++ ) { timeMs(); } } private final void timeMs() throws InterruptedException { ThreadPoolExecutor executor = ( ThreadPoolExecutor ) Executors.newFixedThreadPool( WORKER_THREADS ); CountDownLatch startSignal = new CountDownLatch( 1 ); InsertActivity insertionTask = new InsertActivity( getSessions(), startSignal ); SearchActivity searchTask = new SearchActivity( getSessions(), startSignal ); UpdateActivity updateTask = new UpdateActivity( getSessions(), startSignal ); //we declare needed activities in order, scheduler will "mix": for ( int batch = 0; batch <= TOTAL_WORK_BATCHES; batch++ ) { for ( int inserters = 0; inserters < INSERTIONS_PER_BATCH; inserters++ ) { executor.execute( insertionTask ); } for ( int searchers = 0; searchers < SEARCHERS_PER_BATCH; searchers++ ) { executor.execute( searchTask ); } for ( int updaters = 0; updaters < UPDATES_PER_BATCH; updaters++ ) { executor.execute( updateTask ); } } executor.shutdown(); long startTime = System.nanoTime(); startSignal.countDown();//start! executor.awaitTermination( 600, TimeUnit.SECONDS ); long endTime = System.nanoTime(); System.out.println( "Performance test for " + getReaderStrategyName() + ": " + TimeUnit.NANOSECONDS.toMillis( endTime - startTime ) + "ms. (" + ( TOTAL_WORK_BATCHES * SEARCHERS_PER_BATCH ) + " searches, " + ( TOTAL_WORK_BATCHES * INSERTIONS_PER_BATCH ) + " insertions, " + ( TOTAL_WORK_BATCHES * UPDATES_PER_BATCH ) + " updates)" ); } }