package org.apache.lucene.benchmark.byTask; /** * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ import java.io.File; import java.io.FileReader; import java.io.Reader; import org.apache.lucene.benchmark.byTask.utils.Algorithm; import org.apache.lucene.benchmark.byTask.utils.Config; /** * Run the benchmark algorithm. * <p>Usage: java Benchmark algorithm-file * <ol> * <li>Read algorithm.</li> * <li> Run the algorithm.</li> * </ol> * Things to be added/fixed in "Benchmarking by tasks": * <ol> * <li>TODO - report into Excel and/or graphed view.</li> * <li>TODO - perf comparison between Lucene releases over the years.</li> * <li>TODO - perf report adequate to include in Lucene nightly build site? (so we can easily track performance changes.)</li> * <li>TODO - add overall time control for repeated execution (vs. current by-count only).</li> * <li>TODO - query maker that is based on index statistics.</li> * </ol> */ public class Benchmark { private PerfRunData runData; private Algorithm algorithm; private boolean executed; public Benchmark (Reader algReader) throws Exception { // prepare run data try { runData = new PerfRunData(new Config(algReader)); } catch (Exception e) { e.printStackTrace(); throw new Exception("Error: cannot init PerfRunData!",e); } // parse algorithm try { algorithm = new Algorithm(runData); } catch (Exception e) { throw new Exception("Error: cannot understand algorithm!",e); } } public synchronized void execute() throws Exception { if (executed) { throw new IllegalStateException("Benchmark was already executed"); } executed = true; runData.setStartTimeMillis(); algorithm.execute(); } /** * Run the benchmark algorithm. * @param args benchmark config and algorithm files */ public static void main(String[] args) { // verify command line args if (args.length < 1) { System.err.println("Usage: java Benchmark <algorithm file>"); System.exit(1); } // verify input files File algFile = new File(args[0]); if (!algFile.exists() || !algFile.isFile() || !algFile.canRead()) { System.err.println("cannot find/read algorithm file: "+algFile.getAbsolutePath()); System.exit(1); } System.out.println("Running algorithm from: "+algFile.getAbsolutePath()); Benchmark benchmark = null; try { benchmark = new Benchmark(new FileReader(algFile)); } catch (Exception e) { e.printStackTrace(); System.exit(1); } System.out.println("------------> algorithm:"); System.out.println(benchmark.getAlgorithm().toString()); // execute try { benchmark.execute(); } catch (Exception e) { System.err.println("Error: cannot execute the algorithm! "+e.getMessage()); e.printStackTrace(); } System.out.println("####################"); System.out.println("### D O N E !!! ###"); System.out.println("####################"); } /** * @return Returns the algorithm. */ public Algorithm getAlgorithm() { return algorithm; } /** * @return Returns the runData. */ public PerfRunData getRunData() { return runData; } }