org.hibernate.search.test.performance.reader.ReaderPerformance.java Source code

Java tutorial

Introduction

Here is the source code for org.hibernate.search.test.performance.reader.ReaderPerformance.java

Source

/*
 * Hibernate, Relational Persistence for Idiomatic Java
 *
 *  Copyright (c) 2010, Red Hat, Inc. and/or its affiliates or third-party contributors as
 *  indicated by the @author tags or express copyright attribution
 *  statements applied by the authors.  All third-party contributions are
 *  distributed under license by Red Hat, Inc.
 *
 *  This copyrighted material is made available to anyone wishing to use, modify,
 *  copy, or redistribute it subject to the terms and conditions of the GNU
 *  Lesser General Public License, as published by the Free Software Foundation.
 *
 *  This program is distributed in the hope that it will be useful,
 *  but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
 *  or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU Lesser General Public License
 *  for more details.
 *
 *  You should have received a copy of the GNU Lesser General Public License
 *  along with this distribution; if not, write to:
 *  Free Software Foundation, Inc.
 *  51 Franklin Street, Fifth Floor
 *  Boston, MA  02110-1301  USA
 */
package org.hibernate.search.test.performance.reader;

import java.io.File;
import java.io.IOException;
import java.util.concurrent.CountDownLatch;
import java.util.concurrent.Executors;
import java.util.concurrent.ThreadPoolExecutor;
import java.util.concurrent.TimeUnit;

import org.apache.lucene.analysis.SimpleAnalyzer;
import org.apache.lucene.analysis.StopAnalyzer;
import org.apache.lucene.index.CorruptIndexException;
import org.apache.lucene.index.IndexWriter;
import org.apache.lucene.store.FSDirectory;
import org.apache.lucene.store.LockObtainFailedException;

import org.hibernate.search.Environment;
import org.hibernate.search.test.SearchTestCase;
import org.hibernate.search.util.impl.FileHelper;

/**
 * @author Sanne Grinovero
 */
public abstract class ReaderPerformance extends SearchTestCase {

    //more iterations for more reliable measures:
    private static final int TOTAL_WORK_BATCHES = 10;
    //the next 3 define the kind of workload mix to test on:
    private static final int SEARCHERS_PER_BATCH = 10;
    private static final int UPDATES_PER_BATCH = 2;
    private static final int INSERTIONS_PER_BATCH = 1;

    private static final int WORKER_THREADS = 20;

    private static final int WARMUP_CYCLES = 6;

    protected void setUp() throws Exception {
        File baseIndexDir = getBaseIndexDir();
        baseIndexDir.mkdir();
        File[] files = baseIndexDir.listFiles();
        for (File file : files) {
            FileHelper.delete(file);
        }
        super.setUp();
    }

    private void buildBigIndex()
            throws InterruptedException, CorruptIndexException, LockObtainFailedException, IOException {
        System.out.println("Going to create fake index...");
        FSDirectory directory = FSDirectory.open(new File(getBaseIndexDir(), Detective.class.getCanonicalName()));
        IndexWriter.MaxFieldLength fieldLength = new IndexWriter.MaxFieldLength(
                IndexWriter.DEFAULT_MAX_FIELD_LENGTH);
        IndexWriter iw = new IndexWriter(directory, new SimpleAnalyzer(), true, fieldLength);
        IndexFillRunnable filler = new IndexFillRunnable(iw);
        ThreadPoolExecutor executor = (ThreadPoolExecutor) Executors.newFixedThreadPool(WORKER_THREADS);
        for (int batch = 0; batch <= 5000000; batch++) {
            executor.execute(filler);
        }
        executor.shutdown();
        executor.awaitTermination(600, TimeUnit.SECONDS);
        iw.commit();
        iw.optimize();
        iw.close();
        System.out.println("Index created.");
    }

    @SuppressWarnings("unchecked")
    protected Class<?>[] getAnnotatedClasses() {
        return new Class[] { Detective.class, Suspect.class };
    }

    protected void tearDown() throws Exception {
        super.tearDown();
        FileHelper.delete(getBaseIndexDir());
    }

    protected void configure(org.hibernate.cfg.Configuration cfg) {
        super.configure(cfg);
        cfg.setProperty("hibernate.search.default.directory_provider", "filesystem");
        cfg.setProperty("hibernate.search.default.indexBase", getBaseIndexDir().getAbsolutePath());
        cfg.setProperty("hibernate.search.default.optimizer.transaction_limit.max", "10"); // workaround too many open files
        cfg.setProperty("hibernate.search.default." + Environment.EXCLUSIVE_INDEX_USE, "true");
        cfg.setProperty(Environment.ANALYZER_CLASS, StopAnalyzer.class.getName());
        cfg.setProperty(Environment.READER_STRATEGY, getReaderStrategyName());
    }

    protected abstract String getReaderStrategyName();

    public final void testPerformance()
            throws InterruptedException, CorruptIndexException, LockObtainFailedException, IOException {
        buildBigIndex();
        for (int i = 0; i < WARMUP_CYCLES; i++) {
            timeMs();
        }
    }

    private final void timeMs() throws InterruptedException {
        ThreadPoolExecutor executor = (ThreadPoolExecutor) Executors.newFixedThreadPool(WORKER_THREADS);
        CountDownLatch startSignal = new CountDownLatch(1);
        InsertActivity insertionTask = new InsertActivity(getSessions(), startSignal);
        SearchActivity searchTask = new SearchActivity(getSessions(), startSignal);
        UpdateActivity updateTask = new UpdateActivity(getSessions(), startSignal);
        //we declare needed activities in order, scheduler will "mix":
        for (int batch = 0; batch <= TOTAL_WORK_BATCHES; batch++) {
            for (int inserters = 0; inserters < INSERTIONS_PER_BATCH; inserters++) {
                executor.execute(insertionTask);
            }
            for (int searchers = 0; searchers < SEARCHERS_PER_BATCH; searchers++) {
                executor.execute(searchTask);
            }
            for (int updaters = 0; updaters < UPDATES_PER_BATCH; updaters++) {
                executor.execute(updateTask);
            }
        }
        executor.shutdown();
        long startTime = System.nanoTime();
        startSignal.countDown();//start!
        executor.awaitTermination(600, TimeUnit.SECONDS);
        long endTime = System.nanoTime();
        System.out.println("Performance test for " + getReaderStrategyName() + ": "
                + TimeUnit.NANOSECONDS.toMillis(endTime - startTime) + "ms. ("
                + (TOTAL_WORK_BATCHES * SEARCHERS_PER_BATCH) + " searches, "
                + (TOTAL_WORK_BATCHES * INSERTIONS_PER_BATCH) + " insertions, "
                + (TOTAL_WORK_BATCHES * UPDATES_PER_BATCH) + " updates)");
    }

}