lucene-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Doron Cohen (JIRA)" <>
Subject [jira] Commented: (LUCENE-675) Lucene benchmark: objective performance test for Lucene
Date Wed, 15 Nov 2006 09:12:41 GMT
    [ ] 
Doron Cohen commented on LUCENE-675:

Would be nice to get some feedback on what I already have at this point for the "task based
benchmark framework for Lucene".  

So I am packing it as a zip file. I would probably resubmit as a patch when Grant commits
the current benchmark code.
See attached

To try out taskBenchmark, unzip under contrib/benchmark, on top of Grant's benchmark.patch.
This would do 3 changes:

1. replace build.xml - only change there is adding two targets: run-task-standard and run-task-micro-standard.

2. add 4 new files under conf:
 - task-standard.alg
 - task-micro-standard.alg

3. add a src package 'taskBenchmark' side by side with current 'benchmark' package.

To try it out, go to contrib/benchmark and try 'ant run-task-standard' or 'ant run-task-micro-standard'.

See inside the .alg files for how a test is specified.

The algorithm syntax and the entire package is documented in the package javadoc for taskBenchmark


> Lucene benchmark: objective performance test for Lucene
> -------------------------------------------------------
>                 Key: LUCENE-675
>                 URL:
>             Project: Lucene - Java
>          Issue Type: Improvement
>            Reporter: Andrzej Bialecki 
>         Assigned To: Grant Ingersoll
>         Attachments: benchmark.patch,, extract_reuters.plx,,,, tiny.alg,
> We need an objective way to measure the performance of Lucene, both indexing and querying,
on a known corpus. This issue is intended to collect comments and patches implementing a suite
of such benchmarking tests.
> Regarding the corpus: one of the widely used and freely available corpora is the original
Reuters collection, available from
or I
propose to use this corpus as a base for benchmarks. The benchmarking suite could automatically
retrieve it from known locations, and cache it locally.

This message is automatically generated by JIRA.
If you think it was sent incorrectly contact one of the administrators:
For more information on JIRA, see:


To unsubscribe, e-mail:
For additional commands, e-mail:

View raw message