Skip to content

misev/benchy

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

24 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

benchy

This is a generic and flexible tool for setting up benchmarks. The benchmark is self-contained in a single directory given as a parameter:

./benchy.sh -d <benchmark_suite_path> [OPTION]...

The tool uses /usr/bin/time for instrumentation, and is probably not extremely precise. It is mainly meant for quickly setting up system/macro benchmarks. The main principles guiding its design are simplicity and ease of use. For now at least, benchy is a single script that depends on very standard Linux stuff: Bash and awk. So installation is extremely straightforward (just download benchy.sh).

$ ./benchy.sh -h

Usage: benchy.sh -d <benchmark_suite_path> [OPTION]...

A generic and flexible tool for setting up macro/system benchmarks.

Options:
  -d, --suite-path
    directory containing the benchmark suite to be evaluated
  -r, --results-path
    benchmark results will be saved in a timestamped subdirectory of the
    directory specified with this argument; by default this is /tmp
  --init
    initialize new benchmark suite; the path specified with -d must be an
    empty directory in this case
  --quiet
    do not print any messages, except for errors
  -h, --help
    display this help and exit

Nomenclature

  • Benchmark suite: a collection of benchmark groups
  • Benchmark group: a collection of benchmarks
  • Benchmark: a single benchmark that is evaluated and for which execution time and other statistics are collected.

Directory structure

The benchmark directory (suite below) should have a particular structure, e.g:

suite/
|-- group_1/
|   |-- benchmark_1.query
|   |-- benchmark_2.sh
|   |-- benchmark_3.jar
|   |-- benchmark.conf
|-- group_2/
|-- group_3/
|-- benchmark.conf
  • directories (group_1, group_2, etc.) are equivalent to benchmark groups, so all related stuff that will be benchmarked should be put in the same directory. E.g, benchmarking 10 queries on growing data size could be put in one group.
  • within a benchmark group we have files that can be executed as benchmarks or as supplementary stuff before or after a benchmark (e.g. load data, prepare system, etc.)
  • benchmark.conf files are Bash scripts that customize how the benchmark is executed. There can be a benchmark.conf on suite level, as well as more specifc ones per group; in either case they are optional.

There is no restriction on how any file/directory is named with the exception of benchmark.conf.

benchmark.conf

benchy.sh provides just the functionality for timing, monitoring, averaging results, and aggregating them into a report. How the stuff that should be benchmarked is actually executed is specified by benchmark.conf configuration files.

benchmark.conf is a Bash script that can define particular variables and functions. benchy.sh first loads the benchmark suite configuration, and then as it goes through each group it loads its configuration as well if it exists, allowing to override the suite-wide configuration with more specific group configurations.

The variables and functions in benchmark.conf recognized in benchy.sh are documented below.

Variables

  • BENCHMARK_REPEAT: the number of times to repeat a benchmark; the more times it is evaluated, the more accurate will the measurements be. By default five repetitions are done.
  • BENCHMARK_RETRY: the number of times to retry a benchmark when it fails (run_benchmark return non-zero). By default three retries are attempted.

Functions

Benchmark suite

  • before_suite
    • args: suite name
    • description: executed before starting the benchmark suite evaluation
  • after_suite
    • args: suite name
    • description: executed when the benchmark suite evaluation is finished

Benchmark groups

  • is_group
    • args: group name
    • return: 0 if the given group should be benchmarked, non-zero otherwise
    • description: check whether a group should be benchmarked; if undefined, all groups are benchmarked.
  • before_group
    • args: group (directory) name
    • description: executed before a new group is evaluated
  • after_group
    • args: group name
    • description: executed when a group evaluation is finished

Single benchmark

  • is_benchmark
    • args: benchmark (file) name, group name
    • return: 0 if the given file should be benchmarked, non-zero otherwise
    • description: check whether a file should be benchmarked; if undefined, all files are considered benchmarks.
  • run_benchmark
    • args: benchmark (file) name, group name, repetition, group results directory
    • return: 0 if the benchmark executed successfully, non-zero otherwise
    • description: run the executable that needs to be benchmarked
  • run_non_benchmark
    • args: benchmark (file) name, group name
    • return: 0 if it executed successfully, non-zero otherwise
    • description: run non-benchmark file
  • before_benchmark
    • args: benchmark (file) name, group name
    • description: executed before starting a benchmark evaluation
  • after_benchmark
    • args: benchmark (file) name, group name
    • description: executed when a benchmark evaluation is finished
  • before_benchmark_repetition
    • args: benchmark (file) name, group name, repetition
    • description: executed before starting a benchmark repetition
  • after_benchmark_repetition
    • args: benchmark (file) name, group name, repetition
    • description: executed when a benchmark repetition is finished

About

A generic and flexible tool for setting up macro/system benchmarks.

Resources

License

Stars

Watchers

Forks

Packages

No packages published