perl-Benchmark-Timer

Benchmarking with statistical confidence

The Benchmark::Timer class allows you to time portions of code conveniently, as well as benchmark code by allowing timings of repeated trials. It is perfect for when you need more precise information about the running time of portions of your code than the Benchmark module will give you, but don't want to go all out and profile your code. The methodology is simple; create a Benchmark::Timer object, and wrap portions of code that you want to benchmark with 'start()' and 'stop()' method calls. You can supply a tag to those methods if you plan to time multiple portions of code. If you provide error and confidence values, you can also use 'need_more_samples()' to determine, statistically, whether you need to collect more data. After you have run your code, you can obtain information about the running time by calling the 'results()' method, or get a descriptive benchmark report by calling 'report()'. If you run your code over multiple trials, the average time is reported. This is wonderful for benchmarking time-critical portions of code in a rigorous way. You can also optionally choose to skip any number of initial trials to cut down on initial case irregularities.

There is no official package available for openSUSE Leap 15.5

Distributions

openSUSE Tumbleweed

openSUSE Leap 15.6

openSUSE Leap 15.5

SUSE SLE-15-SP1

SUSE SLE-11 SP 4

Unsupported distributions

The following distributions are not officially supported. Use these packages at your own risk.