1
0
mirror of https://github.com/gnss-sdr/gnss-sdr synced 2024-11-15 06:14:58 +00:00
gnss-sdr/tests/benchmarks
2024-10-06 08:14:46 +02:00
..
benchmark_atan2.cc Remove build and data folders, move tests and utils to the base of the source tree 2024-10-04 11:55:09 +02:00
benchmark_copy.cc Remove build and data folders, move tests and utils to the base of the source tree 2024-10-04 11:55:09 +02:00
benchmark_crypto.cc Remove build and data folders, move tests and utils to the base of the source tree 2024-10-04 11:55:09 +02:00
benchmark_detector.cc Fix warnings when building benchmarks 2024-10-06 08:14:46 +02:00
benchmark_osnma.cc Remove build and data folders, move tests and utils to the base of the source tree 2024-10-04 11:55:09 +02:00
benchmark_preamble.cc Remove build and data folders, move tests and utils to the base of the source tree 2024-10-04 11:55:09 +02:00
benchmark_reed_solomon.cc Remove build and data folders, move tests and utils to the base of the source tree 2024-10-04 11:55:09 +02:00
CMakeLists.txt Fix warnings when building benchmarks 2024-10-06 08:14:46 +02:00
README.md Remove build and data folders, move tests and utils to the base of the source tree 2024-10-04 11:55:09 +02:00

Benchmarks

This is a collection of implementation benchmarks based on Benchmark. It is useful to developers for assessing which is the fastest implementation of a given snippet of code in their machine, and to keep a track about them over different machines and compilers. The results may vary over different system architectures, compiler versions, building modes, C++ standard version used by the compiler, etc.

If Benchmark is not found in your system, CMake will download, build and statically link against that library for you at building time.

This collection is only built if the option ENABLE_BENCHMARKS is enabled when configuring gnss-sdr project's building:

$ cmake -DENABLE_BENCHMARKS=ON ..

Basic usage

Just execute the binaries generated in your install folder.

Example:

$ cd ../install
$ ./benchmark_copy

Output formats

The benchmarks support multiple output formats. Use the --benchmark_format=<console|json|csv> flag (or set the BENCHMARK_FORMAT=<console|json|csv> environment variable) to set the format type. console is the default format.

Write benchmark results to a file with the --benchmark_out=<filename> option.

Example:

$ ./benchmark_copy --benchmark_format=json --benchmark_out=benchmark_copy.json

Statistics

The number of runs of each benchmark is specified globally by the --benchmark_repetitions flag. When a benchmark is run more than once the mean, median and standard deviation of the runs will be reported.

Example:

$ ./benchmark_copy --benchmark_repetitions=10