Loading...
Searching...
No Matches
Benchmark Taskflow

Compile and Run Benchmarks

To build the benchmark code, enable the CMake option TF_BUILD_BENCHMARKS to ON as follows:

# under /taskflow/build
~$ cmake ../ -DTF_BUILD_BENCHMARKS=ON
~$ make

After you successfully build the benchmark code, you can find all benchmark instances in the benchmarks/ folder. You can run the executable of each instance in the corresponding folder.

~$ cd benchmarks & ls
bench_black_scholes bench_binary_tree bench_graph_traversal ...
~$ ./bench_graph_traversal
|V|+|E| Runtime
2 0.197
842 0.198
3284 0.488
7288 0.774
... ...
... ...
619802 75.135
664771 77.436
711200 83.957

You can display the help message by giving the option --help.

~$ ./bench_graph_traversal --help
Graph Traversal
Usage: ./bench_graph_traversal [OPTIONS]
Options:
-h,--help Print this help message and exit
-t,--num_threads UINT number of threads (default=1)
-r,--num_rounds UINT number of rounds (default=1)
-m,--model TEXT model name tbb|omp|tf (default=tf)

We currently implement the following instances that are commonly used by the parallel computing community to evaluate the system performance.

Instance Description
bench_binary_tree traverses a complete binary tree
bench_black_scholes computes option pricing with Black-Shcoles Models
bench_graph_traversal traverses a randomly generated direct acyclic graph
bench_linear_chain traverses a linear chain of tasks
bench_mandelbrot exploits imbalanced workloads in a Mandelbrot set
bench_matrix_multiplication multiplies two 2D matrices
bench_mnist trains a neural network-based image classifier on the MNIST dataset
bench_parallel_sort sorts a range of items
bench_reduce_sum sums a range of items using reduction
bench_wavefront propagates computations in a 2D grid
bench_linear_pipeline performs pipeline parallelism on a linear chain of pipes
bench_graph_pipeline performs pipeline parallelism on a graph of pipes
bench_deferred_pipeline performs pipeline parallelism with dependencies from future pipes
bench_data_pipeline performs pipeline parallelisms on a cache-friendly data wrapper
bench_thread_pool uses our executor as a simple thread pool
bench_for_each performs parallel-iteration algorithms
bench_scan performs parallel-scan algorithms
bench_async_task creates asynchronous tasks
bench_fibonacci finds Fibonacci numbers using recursive asynchronous tasking
bench_nqueens parallelizes n-queen search using recursive asynchronous tasking
bench_integrate parallelizes integration using recursive asynchronous tasking
bench_primes finds a range of prime numbers using parallel-reduction algorithms
bench_skynet traverses a 10-ray tree using recursive asynchronous tasking

Configure Run Options

We implement consistent options for each benchmark instance. Common options are:

option value function
-h none displays the help message
-t integer configures the number of threads to run
-r integer configures the number of rounds to run
-m string configures the baseline models to run, tbb, omp, or tf

You can configure the benchmarking environment by giving different options.

Specify the Run Model

In addition to a Taskflow-based implementation for each benchmark instance, we have implemented two baseline models using the state-of-the-art parallel programming libraries, OpenMP and Intel TBB, to measure and evaluate the performance of Taskflow. You can select different implementations by passing the option -m.

~$ ./bench_graph_traversal -m tf # run the Taskflow implementation (default)
~$ ./bench_graph_traversal -m tbb # run the TBB implementation
~$ ./bench_graph_traversal -m omp # run the OpenMP implementation

Specify the Number of Threads

You can configure the number of threads to run a benchmark instance by passing the option -t. The default value is one.

# run the Taskflow implementation using 4 threads
~$ ./bench_graph_traversal -m tf -t 4

Depending on your environment, you may need to use taskset to set the CPU affinity of the running process. This allows the OS scheduler to keep process on the same CPU(s) as long as practical for performance reason.

# affine the process to 4 CPUs, CPU 0, CPU 1, CPU 2, and CPU 3
~$ taskset -c 0-3 bench_graph_traversal -t 4

Specify the Number of Rounds

Each benchmark instance evaluates the runtime of the implementation at different problem sizes. Each problem size corresponds to one iteration. You can configure the number of rounds per iteration to average the runtime.

# measure the %Taskflow runtime by averaging the results over 10 runs
~$ ./bench_graph_traversal -r 10 -m tf
|V|+|E| Runtime
2 0.109 # the runtime value 0.109 is an average of 10 runs
842 0.298
... ...
619802 73.135
664771 74.436