fgiraffe / ekclib

a simple library for classes, functions, and algorithms that I use in multiple projects.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

ekclib

a simple library for classes, functions, and algorithms that I use in multiple projects.

scheduler.h

This header contains threading code - it makes it easy to multi-thread certain workflows. Current design is based on CUDA, creates and passes thread_idx equivalent. Uses C++ threads, creating a pool equal to the number of CPUs.
To use, include it. Then derive your base class from worker. Override the do_work method. Create a scheduler object with a pointer to your class and the amount of work. run the class, then join which will complete when the work is done. Your do_work will be called with an integer, which tells it which block of work to do for that invocation. The scheduler class handles creating threads, etc. Your code just has to map that int to what needs to be done (ex rows in an image, files to be processes, etc).

Usage: to leverage multiple CPU cores as easy as possible. Use with non-uniform workloads, so that the work is balanced across multiple cores at runtime, rather than at compile time.

Example:

#include "scheduler.h"

struct raytrace : worker {
  void do_work(int work) {
    // do whatever work is needed for pixel row "work"
    // write it to _bitmap
  }
  Bitmap  _bitmap;
}

main() {
  raytrace ray;
  scheduler s(&ray, 1024);  // for a 1024 row image
  s.run();
  s.join();
  // ray._bitmap is now filled in, write it to disk, show it, etc
}

tests

test_scheduler1

This is a test for the scheduler and only does a wait as its workload. It is a performance test to show that the threading is working. It has 80 work-units that take 50 ms each. On an 8 core machine, it should take 0.5 seconds walltime, and 4.0 seconds walltime on a single thread. The results on my machine are as follows:

Simple scheduler test that performs uniform waits as 'work'.
---  Time using scheduler  ---
Wall Time = 0.543389
CPU Time  = 0.0035890

---  Time using single CPU core  ---
Wall Time = 4.02376
CPU Time  = 0.00142

test_scheduler2

This is a test for the scheduler and only does a wait as its workload. It is different from the first test in that the workloads are non-uniform, the timings are stored in the object in a vector - this represents workloads that take varying times. This one then compares the timings of scheduler, single core, and 8 threads that assign 1/8th the range to each thread. The timings show that scheduler is faster, even with its internal locks and bookkeeping, on non-uniform loads vs a fixed range of work assigned to a thread. The results on my machine are as follows:

Simple scheduler test that performs 8 non-uniform waits as 'work'.
---  Time using scheduler  ---
Wall Time = 0.537011
CPU Time  = 0.002407

---  Time using single CPU core  ---
Wall Time = 3.97234
CPU Time  = 0.001393

---  Time using 8 threads  ---
Wall Time = 0.571863
CPU Time  = 0.001889

test_scheduler3

This test simulates a main thread that is doing some work to generate the payloads. This would be like a JPEG decoder doing a Huffman decode in the main thread, and when each block of the image is decoded, then let the non-uniform thread scheduler handle the DCT and color conversion for each block (with a block == a workload).

---  standard scheduler (zero overhead)  ---
Wall Time = 0.344193
CPU Time  = 0.001445

---  1 millisecond to per workload to generate  ---
Wall Time = 0.354235
CPU Time  = 0.002291

---  10 millisecond to per workload to generate  ---
Wall Time = 0.554301
CPU Time  = 0.00307

---  100 millisecond to per workload to generate  ---
Wall Time = 4.2752
CPU Time  = 0.00309

We should see something like the above when testing. It shows that with 1 ms and baseline scheduler, there is almost no difference in timings, even though there is 1ms x 40 overhead; overhead is covered by the threads doing work. With the 10 ms overhead, this is masking the much of the threads work, because 10 x 40 ms = 0.400 seconds, then + 0.290 of thread work = .690 seconds, but we see less than that, implying that the masking is working. In the case of 100 ms overhead, it is much longer than any of the workloads, so it will be the same speed as if it were single threaded 40 x 100 ms = 4.0 seconds. With the _DEBUG numbers turned on, we can see the distribution of WORKLOAD per thread. Since the workloads are non-uniform, when there is little overhead, the thread calls are uneven. But as we increase the overhead, the thread calls become more uniform.

About

a simple library for classes, functions, and algorithms that I use in multiple projects.

License:MIT License


Languages

Language:C++ 91.4%Language:C 6.7%Language:Makefile 1.8%