Skip to content

Latest commit

 

History

History
115 lines (75 loc) · 3.75 KB

README.md

File metadata and controls

115 lines (75 loc) · 3.75 KB

# DotOpt - Optimized matrix multiplication

Development environment configuration

  1. Update the container.
./scripts/update-all.sh
  1. Install the required dependencies.
./scripts/install-deps.sh
  1. Fetch third party libraries
./scripts/fetch-libs.sh

WARNING: Benchmarking requires SYS_CAP_ADMIN.

Project overview

Main library

The main library is implemented as a python module which computes matrix multiplication operations using NumPy types.

Testing

Tests are implemented in Python, and added to test/CMakeLists.txt so that they are executed when testing.

Benchmarking

Benchmarks are implemented in C/C++ using Google's benchmarking library.

Warning: CPU frequency scaling should be disabled during the benchmarks.

# To query the current CPU frequency scaler:
sudo cpupower frequency-info

# To force the CPU to run at full power:
sudo cpupower frequency-set --governor performance

# To restore the previous CPU fequency scaler profile:
sudo cpupower frequency-set --governor <previous governor>

Algorithm implementations

  1. Sequential algorithm.

OpenMP parallelism

  1. Parallel algorithm using OpenMP's loop parallelism.

  2. Parallel algorithm using OpenMP's task parallelism.

  3. Parallel algorithm with row/col tiling using OpenMP's loop parallelism.

  4. Parallel algorithm with row/col tiling using OpenMP's task parallelism.

  5. Parallel algorithm with 2d tiling using OpenMP's loop parallelism.

  6. Parallel algorithm with 2d tiling using OpenMP's task parallelism.

Our IMTS implementation

  1. Parallel algorithm using 1 level IMTS with random tile ordering.

  2. Parallel algorithm using 1 level IMTS with tile ordering.

  3. Parallel algorithm using 2 level ITMS with random tile ordering.

  4. Parallel algorithm using 2 level ITMS with tile ordering.

  5. Parallel algorithm using 3 level ITMS with random tile ordering.

  6. Parallel algorithm using 3 level ITMS with tile ordering.

Extras

  1. The best version to date with a hand-crafted matrix multiplication assembly code.
  2. The hand-crafted version with 4-level ITMS tiling, where the smallest tiles fit into the vector registers. In other words, a hand crafted tiling level for our hand-crafted matrix multiplication implementation.

GPU using Vulkan compute shaders

  1. (Not implemented) A GPU version where each thread computes an output.
  2. (Not implemented) A tiling GPU version where each thread computes a 4x4 tile (GLSL).
  3. (Not implemented) A tiling GPU version where each SM computes a tile using the shared memory as a scratchpad for inputs and/or outputs.

GPU using OpenCL (if we can make it work)

  1. (Not implemented) Same as #16.
  2. (Not implemented) Same as #17.
  3. (Not implemented) Same as #18.

External implementations (for reference when comparing)

  1. (Testing only) NumPy implementation.
  2. (Not used) TensorFlow implementation.
  3. (Not used) OpenBLAS (maybe?).
  4. (Not used) clBLAS (maybe?).
  5. (Not used) ArrayFire (OpenCL; maybe?).

Intelligent Multilevel Tiling Scheduler

The IMTS scheduler divides the workload into 2d tiles taking into account the sizes of the LLC. Those are then passed on to another tiler which further divides the workload into smaller tiles which fit into the next-to-last level cache. The same process is repeated until the first level cache is reached.

The order in which the tiles are computed is predefined to minify the number of transfers between the caches. For example, a cache will need both input tiles and the output. When switching tiles, only one of those will be loaded, and preferably one of the inputs, since the outputs will have to be transferred back to the next level.