Skip to content

OlivierSohn/gpgpu-experiments

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

What is it?

This is a project to experiment GPGPU using OpenCL.

Editing the main.cpp file will let you switch between the different examples.

The code of the first example is a slightly modified version of this excellent tutorial.

Why?

First, parallelizing algorithms is fun!

And in another project, I implemented a convolution reverb algorithm using ffts. Ffts are the bottleneck of the implementation, so I wanted to see if using the GPU to do the ffts would improve the performance, and by how much.

How?

As always, I try to tackle "one small problem at a time":

First, I started just with the tutorial, and verified I could make it run on my machine.

Then, I iteratively complexified the tutorial, adding one little new aspect at a time, before doing the fft (cooley-tukey algorithm without bitreversal of the input).

At every iteration, I was checking that the kernel code was behaving as intended by comparing the results with an equivalent cpu-based implementation.

And using the environment variable CL_LOG_ERRORS=stdout made debugging kernel compilation errors a lot easier!

Next Steps

  • experiment changing the number of items in the workgroup (compensate with the numner of local butterflies): is it best to have a lot of items or a lot of local butterflies? Should we auto-tune that?
  • experiment changing the radix, auto-tune that.
  • use images to have faster access to global memory:
    • To have faster read only access to inputs, use an image + float4 read_imagef
    • To have fater write to output, use an image + write_imagef
    • read/write images are opencl 2.0 only, but in practice passing the image twice with different qualifiers can work, depending on the driver + hardware.
  • use images for twiddles, see if it is faster than computing them on the fly (especially for high precision, and double).
  • Alternate global memory reads with computations for the first level to hide the compute time in the memory latency.
  • use the idea in https://mc.stanford.edu/cgi-bin/images/7/75/SC08_FFT_on_GPUs.pdf where private memory is used
  • instead of doing all levels in a single kernel, try doing one kernel per level, and use images to store intermediate results. The code will be more optimal because more stuff will be precomputed, and possibly less registers will be used.
  • Try stockham for big ffts.
  • Compare with other (open source) fft implementations on the gpu (for example, https://github.com/clMathLibraries/clFFT)
  • Implement in-place fft.

Platforms

Using CMake you can build and run it on a recent OSX.

Other platforms are not supported, but I think it's just a matter of making the CMakeLists.txt more general regarding the way to link to the OpenCL library.

Contributions

PRs are welcome, for example to generalize the CMakeLists.txt file to make it build and run on Linux or Windows.

About

experiments with OpenCL to do GPGPU.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages