CMSC714 - High Performance Computing (Spring 2025)

High Performance Computing (CMSC714)

Project 3: CUDA

Due: March 24, 2025 @ 7:00 PM

The purpose of this programming assignment is to gain experience in writing GPU kernels. For this project you will apply image effects to videos.

Serial Algorithm

You will implement convolutions with image kernels on a video file. Image kernels are used for blurring, edge detection, and other effects. An image kernel is a square matrix which is applied to each pixel in an image. It is applied by setting the pixel to the dot product of its neighbors and the image kernel. Wikipedia and this visualization contain further reading.

For example, if k is a 3x3 kernel, then pixel i, j of img would be computed as


        new_img[i][j] = img[i-1][j-1]*k[0][0] + img[i-1][j]*k[0][1] + img[i-1][j+1]*k[0][2] +
                        img[i][j-1]*k[1][0] +   img[i][j]*k[1][1] +   img[i][j+1]*k[1][2] +
                        img[i+1][j-1]*k[2][0] + img[i+1][j]*k[2][1] + img[i+1][j+1]*k[2][2] 
        

The same pattern can be extended for kernels of size 5x5, 7x7, ... In this assignment you will apply a kernel to each pixel of each frame in a video. The pixels are in BGR format. You should convolve each color channel independently. Additionally, border pixels will be ignored.

GPU Algorithm

Your task is to implement this procedure on the GPU. We have provided starter code to handle reading/writing the video file (driver.cpp) and moving data on/off the GPU (video-effect.cu). You need to implement the convolveGPU kernel in video-effect.cu. A serial implementation of this algorithm is provided in the convolveCPU function. driver.cpp contains code to read/write the video and does not need to be edited.

The GPU kernel is structured as follows:


        __global__ void convolveGPU(float const* in, float *out, int width, int height, float const* kernel, int kernelWidth, int kernelHeight) {
                /* your code here */
        }
        

in is the video frame you are convolving. out is the frame you are storing results into. The frames are in packed BGR format (see the convolveCPU example for how to read each color channel). width and height give the dimensions of the video frame. kernel, kernelWidth, and kernelHeight provide the image kernel. You can assume the kernel is square with an odd dimension >= 3.

convolveGPU will be called with block size 8x8. You can change the block size with the blockDimSize variable. Your kernel should be able to handle when there are more pixels than threads available on the GPU (i.e. you need to implement striding).

Running

The starter code is setup to take an input video file name, output file name, and kernel name. The available kernel names are blur, edge, sharpen, and identity. You can also provide an optional frame number which will be dumped into a csv file. Dumped frames can be used to check for correctness. The command line arguments are:


        ./video-effect <input_filename> <output_filename> <kernel_name> <gridSizeX> <gridSizeY> <optional:frame_number> <optional:frame_filename>
        

You can check your output for this video (download the 1920x1080 version). We provide the output for frame 100 with the edge kernel here.

To get a GPU in your sbatch job add the following setting to the top


            #SBATCH --gres=gpu:a100_1g.5gb
          

A sample sbatch script can be found in submit.sh with the starter code. If you use interactive jobs, then add the --gres=gpu:a100_1g.5gb argument when you run salloc. For this code, you need to load cuda and opencv with module load cuda opencv.

Both the starter code and the sample output file are available on zaratan under /home/asussman/public/714/CUDA.

Output

The program will write out a video file with _out appended to the base filename. For instance ./video-effect video.mp4 blur will write out video_out.mp4 with the blur effect. You can copy videos from zaratan to your local machine to view them.

Running ./video-effect video.mp4 edge 100 frame.csv will write the 100th frame into frame.csv. Your output should match forest-edge-100.csv.

The program will output to stdout:


          Total time: 3.221 s
          Frames processed per second: 80.039 frames/s
        

What to Submit

You must submit the following files and no other files:

  • video-effect.cu: file with convolveGPU implemented
  • driver.cpp: file with changes to the command line arguments implemented
  • Makefile that will compile your code successfully on zaratan when using nvcc. A Makefile is provided with the starter code. Make sure that the executable name is video-effect, and do not include the executables in the tarball.
  • You must also submit a short report (LastName-FirstName-report.pdf) with performance results (one line plot). The line plots should present the execution times to run the GPU version respectively on the input file video.mp4 (for different block sizes). You can change the block size by changing the blockDimSize variable. In the report, you should include:
    • how you wrote the kernel
    • how was the initial data distribution done
    • what are the performance results, and are they what you expected
You should put the code, Makefile and report in a single directory (named LastName-assign3), compress it to .tar.gz (LastName-assign3.tar.gz) and upload that to ELMS.

Feel free to try different videos and image kernels. Keep in mind that some videos may drastically increase in file size if the kernel reduces the compression rate of the video data.

Grading

The project will be graded as follows:

Component Percentage
Runs correctly with the identity kernel 15
Runs correctly with other three kernels 75 (25% each kernel)
Writeup 10

  
Web Accessibility