Skip to content

samiran-kawtikwar/gpu-algorithms-labs

 
 

Repository files navigation

ECE/CS 508 Fall 2021

Welcome to ECE/CS 508!

This page contains information about how to access and submit the labs.

Install and Setup

Clone this repository to get the project folder.

git clone https://github.com/illinois-impact/gpu-algorithms-labs.git

Download the RAI binary for your platform below. You will probably use it for development, and definitely use it for submission.

Linux and Darwin RAI binaries

In the documentation below, we refer to RAI as the RAI binary. Either rename your binary to rai or substitute the downloaded name when executing. When executing, you will need to include the full path to RAI or add RAI's location to your $PATH.

Finally, the downloaded file will not have execute privileges. Use the 'chmod ugo+x rai' command to make the file executable.

You should have received a .rai_profile file by email. Put that file in ~/.rai_profile. Your .rai_profile should look something like this (indented with tabs!)

profile:
    firstname: <your-given-name>
    lastname: <your-surname>
    username: <your-username>
    email: <your-institution-email>
    access_key: <your-access-key>
    secret_key: <your-secret-key>
    affiliation: <your-affiliation>

To test the configurations, execute RAI with the device_query lab. For example, if you are in the parent directory of this repo, run:

rai -p ./labs/device_query

Windows


On Windows, you'll need to install WSL and a virtual linux OS. Several Linux versions are available through the Microsoft Store. Prof. Lumetta uses RAI on WSL (version 1)/Ubuntu.

Labs

Several labs will be assigned over the course of the semester

The device query lab (the first one) simply tests your RAI use; no changes should be necessary--if it doesn't work, you need to debug your setup.

For most labs, the main code of is in main.cu, which is the file you will be editing. Helper code that's specific to the lab is in the helper.hpp file and the common code across the labs in the common folder. You are free to add/delete/rename files but you need to make the appropriate changes to the CMakeLists.txt file.

To run any lab you cd into that directory, cd labs/device_query for example, and run rai -p . . From a user's point a view when the client runs as if it was local.

Code Development Tools

Throughout the semester, you'll be developing the labs. The following information is common through all the labs and might be helpful while developing.

We will also take a closer look at using the recent NVIDIA profiling tools, and I will integrate that material into the labs as the semester progresses.

Timing the Code Sections

It might be useful to figure out the time of each code section to identify the bottleneck code. In common/utils.hpp a function called timer_start/timer_stop which allows you to get the current time at a high resolution. To measure the overhead of a function f(args...), the pattern to use is:

timer_start(/* msg */ "calling the f(args...) function");
f(args...);
timer_stop();

This will print the time as the code is running.

Utility Functions

We provide some helper utility functions in the common/utils.hpp file.

Verifying the Results

Each lab contains the code to compute the golden (true) solution of the lab. We use Catch2 to perform tests to verify the results are accurate within the error tollerance. You can read the Catch2 tutorial if you are interested in how this works.

Subsets of the test cases can be run by executing a subset of the tests. We recomend running the first lab with -h option to understand what you can perform, but the rough idea is if you want to run a specific section (say [inputSize:1024]) then you pass -c "[inputSize:1024]" to the lab.

NOTE: The labs are configured to abort on the first error (using the -a option in the rai_build.yml file). You may need to change this to show the full list of errors.

Checking Errors

To check and throw CUDA errors, use the THROW_IF_ERROR function. This throws an error when a CUDA error is detected which you can catch if you need special handling of the error.

THROW_IF_ERROR(cudaMalloc((void **)&deviceW, wByteCount));

Profiling

Profiling can be performed using nvprof. Place the following build commands in your rai-build.yml file

    - >-
      nvprof --cpu-profiling on --export-profile timeline.nvprof --
      ./mybinary -i input1,input2 -o output
    - >-
      nvprof --cpu-profiling on --export-profile analysis.nvprof --analysis-metrics --
      ./mybinary -i input1,input2 -o output

You could change the input and test datasets. This will output two files timeline.nvprof and analysis.nvprof which can be viewed using the nvvp tool (by performing a file>import). You will have to install the nvvp viewer on your machine to view these files.

NOTE: nvvp will only show performance metrics for GPU invocations, so it may not show any analysis when you only have serial code.

You will need to install the nvprof viewer for the CUDA website and the nvprof GUI can be run without CUDA on your machine.

Downloading Your Output Code

All data files that are in the /build folder are temporarily uploaded to a file storage server. This allows you to dump output files during execution and analyze them locally.

✱ The build folder has been uploaded to http://s3.amazonaws.com/files.rai-project.com/abc.tar.gz. The data will be present for only a short duration of time.

Enabling Debug builds

Within the rai_build.yml environment, run cmake -DCMAKE_BUILD_TYPE=Debug /src this will enable debugging mode. You may also want to pass the -g -pg -lineinfo option to the compiler.

Offline Development

You can use the docker image and or install CMake within a CUDA envrionment. Then run cmake [lab] and then make. We do not recommend using your own machine, and we will not be debugging your machine/installation setup.

Issues

Please use the Github issue manager to make any suggestions on the structure of the labs. These will be provided for future semesters. Questions and problems for this semester should be posted to the discussion forum for the class.

If submitting a bug report, be sure to include the outputs of

rai version

as well as the output of

rai buildtime

in your report. You can also invoke the rai command with verbose and debug outputs using

rai --verbose --debug

License

NCSA/UIUC © Abdul Dakkak

About

IMPACT GPU Algorithms Teaching Labs

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • C++ 98.2%
  • Cuda 1.1%
  • Other 0.7%