CUDA

Allocating Device Memory

cudaMalloc(LOCATION, SIZE)

  • LOCATION: Memory location on Device to allocate memory, an address in the GPU's memory

  • SIZE: number of bytes to allocate

De-Allocate: cudaFree()

Copy Data between Host and Device

cudaMemory(DST, SRC, NUM_BYTES, DIRECTION)

  • DST: An address of the memory to copy into

  • SRC: An address of the memory to copy from

  • NUM_BYTES: N * sizeof(type)

  • DIRECTION:

    • cudaMemcpyHostToDevice

    • cudaMemcpyDeviceToHost

Define the Kernel

__global__ void kernel(int *d_out, int *d_in) {
    d_out[0] = d_in[0];
}

Thread Index

In kernel definition, built-in variable threadIdx is accessible to get thread index within the thread block for each thread.

It has 3 dimensions: threadIdx.x, threadIdx.y and threadIdx.z.

Block Index

Index of a block: blockIdx.x, blockIdx.y and blockIdx.z.

Indexing within Grid

i = threadIdx.x + blockIdx.x * blockDim.x;

__syncthreads

To explicitly synchronize all threads (adding barriers), use __syncthreads.

int temp = a[i + 1];
__syncthreads;
a[i] = temp;
__syncthreads;

Launch the Kernel

int *h_c, *d_c;

// Allocate memeory on the device
cudaMalloc((void**)&d_c, sizeof(int));
// Copy content from host to device
cudaMemcpy(d_c, h_c, sizeof(int), cudaMemcpyHostToDevice);

// Launch the kernel
dim3 grid_size(1);
dim3 block_size(1);
kernel<<<grid_size, block_size>>>(...);

// Force host to wait on the completion of the kernel
cudaDeviceSynchronize();

// Copy data back to host
cudaMemcpy(h_c, d_c, sizeof(int), cudaMemcpyDeviceToHost);

// Clean up memory
cudaFree(d_c);
free(h_c);

Last updated