Thanks for using Compiler Explorer
Sponsors
Jakt
C++
Ada
Algol68
Analysis
Android Java
Android Kotlin
Assembly
C
C3
Carbon
C with Coccinelle
C++ with Coccinelle
C++ (Circle)
CIRCT
Clean
CMake
CMakeScript
COBOL
C++ for OpenCL
MLIR
Cppx
Cppx-Blue
Cppx-Gold
Cpp2-cppfront
Crystal
C#
CUDA C++
D
Dart
Elixir
Erlang
Fortran
F#
GLSL
Go
Haskell
HLSL
Hook
Hylo
IL
ispc
Java
Julia
Kotlin
LLVM IR
LLVM MIR
Modula-2
Mojo
Nim
Numba
Nix
Objective-C
Objective-C++
OCaml
Odin
OpenCL C
Pascal
Pony
PTX
Python
Racket
Raku
Ruby
Rust
Sail
Snowball
Scala
Slang
Solidity
Spice
SPIR-V
Swift
LLVM TableGen
Toit
Triton
TypeScript Native
V
Vala
Visual Basic
Vyper
WASM
Zig
Javascript
GIMPLE
Ygen
sway
cuda source #1
Output
Compile to binary object
Link to binary
Execute the code
Intel asm syntax
Demangle identifiers
Verbose demangling
Filters
Unused labels
Library functions
Directives
Comments
Horizontal whitespace
Debug intrinsics
Compiler
10.0.0 sm_75 CUDA-10.2
10.0.1 sm_75 CUDA-10.2
11.0.0 sm_75 CUDA-10.2
16.0.0 sm_90 CUDA-11.8
17.0.1(libc++) sm_90 CUDA-12.1
18.1.0(libc++) sm_90 CUDA-12.3.1
19.1.0 sm_90 CUDA-12.5.1
20.1.0 sm_90 CUDA-12.5.1
20.1.0 sm_90 CUDA-12.6.1
20.1.0 sm_90 CUDA-12.6.2
NVCC 10.0.130
NVCC 10.1.105
NVCC 10.1.168
NVCC 10.1.243
NVCC 10.2.89
NVCC 11.0.2
NVCC 11.0.3
NVCC 11.1.0
NVCC 11.1.1
NVCC 11.2.0
NVCC 11.2.1
NVCC 11.2.2
NVCC 11.3.0
NVCC 11.3.1
NVCC 11.4.0
NVCC 11.4.1
NVCC 11.4.2
NVCC 11.4.3
NVCC 11.4.4
NVCC 11.5.0
NVCC 11.5.1
NVCC 11.5.2
NVCC 11.6.0
NVCC 11.6.1
NVCC 11.6.2
NVCC 11.7.0
NVCC 11.7.1
NVCC 11.8.0
NVCC 12.0.0
NVCC 12.0.1
NVCC 12.1.0
NVCC 12.2.1
NVCC 12.3.1
NVCC 12.4.1
NVCC 12.5.1
NVCC 12.6.1
NVCC 12.6.2
NVCC 12.8.1
NVCC 12.9.0
NVCC 12.9.1
NVCC 13.0.0
NVCC 13.0.1
NVCC 9.1.85
NVCC 9.2.88
NVRTC 11.0.2
NVRTC 11.0.3
NVRTC 11.1.0
NVRTC 11.1.1
NVRTC 11.2.0
NVRTC 11.2.1
NVRTC 11.2.2
NVRTC 11.3.0
NVRTC 11.3.1
NVRTC 11.4.0
NVRTC 11.4.1
NVRTC 11.5.0
NVRTC 11.5.1
NVRTC 11.5.2
NVRTC 11.6.0
NVRTC 11.6.1
NVRTC 11.6.2
NVRTC 11.7.0
NVRTC 11.7.1
NVRTC 11.8.0
NVRTC 12.0.0
NVRTC 12.0.1
NVRTC 12.1.0
NVRTC 12.2.1
NVRTC 12.3.1
NVRTC 12.4.1
NVRTC 12.5.1
NVRTC 12.6.1
NVRTC 12.6.2
NVRTC 12.8.1
NVRTC 12.9.0
NVRTC 12.9.1
NVRTC 13.0.0
NVRTC 13.0.1
clang 7.0.0 sm_70 CUDA-9.1
clang 8.0.0 sm_75 CUDA-10.0
clang 9.0.0 sm_75 CUDA-10.1
clang rocm-4.5.2
clang rocm-5.0.2
clang rocm-5.1.3
clang rocm-5.2.3
clang rocm-5.3.2
clang rocm-5.7.0
clang rocm-6.0.2
clang rocm-6.1.2
clang rocm-6.2.4
clang rocm-6.3.3
clang rocm-6.4.0
clang staging rocm-6.1.2
clang staging rocm-6.2.4
clang staging rocm-6.3.3
clang staging rocm-6.4.0
clang trunk rocm-6.1.2
clang trunk rocm-6.2.4
clang trunk rocm-6.3.3
clang trunk rocm-6.4.0
trunk sm_100a CUDA-12.8.1
Options
Source code
#include <cstdio> #include <cuda_runtime.h> #ifndef CUDA_CHECK #define CUDA_CHECK(x) do { cudaError_t e = (x); if (e != cudaSuccess) { \ fprintf(stderr, "CUDA error %s:%d: %s\n", __FILE__, __LINE__, cudaGetErrorString(e)); \ std::exit(1); } } while (0) #endif __global__ void strided_read_kernel(const float* __restrict__ in, float* __restrict__ out, size_t N, int stride) { const size_t t = blockIdx.x * blockDim.x + threadIdx.x; const size_t T = gridDim.x * (size_t)blockDim.x; float acc = 0.f; for (size_t j = (size_t)t * (size_t)stride; j < N; j += (size_t)T * (size_t)stride) { // across a warp, addresses differ by (stride * sizeof(float)) float v = in[j]; // perfectly coalesced for stride == 1 acc = acc * 1.000000119f + v; // keep compiler from optimizing out loads } // do one write per thread (no stride so coalesced, negligible vs reads) if (t < N) out[t] = acc; } static float run_case(const float* d_in, float* d_out, size_t N, int stride, int iters, dim3 grid, dim3 block) { // do a warmup run strided_read_kernel<<<grid, block>>>(d_in, d_out, N, stride); CUDA_CHECK(cudaGetLastError()); CUDA_CHECK(cudaDeviceSynchronize()); // use cudaEvents for timing cudaEvent_t start, stop; CUDA_CHECK(cudaEventCreate(&start)); CUDA_CHECK(cudaEventCreate(&stop)); // run benchmark CUDA_CHECK(cudaEventRecord(start)); for (int i = 0; i < iters; ++i) { strided_read_kernel<<<grid, block>>>(d_in, d_out, N, stride); } CUDA_CHECK(cudaEventRecord(stop)); CUDA_CHECK(cudaEventSynchronize(stop)); float ms = 0.f; CUDA_CHECK(cudaEventElapsedTime(&ms, start, stop)); CUDA_CHECK(cudaEventDestroy(start)); CUDA_CHECK(cudaEventDestroy(stop)); // bytes read ~= (N/stride) floats per kernel launch const double bytes_read = (double)iters * (double)(N / (size_t)stride) * sizeof(float); const double sec = ms / 1e3; const double GBps = bytes_read / sec / 1e9; return (float)GBps; } int main(int argc, char** argv) { // needs to be big enough to exceed cache size const size_t N = (size_t)1 << 26; // 67,108,864 floats ~= 256 MB const int iters = 10; // choose sensible launch parameters const int block_size = 256; int dev = 0; cudaDeviceProp prop{}; CUDA_CHECK(cudaGetDevice(&dev)); CUDA_CHECK(cudaGetDeviceProperties(&prop, dev)); const int blocks_per_sm = 8; const int grid_size = blocks_per_sm * prop.multiProcessorCount; float *d_in = nullptr, *d_out = nullptr; CUDA_CHECK(cudaMalloc(&d_in, N * sizeof(float))); CUDA_CHECK(cudaMalloc(&d_out, N * sizeof(float))); // initialize input CUDA_CHECK(cudaMemset(d_in, 0, N * sizeof(float))); CUDA_CHECK(cudaMemset(d_out, 0, N * sizeof(float))); printf("# Device: %s (SM %d)\n", prop.name, prop.major * 10 + prop.minor); printf("# N = %zu floats (%.1f MB), iters = %d\n", N, N * sizeof(float) / (1024.0*1024.0), iters); printf("%6s %10s\n", "stride", "GB/s"); // run multiple strides const int strides[] = {1,2,4,8,16,32,64,128,256,512,1024,2048}; for (int s : strides) { // (N รท stride) must be an integer if ((N % (size_t)s) != 0) continue; float gbps = run_case(d_in, d_out, N, s, iters, dim3(grid_size), dim3(block_size)); printf("%6d %10.1f\n", s, gbps); } CUDA_CHECK(cudaFree(d_out)); CUDA_CHECK(cudaFree(d_in)); return 0; }
Become a Patron
Sponsor on GitHub
Donate via PayPal
Compiler Explorer Shop
Source on GitHub
Mailing list
Installed libraries
Wiki
Report an issue
How it works
Contact the author
CE on Mastodon
CE on Bluesky
Statistics
Changelog
Version tree