29.01.2013 Views

Tutorial CUDA

Tutorial CUDA

Tutorial CUDA

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Execution Configuration: Heuristics<br />

(# of threads per block) = multiple of warp size<br />

© NVIDIA Corporation 2008<br />

To avoid wasting computation on under-populated warps<br />

(# of blocks) / (# of multiprocessors) > 1<br />

So all multiprocessors have at least a block to execute<br />

Per-block resources (shared memory and registers) at most<br />

half of total available<br />

And: (# of blocks) / (# of multiprocessors) > 2<br />

To get more than 1 active block per multiprocessor<br />

With multiple active blocks that aren’t all waiting at a<br />

__syncthreads(), the multiprocessor can stay busy<br />

(# of blocks) > 100 to scale to future devices<br />

Blocks stream through machine in pipeline fashion<br />

1000 blocks per grid will scale across multiple generations<br />

Very application-dependent: experiment!

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!