Cuda learn (1)

[toc]

Basic Concepts

A GPU is built around an array of Streaming Multiprocessors (SMs) (see Hardware Implementation for more details). A multi-threaded program is partitioned into blocks of threads that execute independently from each other, so that a GPU with more multiprocessors will automatically execute the program in less time than a GPU with fewer multiprocessors.

here is your content.