I use Thrust in my current project, so I donβt need to write abstraction or (segmented) scans of the device_vector kernel.
So far, I have done all my work using traction abstractions, but for simple kernels or kernels that donβt easily transform into for_each or transform abstractions, I would rather write my own kernels at some point.
So my question is: can I ask through Thrust (or, possibly, CUDA) which device is currently used and what properties it has (maximum block size, maximum shared memory, all this)?
If I cannot get the current device, is there any way to get me to calculate the kernel sizes if I provide kernel registers and shared memory requirements?
source share