In CUDA C++ it's straightforward to define a shared memory of size specified at runtime. How can I do this with Numba/NumbaPro CUDA?
What I've done so far has only resulted in errors with the message
Argument 'shape' must be a constant
EDIT: Just to clarify, what I want is an equivalent to the following in CUDA C++ (example taken and adapted from here:
__global__ void dynamicReverse(int *d, int n)
{
extern __shared__ int s[];
// some work in the kernel with the shared memory
}
int main(void)
{
const int n = 64;
int a[n];
// run dynamic shared memory version
dynamicReverse<<<1,n,n*sizeof(int)>>>(a, n);
}