All Questions

72
votes
3answers
66072 views

How do I choose grid and block dimensions for CUDA kernels?

This is a question about how to determine the CUDA grid, block and thread sizes. This is an additional question to the one posted here: https://stackoverflow.com/a/5643838/1292251 Following this lin...
70
votes
4answers
24253 views

How does OpenGL work at the lowest level?

I understand how to write OpenGL/DirectX programs, and I know the maths and the conceptual stuff behind it, but I'm curious how the GPU-CPU communication works on a low level. Say I've got an OpenGL ...
8
votes
1answers
29865 views

Modifying registry to increase GPU timeout, windows 7

Im trying to increase the timeout on the GPU from its default setting of 2 seconds to something a little longer. I found the following link but it appears its slightly different in windows 7 as i cant...
4
votes
1answers
4236 views

sending 3d array to CUDA kernel

I took the code given as an answer for How can I add up two 2d (pitched) arrays using nested for loops? and tried to use it for 3D instead of 2D and changed other parts slightly too, now it looks as f...
114
votes
14answers
57245 views

Utilizing the GPU with c#

I am trying to get more processing power out of my grid. I am using all cpus/cores, is it possible to utilize the GPU with C#. Anyone know any libraries or got any sample code?
87
votes
6answers
57301 views

GPU Emulator for CUDA programming without the hardware

Question: Is there an emulator for a Geforce card that would allow me to program and test CUDA without having the actual hardware? Info: I'm looking to speed up a few simulations of mine in CUDA, ...
16
votes
2answers
10569 views

How to measure the inner kernel time in NVIDIA CUDA?

I want to measure time inner kernel of GPU, how how to measure it in NVIDIA CUDA? e.g. __global__ void kernelSample() { some code here get start time some code here get stop time some co...
35
votes
3answers
29881 views

How to get current available GPUs in tensorflow?

I have a plan to use distributed TensorFlow, and I saw TensorFlow can use GPUs for training and testing. In a cluster environment, each machine could have 0 or 1 or more GPUs, and I want to run my Ten...
18
votes
1answers
12402 views

multi-GPU basic usage

How can I use two devices in order to improve for example the performance of the following code (sum of vectors)? Is it possible to use more devices "at the same time"? If yes, how can I manage the a...
16
votes
1answers
3268 views

AMD equivalent to NvOptimusEnablement

For Intel+NVIDIA dual-GPU "Optimus" setups, an application can export NvOptimusEnablement as explained in OptimusRenderingPolicies.pdf. This option allows an application to ensure the use of the high...

Previous Next