What happen to the CUDA cache mem?
3 views (last 30 days)
Show older comments
Hello there. I am a newbie with the GPU computing with Matlab, so apologize if the question sounds silly. I am trying to optimise some computation I am doing with the GPU. I believe it is well configured. I am doing some testing to understand how the GPU reacts to different commands and choose the best programming strategy. I have incurred in the following thing. I would appreciate some elucidation about the mechanism by which this feature happens. I am running a Geforce GTX 1080 ti. I do the following:
A = rand([100 100 100 100 10],'single','gpuArray')
tic,permute(A,[3 2 1 5 4]),wait(M.SelectedDevice()),toc
(trying to see how long does it take to permute a matrix)
now, if I ask the parallel.gpu.GPUDeviceManager.instance.SelectedDevice().AvailableMemory (read the available memory), then I can run a permute again. However, if I run two consecutive permute, I get the following
Error using gpuArray/permute Out of memory on device. To view more detail about available memory on the GPU, use 'gpuDevice()'. If the problem persists, reset the GPU by calling 'gpuDevice(1)'.
WHY?
2 Comments
Answers (2)
Joss Knight
on 19 Dec 2017
The result is stored as the variable ans, which means you have less memory the second time round.
4 Comments
giovanni esposito
on 18 Jul 2018
Edited: giovanni esposito
on 18 Jul 2018
hence, for example this code shall free all gpus memory ad the end of each loop, correct ? I try to do this but memory is still busy at the end of each loop.
clear all
RefreshGPU = 100;
NW = gpuDeviceCount;
nw = 1:NW;
poolobj = gcp('nocreate'); % If no pool, do not create new one.
if isempty(poolobj)
ParObj = parpool('local',NW);
else
delete(gcp);
ParObj = parpool('local',NW);
end
a=rand(NW,1e5);
Nloop = 1e5;
for kk=1:Nloop
spmd
b = somefunction(a(labindex,:)); % this function do something on GPUs
end
clear b
end
Joss Knight
on 18 Jul 2018
No, you are calling clear b on the client. You need to do it inside the SPMD block.
Jeffrey Daniels
on 12 Mar 2018
FYI - For anyone else having similar problems, I get similar errors when I run too many workers. The GPU is being shared by each of the CPU workers and if you have too large or too many GPU matricies you will run out of memory on the GPU. One solution is to open the Cluster Profile Manager from the Parallel menu and reduce the number of workers in your Cluster Profile.
0 Comments
See Also
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!