Choose a graphic card to train SegNet for deep learning
1 view (last 30 days)
Show older comments
I just started to learn semantic segmentation using SegNet. I'm in the stage to choose a graphic card between GTX 1060 6gb or GTX 1060 3gb or GTX 1050 Ti 4gb. I'm very budget tight so I won't consider anything beyond GTX 1060 6gb. I need suggestion whether GTX 1060 6gb is sufficient to train SegNet? My data base is small and it won't exceed 1gb of images. Or can I start from GTX 1050 Ti 4gb? I don't mind about slower training speed (I can leave my computer on overnight) but I'm a bit worried if 4gb is enough?
Let's say can I use GTX 1050 Ti 4GB for this Matlab tutorial
https://au.mathworks.com/help/vision/examples/semantic-segmentation-using-deep-learning.html
Thanks.
4 Comments
Joss Knight
on 20 Nov 2017
To answer the question on precision, all Deep Learning is single precision. This is also true of most image processing, although there are some notable exceptions (for instance imgaussfilt which by default computes in double precision for accuracy reasons).
Answers (1)
Joss Knight
on 19 Nov 2017
The GeForce 10 series will all work fine, they just have different capabilities and constraints. On a cheaper GPU you may need to keep the training MiniBatchSize down in order to lower the stress on memory and compute power. I've seen quite a lot of problems with kernel timeouts and other issues posted to Answers on the 960, 970, 1050 and 1060, but I do think most of the problems can be dealt with that way.
0 Comments
See Also
Categories
Find more on Image Data Workflows in Help Center and File Exchange
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!