Problem with the "trainNetwork" function of neural network toolbox
Show older comments
I'm trying to do a transfer learning using AlexNet in version R2016a. But when my code ran to the "trainNetwork" function, I received these error messages:
Undefined function 'apply' for input arguments of type 'single'.
Error in trainNetwork>iComputeAverageImage (line 209)
X = apply(augmentations, X);
Error in trainNetwork>iInitializeNetworkNormalizations
(line 190)
avgI = iComputeAverageImage(data, augmentations);
Error in trainNetwork (line 78)
trainedNet = iInitializeNetworkNormalizations(trainedNet,
X, Y, opts, precision);
Error in trainingCompositionData (line 68)
convnetTransfer = trainNetwork(imds,transLayers,optionsTransfer);
I think the problem is not from my code, because the variable "augmentation" is an empty array and the variable "X" is a 4-D single matrix, which is specified by the "trainNetwork" itself.
Furthermore, I tried to find the code of "apply" function by open(F4) it and MATLAB redirected the page to the "ImageTransform" class, below is the "apply" function part of the class:
methods(Sealed)
%------------------------------------------------------------------
% Applies an array of transforms to a batch of RGB or grayscale
% images.
%------------------------------------------------------------------
function y = apply(this, batch)
y = batch;
for i = 1:numel(this)
y = doTransform(this(i), y);
end
end
end
I think this one is the correct code. But according to the error message, it seems that MATLAB doesn't use this function but other "apply" function in executing time.
Did anyone encounter this problem too?
3 Comments
Stefano Giorgi
on 8 Jan 2017
I had the same issue doing transfer learning from alexnet provided in the image catecory classification example on matlab website. I solved it re-creating the input image layer:
layersTransfer(1)=imageInputLayer([227 227 3],'normalization','zerocenter')
Note thet I re-created it exactly as it was before. I don't know the reason but it seems that input image normalization part won't succeed if you don't re-create the input layer in transfer learning. Not too bad since you don't loose anything re-creating it. After thet I was able to train it, too bad that on my poor 2 GB GPU I ran out of memory with batch size > 2, so GPU is unfeasible for me. This is a bit strange because the original Alexnet was trained with a batch of 128 on 2x3Gb GPU. Is it matlab implementation a bit too much memory hungry?
Joss Knight
on 12 Jan 2017
MATLAB is too memory hungry. This has been improved in a future release. 2 GB is still a bit restrictive for training AlexNet though, if the GPU is also driving the display.
Alam Noor
on 1 May 2018
Stefano Giorgi, can you please help for error of Error using classify (line 122) Requires at least three arguments.
Error in AlexNet_1 (line 12) label=classify(net,pic);
Answers (2)
Walter Roberson
on 6 Oct 2016
0 votes
Try training on double precision values not single precision.
flysuya
on 5 Dec 2016
0 votes
yes, I meet the same question. Do you find solution?
2 Comments
Walter Roberson
on 5 Dec 2016
What happened when you changed your data to double precision?
Roopam Gupta
on 5 Dec 2017
it does not change the error
Categories
Find more on Deep Learning Toolbox in Help Center and File Exchange
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!