diff --git a/examples/cnn_train.m b/examples/cnn_train.m index 6bbeabfc..ec53ebef 100644 --- a/examples/cnn_train.m +++ b/examples/cnn_train.m @@ -348,7 +348,7 @@ function [net,res] = accumulate_gradients(opts, lr, batchSize, net, res, mmap) % ------------------------------------------------------------------------- for l=numel(net.layers):-1:1 - for j=1:min(numel(res(l).dzdw),1) + for j=1:numel(res(l).dzdw) thisDecay = opts.weightDecay * net.layers{l}.weightDecay(j) ; thisLR = lr * net.layers{l}.learningRate(j) ;