Skip to content

Commit 08cda91

Browse files
committed
updated adam solver to v8 of paper
lambda parameter was redundant and has been removed.
1 parent b1c7c7f commit 08cda91

File tree

2 files changed

+5
-9
lines changed

2 files changed

+5
-9
lines changed

Diff for: demo/js/trainers.js

+1-1
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ nets = [];\n\
1515
trainer_defs = [];\n\
1616
trainer_defs.push({learning_rate:LR, method: 'sgd', momentum: 0.0, batch_size:BS, l2_decay:L2});\n\
1717
trainer_defs.push({learning_rate:LR, method: 'sgd', momentum: 0.9, batch_size:BS, l2_decay:L2});\n\
18-
trainer_defs.push({learning_rate:LR, method: 'adam', eps: 1e-8, beta1: 0.9, beta2: 0.99, lambda: 1-1e-8, batch_size:BS, l2_decay:L2});\n\
18+
trainer_defs.push({learning_rate:LR, method: 'adam', eps: 1e-8, beta1: 0.9, beta2: 0.99, batch_size:BS, l2_decay:L2});\n\
1919
trainer_defs.push({learning_rate:LR, method: 'adagrad', eps: 1e-6, batch_size:BS, l2_decay:L2});\n\
2020
trainer_defs.push({learning_rate:LR, method: 'windowgrad', eps: 1e-6, ro: 0.95, batch_size:BS, l2_decay:L2});\n\
2121
trainer_defs.push({learning_rate:1.0, method: 'adadelta', eps: 1e-6, ro:0.95, batch_size:BS, l2_decay:L2});\n\

Diff for: src/convnet_trainers.js

+4-8
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,6 @@
1818
this.eps = typeof options.eps !== 'undefined' ? options.eps : 1e-8; // used in adam or adadelta
1919
this.beta1 = typeof options.beta1 !== 'undefined' ? options.beta1 : 0.9; // used in adam
2020
this.beta2 = typeof options.beta2 !== 'undefined' ? options.beta2 : 0.999; // used in adam
21-
this.lambda = typeof options.lambda !== 'undefined' ? options.lambda : 1-1e-8; // used in adam
2221

2322
this.k = 0; // iteration counter
2423
this.gsum = []; // last iteration gradients (used for momentum calculations)
@@ -86,14 +85,11 @@
8685
var xsumi = this.xsum[i];
8786
if(this.method === 'adam') {
8887
// adam update
89-
var bt1 = this.beta1 * Math.pow(this.lambda, this.k-1); // decay first moment running average coefficient
90-
gsumi[j] = gsumi[j] * bt1 + (1-bt1) * gij; // update biased first moment estimate
88+
gsumi[j] = gsumi[j] * this.beta1 + (1- this.beta1) * gij; // update biased first moment estimate
9189
xsumi[j] = xsumi[j] * this.beta2 + (1-this.beta2) * gij * gij; // update biased second moment estimate
92-
var denom = Math.sqrt(xsumi[j]) + this.eps;
93-
var biasCorr1 = 1 - Math.pow(this.beta1, this.k); // correct bias
94-
var biasCorr2 = 1 - Math.pow(this.beta2, this.k); // correct bias
95-
var stepSize = this.learning_rate * Math.sqrt(biasCorr2) / biasCorr1;
96-
var dx = stepSize * gsumi[j] / denom;
90+
var biasCorr1 = gsumi[j] * (1 - Math.pow(this.beta1, this.k)); // correct bias first moment estimate
91+
var biasCorr2 = xsumi[j] * (1 - Math.pow(this.beta2, this.k)); // correct bias second moment estimate
92+
var dx = - this.learning_rate * biasCorr1 / (Math.sqrt(biasCorr2) + this.eps);
9793
p[j] += dx;
9894
} else if(this.method === 'adagrad') {
9995
// adagrad update

0 commit comments

Comments
 (0)