/*
* Encog(tm) Core v3.4 - Java Version
* http://www.heatonresearch.com/encog/
* https://github.com/encog/encog-java-core
* Copyright 2008-2016 Heaton Research, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*
* For more information on Heaton Research copyrights, licenses
* and trademarks visit:
* http://www.heatonresearch.com/copyright
*/
package org.encog.neural.networks.training.propagation.sgd.update;
import org.encog.neural.networks.training.propagation.sgd.StochasticGradientDescent;
/**
* Created by jeffh on 7/15/2016.
*/
public class AdaGradUpdate implements UpdateRule {
private StochasticGradientDescent training;
private double[] cache;
private double eps = 1e-8;
@Override
public void init(StochasticGradientDescent theTraining) {
this.training = theTraining;
this.cache = new double[theTraining.getFlat().getWeights().length];
}
@Override
public void update(double[] gradients, double[] weights) {
for(int i=0;i<weights.length;i++) {
this.cache[i] += gradients[i]*gradients[i];
final double delta = (this.training.getLearningRate()*gradients[i])/(Math.sqrt(cache[i])+this.eps);
weights[i] += delta;
}
}
public double getEps() {
return eps;
}
public void setEps(double eps) {
this.eps = eps;
}
}