/*
* Encog(tm) Core v3.4 - Java Version
* http://www.heatonresearch.com/encog/
* https://github.com/encog/encog-java-core
* Copyright 2008-2016 Heaton Research, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*
* For more information on Heaton Research copyrights, licenses
* and trademarks visit:
* http://www.heatonresearch.com/copyright
*/
package org.encog.neural.networks.training.simple;
import org.encog.mathutil.error.ErrorCalculation;
import org.encog.ml.MLMethod;
import org.encog.ml.TrainingImplementationType;
import org.encog.ml.data.MLData;
import org.encog.ml.data.MLDataPair;
import org.encog.ml.data.MLDataSet;
import org.encog.ml.train.BasicTraining;
import org.encog.neural.NeuralNetworkError;
import org.encog.neural.networks.BasicNetwork;
import org.encog.neural.networks.training.LearningRate;
import org.encog.neural.networks.training.propagation.TrainingContinuation;
/**
* Train an ADALINE neural network.
*/
public class TrainAdaline extends BasicTraining implements LearningRate {
/**
* The network to train.
*/
private final BasicNetwork network;
/**
* The training data to use.
*/
private final MLDataSet training;
/**
* The learning rate.
*/
private double learningRate;
/**
* Construct an ADALINE trainer.
*
* @param network
* The network to train.
* @param training
* The training data.
* @param learningRate
* The learning rate.
*/
public TrainAdaline(final BasicNetwork network, final MLDataSet training,
final double learningRate) {
super(TrainingImplementationType.Iterative);
if (network.getLayerCount() > 2) {
throw new NeuralNetworkError(
"An ADALINE network only has two layers.");
}
this.network = network;
this.training = training;
this.learningRate = learningRate;
}
/**
* {@inheritDoc}
*/
@Override
public boolean canContinue() {
return false;
}
/**
* {@inheritDoc}
*/
@Override
public double getLearningRate() {
return this.learningRate;
}
/**
* {@inheritDoc}
*/
@Override
public MLMethod getMethod() {
return this.network;
}
/**
* {@inheritDoc}
*/
@Override
public void iteration() {
final ErrorCalculation errorCalculation = new ErrorCalculation();
for (final MLDataPair pair : this.training) {
// calculate the error
final MLData output = this.network.compute(pair.getInput());
for (int currentAdaline = 0; currentAdaline < output.size(); currentAdaline++) {
final double diff = pair.getIdeal().getData(currentAdaline)
- output.getData(currentAdaline);
// weights
for (int i = 0; i <= this.network.getInputCount(); i++) {
final double input;
if (i == this.network.getInputCount()) {
input = 1.0;
} else {
input = pair.getInput().getData(i);
}
this.network.addWeight(0, i, currentAdaline,
this.learningRate * diff * input);
}
}
errorCalculation.updateError(output.getData(), pair.getIdeal()
.getData(),pair.getSignificance());
}
// set the global error
setError(errorCalculation.calculate());
}
/**
* {@inheritDoc}
*/
@Override
public TrainingContinuation pause() {
return null;
}
/**
* {@inheritDoc}
*/
@Override
public void resume(final TrainingContinuation state) {
}
/**
* Set the learning rate.
*
* @param rate
* The new learning rate.
*/
@Override
public void setLearningRate(final double rate) {
this.learningRate = rate;
}
}