/* * Encog(tm) Java Examples v3.4 * http://www.heatonresearch.com/encog/ * https://github.com/encog/encog-java-examples * * Copyright 2008-2016 Heaton Research, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. * * For more information on Heaton Research copyrights, licenses * and trademarks visit: * http://www.heatonresearch.com/copyright */ package org.encog.examples.neural.benchmark; import org.encog.Encog; import org.encog.ml.MLMethod; import org.encog.ml.data.MLDataSet; import org.encog.ml.data.basic.BasicMLDataSet; import org.encog.neural.networks.BasicNetwork; import org.encog.neural.networks.training.lma.LevenbergMarquardtTraining; import org.encog.util.simple.EncogUtility; /** * This example implements a Fahlman Encoder. Though probably not invented by Scott * Fahlman, such encoders were used in many of his papers, particularly: * * "An Empirical Study of Learning Speed in Backpropagation Networks" * (Fahlman,1988) * * It provides a very simple way of evaluating classification neural networks. * Basically, the input and output neurons are the same in count. However, * there is a smaller number of hidden neurons. This forces the neural * network to learn to encode the patterns from the input neurons to a * smaller vector size, only to be expanded again to the outputs. * * The training data is exactly the size of the input/output neuron count. * Each training element will have a single column set to 1 and all other * columns set to zero. You can also perform in "complement mode", where * the opposite is true. In "complement mode" all columns are set to 1, * except for one column that is 0. The data produced in "complement mode" * is more difficult to train. * * Fahlman used this simple training data to benchmark neural networks when * he introduced the Quickprop algorithm in the above paper. * */ public class FahlmanEncoder { public static final int INPUT_OUTPUT_COUNT = 10; public static final int HIDDEN_COUNT = 5; public static final int TRIES = 2500; public static final boolean COMPL = false; public static MLDataSet generateTraining(int inputCount, boolean compl) { double[][] input = new double[INPUT_OUTPUT_COUNT][INPUT_OUTPUT_COUNT]; double[][] ideal = new double[INPUT_OUTPUT_COUNT][INPUT_OUTPUT_COUNT]; for (int i = 0; i < inputCount; i++) { for (int j = 0; j < inputCount; j++) { if (compl) { input[i][j] = (j == i) ? 0.0 : 1.0; } else { input[i][j] = (j == i) ? 1.0 : 0.0; } ideal[i][j] = input[i][j]; } } return new BasicMLDataSet(input, ideal); } public static void main(String[] args) { MLDataSet trainingData = generateTraining(INPUT_OUTPUT_COUNT, COMPL); MLMethod method = EncogUtility.simpleFeedForward(INPUT_OUTPUT_COUNT, HIDDEN_COUNT, 0, INPUT_OUTPUT_COUNT, false); LevenbergMarquardtTraining train = new LevenbergMarquardtTraining((BasicNetwork) method, trainingData); EncogUtility.trainToError(train, 0.01); Encog.getInstance().shutdown(); } }