/* * Encog(tm) Core v3.4 - Java Version * http://www.heatonresearch.com/encog/ * https://github.com/encog/encog-java-core * Copyright 2008-2016 Heaton Research, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. * * For more information on Heaton Research copyrights, licenses * and trademarks visit: * http://www.heatonresearch.com/copyright */ package org.encog.neural.networks; import org.encog.engine.network.activation.ActivationSigmoid; import org.encog.mathutil.randomize.ConsistentRandomizer; import org.encog.mathutil.randomize.NguyenWidrowRandomizer; import org.encog.ml.MLError; import org.encog.ml.data.MLDataSet; import org.encog.ml.train.MLTrain; import org.encog.neural.freeform.FreeformLayer; import org.encog.neural.freeform.FreeformNetwork; import org.encog.neural.networks.layers.BasicLayer; import org.junit.Assert; import org.junit.Test; public class NetworkUtil { public static BasicNetwork createXORNetworkUntrained() { // random matrix data. However, it provides a constant starting point // for the unit tests. BasicNetwork network = new BasicNetwork(); network.addLayer(new BasicLayer(null,true,2)); network.addLayer(new BasicLayer(new ActivationSigmoid(),true,4)); network.addLayer(new BasicLayer(new ActivationSigmoid(),false,1)); network.getStructure().finalizeStructure(); (new ConsistentRandomizer(-1,1)).randomize(network); return network; } public static BasicNetwork createXORNetworknNguyenWidrowUntrained() { // random matrix data. However, it provides a constant starting point // for the unit tests. BasicNetwork network = new BasicNetwork(); network.addLayer(new BasicLayer(null,true,2)); network.addLayer(new BasicLayer(new ActivationSigmoid(),true,3)); network.addLayer(new BasicLayer(new ActivationSigmoid(),false,3)); network.addLayer(new BasicLayer(null,false,1)); network.getStructure().finalizeStructure(); (new NguyenWidrowRandomizer()).randomize( network ); return network; } public static void testTraining(MLDataSet dataSet, MLTrain train, double requiredImprove) { train.iteration(); double error1 = train.getError(); for(int i=0;i<10;i++) train.iteration(); double error2 = train.getError(); if( train.getMethod() instanceof MLError ) { double error3 = ((MLError)train.getMethod()).calculateError(dataSet); double improve = (error1-error3)/error1; Assert.assertTrue("Improve rate too low for " + train.getClass().getSimpleName() + ",Improve="+improve+",Needed="+requiredImprove, improve>=requiredImprove); } double improve = (error1-error2)/error1; Assert.assertTrue("Improve rate too low for " + train.getClass().getSimpleName() + ",Improve="+improve+",Needed="+requiredImprove, improve>=requiredImprove); } public static FreeformNetwork createXORFreeformNetworkUntrained() { FreeformNetwork network = new FreeformNetwork(); FreeformLayer inputLayer = network.createInputLayer(2); FreeformLayer hiddenLayer1 = network.createLayer(3); FreeformLayer outputLayer = network.createOutputLayer(1); network.connectLayers(inputLayer, hiddenLayer1, new ActivationSigmoid(), 1.0, false); network.connectLayers(hiddenLayer1, outputLayer, new ActivationSigmoid(), 1.0, false); network.reset(1000); return network; } }