/** * Copyright 2014, Emory University * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package edu.emory.clir.clearnlp.bin.helper; import org.kohsuke.args4j.Option; import edu.emory.clir.clearnlp.classification.configuration.AbstractTrainerConfiguration; import edu.emory.clir.clearnlp.classification.configuration.AdaGradTrainerConfiguration; import edu.emory.clir.clearnlp.classification.model.AbstractModel; import edu.emory.clir.clearnlp.classification.model.SparseModel; import edu.emory.clir.clearnlp.classification.model.StringModel; import edu.emory.clir.clearnlp.classification.trainer.AbstractTrainer; import edu.emory.clir.clearnlp.classification.trainer.AdaGradLR; import edu.emory.clir.clearnlp.classification.trainer.AdaGradSVM; /** * @since 3.0.0 * @author Jinho D. Choi ({@code jinho.choi@emory.edu}) */ public class AdaGradClassify extends AbstractClassifyOnline { @Option(name="-a", usage="the learning rate (default: 0.1)", required=false, metaVar="<double>") private double d_alpha = 0.01; @Option(name="-r", usage="the tolerance of termination criterion (default: 0.1)", required=false, metaVar="<double>") private double d_rho = 0.1; @Option(name="-b", usage="the bias (default: 0.0)", required=false, metaVar="<double>") private double d_bias = 0d; @Option(name="-average", usage="if set, average weights (default: false)", required=false, metaVar="<boolean>") protected boolean b_average = false; @Option(name="-logistic", usage="if set, logistic regression (default: false)", required=false, metaVar="<boolean>") protected boolean b_logistic = false; public AdaGradClassify(String[] args) { super(args); } @Override protected AbstractTrainerConfiguration createTrainConfiguration() { return new AdaGradTrainerConfiguration(i_vectorType, b_binary, i_labelCutoff, i_featureCutoff, i_numberOfThreads, b_average, d_alpha, d_rho, d_bias); } @Override protected AbstractTrainer getTrainer(AbstractTrainerConfiguration trainConfiguration, AbstractModel<?, ?> model) { AdaGradTrainerConfiguration c = (AdaGradTrainerConfiguration)trainConfiguration; if (isSparseModel(model)) { if (b_logistic) return new AdaGradLR ((SparseModel)model, c.isAverage(), c.getLearningRate(), c.getRidge(), c.getBias()); else return new AdaGradSVM((SparseModel)model, c.isAverage(), c.getLearningRate(), c.getRidge(), c.getBias()); } else { if (b_logistic) return new AdaGradLR ((StringModel)model, c.getLabelCutoff(), c.getFeatureCutoff(), c.isAverage(), c.getLearningRate(), c.getRidge(), c.getBias()); else return new AdaGradSVM((StringModel)model, c.getLabelCutoff(), c.getFeatureCutoff(), c.isAverage(), c.getLearningRate(), c.getRidge(), c.getBias()); } } static public void main(String[] args) { new AdaGradClassify(args); } }