package com.antbrains.crf.hadoop; import gnu.trove.map.hash.TObjectIntHashMap; import java.io.FileInputStream; import java.io.FileNotFoundException; import java.io.FileOutputStream; import java.io.IOException; import java.util.ArrayList; import com.antbrains.crf.BESB1B2MTagConvertor; import com.antbrains.crf.FeatureDict; import com.antbrains.crf.TagConvertor; import com.antbrains.crf.TrainingWeights; import com.google.gson.Gson; import de.ruedigermoeller.serialization.FSTObjectInput; import de.ruedigermoeller.serialization.FSTObjectOutput; public class FeatureCountReader { public static void main(String[] args) throws Exception { if (args.length != 1) { System.out.println("Usage: FeatureCountReader <featuredict>"); System.exit(-1); } FSTObjectInput foi = null; FeatureDict fd = null; System.out.println(new java.util.Date() + " start load"); try { foi = new FSTObjectInput(new FileInputStream(args[0])); try { fd = (FeatureDict) foi.readObject(); System.out.println(fd.size()); } catch (ClassNotFoundException e) { throw new IOException(e); } } finally { if (foi != null) { foi.close(); } } } public static void addArray(ArrayList<String> lines, double[] array) { lines.add("" + array.length); for (double d : array) { lines.add("" + d); } } }