/** * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.hadoop.mapred.lib.aggregate; import org.apache.hadoop.fs.*; import org.apache.hadoop.io.*; import org.apache.hadoop.mapred.*; import org.apache.hadoop.mapred.lib.*; import junit.framework.TestCase; import java.io.*; import java.util.*; import java.text.NumberFormat; public class TestAggregates extends TestCase { private static NumberFormat idFormat = NumberFormat.getInstance(); static { idFormat.setMinimumIntegerDigits(4); idFormat.setGroupingUsed(false); } public void testAggregates() throws Exception { launch(); } public static void launch() throws Exception { JobConf conf = new JobConf(TestAggregates.class); FileSystem fs = FileSystem.get(conf); int numOfInputLines = 20; Path OUTPUT_DIR = new Path("build/test/output_for_aggregates_test"); Path INPUT_DIR = new Path("build/test/input_for_aggregates_test"); String inputFile = "input.txt"; fs.delete(INPUT_DIR, true); fs.mkdirs(INPUT_DIR); fs.delete(OUTPUT_DIR, true); StringBuffer inputData = new StringBuffer(); StringBuffer expectedOutput = new StringBuffer(); expectedOutput.append("max\t19\n"); expectedOutput.append("min\t1\n"); FSDataOutputStream fileOut = fs.create(new Path(INPUT_DIR, inputFile)); for (int i = 1; i < numOfInputLines; i++) { expectedOutput.append("count_").append(idFormat.format(i)); expectedOutput.append("\t").append(i).append("\n"); inputData.append(idFormat.format(i)); for (int j = 1; j < i; j++) { inputData.append(" ").append(idFormat.format(i)); } inputData.append("\n"); } expectedOutput.append("value_as_string_max\t9\n"); expectedOutput.append("value_as_string_min\t1\n"); expectedOutput.append("uniq_count\t15\n"); fileOut.write(inputData.toString().getBytes("utf-8")); fileOut.close(); System.out.println("inputData:"); System.out.println(inputData.toString()); JobConf job = new JobConf(conf, TestAggregates.class); FileInputFormat.setInputPaths(job, INPUT_DIR); job.setInputFormat(TextInputFormat.class); FileOutputFormat.setOutputPath(job, OUTPUT_DIR); job.setOutputFormat(TextOutputFormat.class); job.setMapOutputKeyClass(Text.class); job.setMapOutputValueClass(Text.class); job.setOutputKeyClass(Text.class); job.setOutputValueClass(Text.class); job.setNumReduceTasks(1); job.setMapperClass(ValueAggregatorMapper.class); job.setReducerClass(ValueAggregatorReducer.class); job.setCombinerClass(ValueAggregatorCombiner.class); job.setInt("aggregator.descriptor.num", 1); job.set("aggregator.descriptor.0", "UserDefined,org.apache.hadoop.mapred.lib.aggregate.AggregatorTests"); job.setLong("aggregate.max.num.unique.values", 14); JobClient.runJob(job); // // Finally, we compare the reconstructed answer key with the // original one. Remember, we need to ignore zero-count items // in the original key. // boolean success = true; Path outPath = new Path(OUTPUT_DIR, "part-00000"); String outdata = TestMiniMRWithDFS.readOutput(outPath,job); System.out.println("full out data:"); System.out.println(outdata.toString()); outdata = outdata.substring(0, expectedOutput.toString().length()); assertEquals(expectedOutput.toString(),outdata); //fs.delete(OUTPUT_DIR); fs.delete(INPUT_DIR, true); } /** * Launches all the tasks in order. */ public static void main(String[] argv) throws Exception { launch(); } }