/**
* Copyright 2012 - CommonCrawl Foundation
*
* This program is free software: you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation, either version 3 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program. If not, see <http://www.gnu.org/licenses/>.
*
**/
package org.commoncrawl.mapred.pipelineV3.domainmeta.fuzzydedupe;
import java.io.IOException;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.mapred.JobClient;
import org.apache.hadoop.mapred.JobConf;
import org.commoncrawl.crawl.common.internal.CrawlEnvironment;
import org.commoncrawl.mapred.ec2.postprocess.deduper.Stage2Reducer;
import org.commoncrawl.mapred.pipelineV3.CrawlPipelineStep;
import org.commoncrawl.mapred.pipelineV3.CrawlPipelineTask;
import org.commoncrawl.util.JobBuilder;
import org.commoncrawl.util.TextBytes;
import com.google.common.collect.ImmutableList;
/**
*
* @author rana
*
*/
public class FuzzyDedupeStep2 extends CrawlPipelineStep {
private static final Log LOG = LogFactory.getLog(FuzzyDedupeStep2.class);
public static final String OUTPUT_DIR_NAME = "dedupeStage2";
public FuzzyDedupeStep2(CrawlPipelineTask task) throws IOException {
super(task, "fuzzy deduper - stage 2", OUTPUT_DIR_NAME);
}
@Override
public Log getLogger() {
return LOG;
}
@Override
public void runStep(Path outputPathLocation) throws IOException {
LOG.info("Task Identity Path is:" + getTaskIdentityPath());
LOG.info("Temp Path is:" + outputPathLocation);
ImmutableList<Path> paths = new ImmutableList.Builder<Path>().add(
makeUniqueOutputDirPath(_task.getOutputDirForStep(FuzzyDedupeStep1.OUTPUT_DIR_NAME), getTaskIdentityId()))
.build();
JobConf job = new JobBuilder(getPipelineStepName(), getConf())
.inputIsSeqFile()
.inputs(paths)
.mapperKeyValue(TextBytes.class, TextBytes.class)
.outputKeyValue(TextBytes.class, TextBytes.class)
.reducer(Stage2Reducer.class,false)
.numReducers(CrawlEnvironment.NUM_DB_SHARDS / 2)
.outputIsSeqFile()
.output(outputPathLocation)
.build();
LOG.info("Running " + getDescription());
JobClient.runJob(job);
LOG.info("Done Running " + getDescription());
}
}