/**
* (C) Copyright IBM Corp. 2010, 2015
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*
*/
package com.ibm.bi.dml.runtime.matrix.mapred;
import java.io.IOException;
import java.util.regex.Pattern;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.ByteWritable;
import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapred.JobConf;
import org.apache.hadoop.mapred.MapReduceBase;
import org.apache.hadoop.mapred.Mapper;
import org.apache.hadoop.mapred.OutputCollector;
import org.apache.hadoop.mapred.Reporter;
import org.apache.wink.json4j.JSONException;
import com.ibm.bi.dml.runtime.DMLRuntimeException;
import com.ibm.bi.dml.runtime.DMLUnsupportedOperationException;
import com.ibm.bi.dml.runtime.instructions.mr.CSVReblockInstruction;
import com.ibm.bi.dml.runtime.matrix.CSVReblockMR;
import com.ibm.bi.dml.runtime.matrix.CSVReblockMR.OffsetCount;
import com.ibm.bi.dml.runtime.transform.TfUtils;
public class CSVAssignRowIDMapper extends MapReduceBase implements Mapper<LongWritable, Text, ByteWritable, OffsetCount>
{
private ByteWritable outKey=new ByteWritable();
private long fileOffset=0;
private long num=0;
private boolean first=true;
private OutputCollector<ByteWritable, OffsetCount> outCache=null;
private String delim=" ";
private boolean ignoreFirstLine=false;
private boolean realFirstLine=false;
private String filename="";
private boolean headerFile=false;
// members relevant to transform
TfUtils _agents = null;
@Override
public void map(LongWritable key, Text value,
OutputCollector<ByteWritable, OffsetCount> out, Reporter report)
throws IOException
{
if(first) {
first=false;
fileOffset=key.get();
outCache=out;
}
if(key.get()==0 && headerFile)//getting the number of colums
{
if(!ignoreFirstLine)
{
report.incrCounter(CSVReblockMR.NUM_COLS_IN_MATRIX, outKey.toString(), value.toString().split(delim, -1).length);
if(!omit(value.toString()))
num++;
}
else
realFirstLine=true;
}
else
{
if(realFirstLine)
{
report.incrCounter(CSVReblockMR.NUM_COLS_IN_MATRIX, outKey.toString(), value.toString().split(delim, -1).length);
realFirstLine=false;
}
if(!omit(value.toString()))
num++;
}
}
@Override
@SuppressWarnings("deprecation")
public void configure(JobConf job)
{
byte thisIndex;
try {
//it doesn't make sense to have repeated file names in the input, since this is for reblock
thisIndex=MRJobConfiguration.getInputMatrixIndexesInMapper(job).get(0);
outKey.set(thisIndex);
FileSystem fs=FileSystem.get(job);
Path thisPath=new Path(job.get("map.input.file")).makeQualified(fs);
filename=thisPath.toString();
String[] strs=job.getStrings(CSVReblockMR.SMALLEST_FILE_NAME_PER_INPUT);
Path headerPath=new Path(strs[thisIndex]).makeQualified(fs);
if(headerPath.toString().equals(filename))
headerFile=true;
} catch (IOException e) {
throw new RuntimeException(e);
}
try {
CSVReblockInstruction[] reblockInstructions = MRJobConfiguration.getCSVReblockInstructions(job);
for(CSVReblockInstruction ins: reblockInstructions)
{
if(ins.input==thisIndex)
{
delim=Pattern.quote(ins.delim);
ignoreFirstLine=ins.hasHeader;
break;
}
}
} catch (DMLUnsupportedOperationException e) {
throw new RuntimeException(e);
} catch (DMLRuntimeException e) {
throw new RuntimeException(e);
}
// load properties relevant to transform
try {
boolean omit = job.getBoolean(MRJobConfiguration.TF_TRANSFORM, false);
if ( omit )
_agents = new TfUtils(job, true);
}
catch(IOException e) {
throw new RuntimeException(e);
}
catch(JSONException e) {
throw new RuntimeException(e);
}
}
private boolean omit(String line)
{
if(_agents == null)
return false;
return _agents.omit( line.split(delim, -1) );
}
@Override
public void close() throws IOException
{
outCache.collect(outKey, new OffsetCount(filename, fileOffset, num));
}
}