/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.archive.bacon;
import java.io.*;
import java.util.*;
import org.apache.pig.EvalFunc;
import org.apache.pig.PigException;
import org.apache.pig.backend.executionengine.ExecException;
import org.apache.pig.data.BagFactory;
import org.apache.pig.data.DataBag;
import org.apache.pig.data.DataType;
import org.apache.pig.data.Tuple;
import org.apache.pig.data.TupleFactory;
import org.apache.pig.impl.logicalLayer.FrontendException;
import org.apache.pig.impl.logicalLayer.schema.Schema;
import org.apache.pig.FuncSpec;
/**
* Version of the Pig built-in TOKENIZE() function which operates like
* STRSPLIT() but returns the tokens in a Bag rather than a Tuple.
*
* Source code derived from the Pig TOKENIZE() and STRSPLIT()
* built-ins.
*/
public class Tokenize extends EvalFunc<DataBag>
{
TupleFactory mTupleFactory = TupleFactory.getInstance();
BagFactory mBagFactory = BagFactory.getInstance();
@Override
public DataBag exec(Tuple input) throws IOException
{
try
{
if ( input == null || input.size() < 1 ) return null;
String source = (String) input.get(0);
String delim = (input.size() > 1 ) ? (String) input.get(1) : "\\s";
int length = (input.size() > 2 ) ? (Integer) input.get(2) : 0;
if ( source == null || delim == null ) return null;
DataBag output = mBagFactory.newDefaultBag();
for ( String token : source.split(delim, length) )
{
output.add( mTupleFactory.newTuple( token ) );
}
return output;
}
catch (ExecException ee)
{
throw ee;
}
}
@SuppressWarnings("deprecation")
@Override
public Schema outputSchema(Schema input)
{
try
{
Schema.FieldSchema tokenFs = new Schema.FieldSchema("token", DataType.CHARARRAY);
Schema tupleSchema = new Schema(tokenFs);
Schema.FieldSchema tupleFs;
tupleFs = new Schema.FieldSchema("tuple_of_tokens", tupleSchema, DataType.TUPLE);
Schema bagSchema = new Schema(tupleFs);
bagSchema.setTwoLevelAccessRequired(true);
Schema.FieldSchema bagFs = new Schema.FieldSchema("bag_of_tokenTuples",bagSchema, DataType.BAG);
return new Schema(bagFs);
}
catch (FrontendException e)
{
// throwing RTE because above schema creation is not expected
// to throw an exception and also because superclass does not
// throw exception
throw new RuntimeException("Unable to compute Tokenize schema.");
}
}
/* Omit this from our Tokenizer so that the multi-input calls can be mapped to it.
public List<FuncSpec> getArgToFuncMapping() throws FrontendException
{
List<FuncSpec> funcList = new ArrayList<FuncSpec>();
Schema s = new Schema();
s.add(new Schema.FieldSchema(null, DataType.CHARARRAY));
funcList.add(new FuncSpec(this.getClass().getName(), s));
return funcList;
}
*/
}