/** * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.pig.backend.hadoop.executionengine.tez.plan; import java.io.IOException; import java.io.Serializable; import java.util.ArrayList; import java.util.Arrays; import java.util.Collections; import java.util.HashMap; import java.util.Iterator; import java.util.List; import java.util.Map; import java.util.Properties; import java.util.Set; import java.util.Stack; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.io.IntWritable; import org.apache.hadoop.io.LongWritable; import org.apache.hadoop.mapreduce.lib.partition.HashPartitioner; import org.apache.hadoop.util.StringUtils; import org.apache.hadoop.util.hash.Hash; import org.apache.pig.CollectableLoadFunc; import org.apache.pig.FuncSpec; import org.apache.pig.IndexableLoadFunc; import org.apache.pig.LoadFunc; import org.apache.pig.OrderedLoadFunc; import org.apache.pig.PigConfiguration; import org.apache.pig.PigException; import org.apache.pig.backend.executionengine.ExecException; import org.apache.pig.backend.hadoop.datastorage.ConfigurationUtil; import org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler.PigTupleWritableComparator; import org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MergeJoinIndexer; import org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigWritableComparators; import org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.plans.ScalarPhyFinder; import org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.plans.UDFFinder; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.POStatus; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.PhysicalOperator; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.expressionOperators.ConstantExpression; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.expressionOperators.POProject; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.expressionOperators.POUserFunc; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.plans.PhyPlanVisitor; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.plans.PhysicalPlan; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.LitePackager; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POCollectedGroup; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POCounter; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POCross; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.PODistinct; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POFRJoin; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POFilter; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POForEach; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POGlobalRearrange; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POLimit; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POLoad; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POLocalRearrange; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POMergeCogroup; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POMergeJoin; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.PONative; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POPackage; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POPoissonSample; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.PORank; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POReservoirSample; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POSkewedJoin; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POSort; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POSplit; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POStore; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POStream; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POUnion; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.Packager.PackageType; import org.apache.pig.backend.hadoop.executionengine.physicalLayer.util.PlanHelper; import org.apache.pig.backend.hadoop.executionengine.tez.plan.operator.BloomPackager; import org.apache.pig.backend.hadoop.executionengine.tez.plan.operator.NativeTezOper; import org.apache.pig.backend.hadoop.executionengine.tez.plan.operator.POBloomFilterRearrangeTez; import org.apache.pig.backend.hadoop.executionengine.tez.plan.operator.POBuildBloomRearrangeTez; import org.apache.pig.backend.hadoop.executionengine.tez.plan.operator.POCounterStatsTez; import org.apache.pig.backend.hadoop.executionengine.tez.plan.operator.POCounterTez; import org.apache.pig.backend.hadoop.executionengine.tez.plan.operator.POFRJoinTez; import org.apache.pig.backend.hadoop.executionengine.tez.plan.operator.POIdentityInOutTez; import org.apache.pig.backend.hadoop.executionengine.tez.plan.operator.POLocalRearrangeTez; import org.apache.pig.backend.hadoop.executionengine.tez.plan.operator.POLocalRearrangeTezFactory; import org.apache.pig.backend.hadoop.executionengine.tez.plan.operator.POLocalRearrangeTezFactory.LocalRearrangeType; import org.apache.pig.backend.hadoop.executionengine.tez.plan.operator.POPartitionRearrangeTez; import org.apache.pig.backend.hadoop.executionengine.tez.plan.operator.PORankTez; import org.apache.pig.backend.hadoop.executionengine.tez.plan.operator.POShuffledValueInputTez; import org.apache.pig.backend.hadoop.executionengine.tez.plan.operator.POStoreTez; import org.apache.pig.backend.hadoop.executionengine.tez.plan.operator.POValueInputTez; import org.apache.pig.backend.hadoop.executionengine.tez.plan.operator.POValueOutputTez; import org.apache.pig.backend.hadoop.executionengine.tez.plan.udf.FindQuantilesTez; import org.apache.pig.backend.hadoop.executionengine.tez.plan.udf.IsFirstReduceOfKeyTez; import org.apache.pig.backend.hadoop.executionengine.tez.plan.udf.PartitionSkewedKeysTez; import org.apache.pig.backend.hadoop.executionengine.tez.plan.udf.ReadScalarsTez; import org.apache.pig.backend.hadoop.executionengine.tez.runtime.SkewedPartitionerTez; import org.apache.pig.backend.hadoop.executionengine.tez.runtime.WeightedRangePartitionerTez; import org.apache.pig.backend.hadoop.executionengine.tez.util.TezCompilerUtil; import org.apache.pig.data.DataType; import org.apache.pig.impl.PigContext; import org.apache.pig.impl.builtin.DefaultIndexableLoader; import org.apache.pig.impl.builtin.GetMemNumRows; import org.apache.pig.impl.builtin.PartitionSkewedKeys; import org.apache.pig.impl.io.FileLocalizer; import org.apache.pig.impl.io.FileSpec; import org.apache.pig.impl.io.NullableIntWritable; import org.apache.pig.impl.plan.DepthFirstWalker; import org.apache.pig.impl.plan.NodeIdGenerator; import org.apache.pig.impl.plan.Operator; import org.apache.pig.impl.plan.OperatorKey; import org.apache.pig.impl.plan.OperatorPlan; import org.apache.pig.impl.plan.PlanException; import org.apache.pig.impl.plan.VisitorException; import org.apache.pig.impl.util.CompilerUtils; import org.apache.pig.impl.util.MultiMap; import org.apache.pig.impl.util.ObjectSerializer; import org.apache.pig.impl.util.Pair; import org.apache.pig.impl.util.Utils; import org.apache.pig.newplan.logical.relational.LOJoin; import org.apache.tez.dag.api.EdgeProperty.DataMovementType; import org.apache.tez.runtime.library.input.OrderedGroupedKVInput; import org.apache.tez.runtime.library.input.UnorderedKVInput; import org.apache.tez.runtime.library.output.UnorderedKVOutput; import com.google.common.collect.Lists; import com.google.common.collect.Maps; import com.google.common.collect.Sets; /** * The compiler that compiles a given physical plan into a DAG of Tez * operators which can then be converted into the JobControl structure. */ public class TezCompiler extends PhyPlanVisitor { private static final Log LOG = LogFactory.getLog(TezCompiler.class); private PigContext pigContext; private Properties pigProperties; //The plan that is being compiled private PhysicalPlan plan; //The plan of Tez Operators private TezOperPlan tezPlan; //The current Tez Operator that is being compiled private TezOperator curTezOp; //The output of compiling the inputs private TezOperator[] compiledInputs = null; //The split operators seen till now. //During the traversal a split is the only operator that can be revisited //from a different path. private Map<OperatorKey, TezOperator> splitsSeen; private NodeIdGenerator nig; private String scope; private UDFFinder udfFinder; private Map<PhysicalOperator, TezOperator> phyToTezOpMap; // Contains the inputs to operator like join, with the list maintaining the // same order of join from left to right private Map<TezOperator, List<TezOperator>> inputsMap; public static final String USER_COMPARATOR_MARKER = "user.comparator.func:"; public static final String FILE_CONCATENATION_THRESHOLD = "pig.files.concatenation.threshold"; public static final String OPTIMISTIC_FILE_CONCATENATION = "pig.optimistic.files.concatenation"; private int fileConcatenationThreshold = 100; private boolean optimisticFileConcatenation = false; private List<String> readOnceLoadFuncs = null; private Configuration conf; private POLocalRearrangeTezFactory localRearrangeFactory; public TezCompiler(PhysicalPlan plan, PigContext pigContext) throws TezCompilerException { super(plan, new DepthFirstWalker<PhysicalOperator, PhysicalPlan>(plan)); this.plan = plan; this.pigContext = pigContext; pigProperties = pigContext.getProperties(); conf = ConfigurationUtil.toConfiguration(pigProperties, false); splitsSeen = Maps.newHashMap(); tezPlan = new TezOperPlan(); nig = NodeIdGenerator.getGenerator(); udfFinder = new UDFFinder(); List<PhysicalOperator> roots = plan.getRoots(); if((roots == null) || (roots.size() <= 0)) { int errCode = 2053; String msg = "Internal error. Did not find roots in the physical plan."; throw new TezCompilerException(msg, errCode, PigException.BUG); } scope = roots.get(0).getOperatorKey().getScope(); localRearrangeFactory = new POLocalRearrangeTezFactory(scope, nig); phyToTezOpMap = Maps.newHashMap(); inputsMap = Maps.newHashMap(); fileConcatenationThreshold = Integer.parseInt(pigProperties .getProperty(FILE_CONCATENATION_THRESHOLD, "100")); optimisticFileConcatenation = pigProperties.getProperty( OPTIMISTIC_FILE_CONCATENATION, "false").equals("true"); LOG.info("File concatenation threshold: " + fileConcatenationThreshold + " optimistic? " + optimisticFileConcatenation); String loadFuncs = pigContext.getProperties().getProperty( PigConfiguration.PIG_SORT_READONCE_LOADFUNCS); if (loadFuncs != null && loadFuncs.trim().length() > 0) { readOnceLoadFuncs = Arrays.asList(StringUtils.split(loadFuncs.trim())); } } public TezOperPlan getTezPlan() { return tezPlan; } // Segment a single DAG into a DAG graph public TezPlanContainer getPlanContainer() throws PlanException { TezPlanContainer tezPlanContainer = new TezPlanContainer(pigContext); tezPlanContainer.addPlan(tezPlan); return tezPlanContainer; } /** * The front-end method that the user calls to compile the plan. Assumes * that all submitted plans have a Store operators as the leaf. * @return A Tez plan * @throws IOException * @throws PlanException * @throws VisitorException */ public TezOperPlan compile() throws IOException, PlanException, VisitorException { List<PhysicalOperator> leaves = plan.getLeaves(); if (!pigContext.inIllustrator) { for (PhysicalOperator op : leaves) { if (!(op instanceof POStore)) { int errCode = 2025; String msg = "Expected leaf of reduce plan to " + "always be POStore. Found " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG); } } } // get all stores, sort them in order(operator id) and compile their // plans List<POStore> stores = PlanHelper.getPhysicalOperators(plan, POStore.class); List<PONative> nativeTezs= PlanHelper.getPhysicalOperators(plan, PONative.class); List<PhysicalOperator> ops; if (!pigContext.inIllustrator) { ops = new ArrayList<PhysicalOperator>(stores.size() + nativeTezs.size()); ops.addAll(stores); } else { ops = new ArrayList<PhysicalOperator>(leaves.size() + nativeTezs.size()); ops.addAll(leaves); } ops.addAll(nativeTezs); Collections.sort(ops); for (PhysicalOperator op : ops) { compile(op); } for (TezOperator tezOper : splitsSeen.values()) { int idx = 0; List<POStore> strs = PlanHelper.getPhysicalOperators(tezOper.plan, POStore.class); for (POStore op : strs) { op.setIndex(idx++); } tezOper.setClosed(true); } fixScalar(); return tezPlan; } private void fixScalar() throws VisitorException, PlanException { // Mapping POStore to POValueOuptut Map<POStore, POValueOutputTez> storeSeen = new HashMap<POStore, POValueOutputTez>(); for (TezOperator tezOp : tezPlan) { List<POUserFunc> userFuncs = PlanHelper.getPhysicalOperators(tezOp.plan, POUserFunc.class); for (POUserFunc userFunc : userFuncs) { if (userFunc.getReferencedOperator()!=null) { // Scalar POStore store = (POStore)userFunc.getReferencedOperator(); TezOperator from = phyToTezOpMap.get(store); FuncSpec newSpec = new FuncSpec(ReadScalarsTez.class.getName(), from.getOperatorKey().toString()); userFunc.setFuncSpec(newSpec); //Remove unused store filename if (userFunc.getInputs().size() == 2) { userFunc.getInputs().remove(1); } if (storeSeen.containsKey(store)) { storeSeen.get(store).addOutputKey(tezOp.getOperatorKey().toString()); } else { POValueOutputTez output = new POValueOutputTez(OperatorKey.genOpKey(scope)); output.setScalarOutput(true); output.addOutputKey(tezOp.getOperatorKey().toString()); from.plan.remove(from.plan.getOperator(store.getOperatorKey())); from.plan.addAsLeaf(output); storeSeen.put(store, output); } if (tezPlan.getPredecessors(tezOp)==null || !tezPlan.getPredecessors(tezOp).contains(from)) { TezEdgeDescriptor edge = TezCompilerUtil.connect(tezPlan, from, tezOp); //TODO shared edge once support is available in Tez TezCompilerUtil.configureValueOnlyTupleOutput(edge, DataMovementType.BROADCAST); } } } } } /** * Compiles the plan below op into a Tez Operator * and stores it in curTezOp. * @param op * @throws IOException * @throws PlanException * @throws VisitorException */ private void compile(PhysicalOperator op) throws IOException, PlanException, VisitorException { // An artifact of the Visitor. Need to save this so that it is not // overwritten. TezOperator[] prevCompInp = compiledInputs; // Compile each predecessor into the TezOperator and store them away so // that we can use them for compiling op. List<PhysicalOperator> predecessors = plan.getPredecessors(op); if(op instanceof PONative){ // the predecessor (store) has already been processed // don't process it again } else if (predecessors != null && predecessors.size() > 0) { // When processing an entire script (multiquery), we can get into a // situation where a load has predecessors. This means that it // depends on some store earlier in the plan. We need to take that // dependency and connect the respective Tez operators, while at the // same time removing the connection between the Physical operators. // That way the jobs will run in the right order. if (op instanceof POLoad) { if (predecessors.size() != 1) { int errCode = 2125; String msg = "Expected at most one predecessor of load. Got "+predecessors.size(); throw new PlanException(msg, errCode, PigException.BUG); } PhysicalOperator p = predecessors.get(0); TezOperator oper = null; if (p instanceof POStore || p instanceof PONative) { oper = phyToTezOpMap.get(p); } else { int errCode = 2126; String msg = "Predecessor of load should be a store or native oper. Got " + p.getClass(); throw new PlanException(msg, errCode, PigException.BUG); } curTezOp = getTezOp(); curTezOp.plan.add(op); curTezOp.setUseMRMapSettings(true); if (((POLoad) op).getLFile() != null && ((POLoad) op).getLFile().getFuncSpec() != null) { curTezOp.UDFs.add(((POLoad)op).getLFile().getFuncSpec().toString()); } tezPlan.add(curTezOp); phyToTezOpMap.put(op, curTezOp); plan.disconnect(op, p); TezCompilerUtil.connect(tezPlan, oper, curTezOp); oper.segmentBelow = true; return; } Collections.sort(predecessors); if(op instanceof POSplit && splitsSeen.containsKey(op.getOperatorKey())){ // skip follow up POSplit } else { compiledInputs = new TezOperator[predecessors.size()]; int i = -1; for (PhysicalOperator pred : predecessors) { compile(pred); compiledInputs[++i] = curTezOp; } } } else { // No predecessors. Mostly a load. But this is where we start. We // create a new TezOp and add its first operator op. Also this // should be added to the tezPlan. curTezOp = getTezOp(); curTezOp.plan.add(op); curTezOp.setUseMRMapSettings(true); if (op !=null && op instanceof POLoad) { if (((POLoad)op).getLFile()!=null && ((POLoad)op).getLFile().getFuncSpec()!=null) curTezOp.UDFs.add(((POLoad)op).getLFile().getFuncSpec().toString()); } tezPlan.add(curTezOp); phyToTezOpMap.put(op, curTezOp); return; } // Now we have the inputs compiled. Do something with the input oper op. op.visit(this); compiledInputs = prevCompInp; } private void nonBlocking(PhysicalOperator op) throws PlanException, IOException { TezOperator tezOp; if (compiledInputs.length == 1) { tezOp = compiledInputs[0]; if (tezOp.isClosed()) { int errCode = 2027; String msg = "Tez operator has been closed. This is unexpected for a merge."; throw new PlanException(msg, errCode, PigException.BUG); } } else { tezOp = merge(compiledInputs); } tezOp.plan.addAsLeaf(op); curTezOp = tezOp; } private void blocking() throws IOException, PlanException { TezOperator newTezOp = getTezOp(); tezPlan.add(newTezOp); for (TezOperator tezOp : compiledInputs) { tezOp.setClosed(true); TezCompilerUtil.connect(tezPlan, tezOp, newTezOp); } curTezOp = newTezOp; } private POSplit findPOSplit(TezOperator tezOp, OperatorKey splitKey) throws PlanException { POSplit split = (POSplit) tezOp.plan.getOperator(splitKey); if (split != null) { // The split is the leaf operator. return split; } else { // The split is a nested split Stack<POSplit> stack = new Stack<POSplit>(); split = (POSplit) tezOp.plan.getLeaves().get(0); stack.push(split); while (!stack.isEmpty()) { split = stack.pop(); for (PhysicalPlan plan : split.getPlans()) { PhysicalOperator op = plan.getLeaves().get(0); if (op instanceof POSplit) { split = (POSplit) op; if (split.getOperatorKey().equals(splitKey)) { return split; } else { stack.push(split); } } } } } // TODO: what should be the new error code?? throw new PlanException( "Could not find the split operator " + splitKey, 2059, PigException.BUG); } /** * Merges the TezOperators in the compiledInputs into a single merged * TezOperator. * * Care is taken to remove the TezOperators that are merged from the TezPlan * and their connections moved over to the merged map TezOperator. * * Merge is implemented as a sequence of binary merges. * merge(PhyPlan finPlan, List<PhyPlan> lst) := finPlan, merge(p) foreach p in lst * * @param compiledInputs * @return merged TezOperator * @throws PlanException * @throws IOException */ private TezOperator merge(TezOperator[] compiledInputs) throws PlanException { TezOperator ret = getTezOp(); tezPlan.add(ret); Set<TezOperator> toBeConnected = Sets.newHashSet(); List<TezOperator> toBeRemoved = Lists.newArrayList(); List<PhysicalPlan> toBeMerged = Lists.newArrayList(); for (TezOperator tezOp : compiledInputs) { if (!tezOp.isClosed()) { toBeRemoved.add(tezOp); toBeMerged.add(tezOp.plan); List<TezOperator> predecessors = tezPlan.getPredecessors(tezOp); if (predecessors != null) { for (TezOperator predecessorTezOp : predecessors) { toBeConnected.add(predecessorTezOp); } } } else { int errCode = 2027; String msg = "Tez operator has been closed. This is unexpected for a merge."; throw new PlanException(msg, errCode, PigException.BUG); } } merge(ret.plan, toBeMerged); Iterator<TezOperator> it = toBeConnected.iterator(); while (it.hasNext()) { tezPlan.connect(it.next(), ret); } for (TezOperator tezOp : toBeRemoved) { if (tezOp.getRequestedParallelism() > ret.getRequestedParallelism()) { ret.setRequestedParallelism(tezOp.getRequestedParallelism()); } for (String udf : tezOp.UDFs) { if (!ret.UDFs.contains(udf)) { ret.UDFs.add(udf); } } // We also need to change scalar marking for (PhysicalOperator physOp : tezOp.scalars) { if (!ret.scalars.contains(physOp)) { ret.scalars.add(physOp); } } Set<PhysicalOperator> opsToChange = Sets.newHashSet(); for (Map.Entry<PhysicalOperator, TezOperator> entry : phyToTezOpMap.entrySet()) { if (entry.getValue()==tezOp) { opsToChange.add(entry.getKey()); } } for (PhysicalOperator op : opsToChange) { phyToTezOpMap.put(op, ret); } tezPlan.remove(tezOp); } return ret; } /** * The merge of a list of plans into a single plan * @param <O> * @param <E> * @param finPlan - Final Plan into which the list of plans is merged * @param plans - list of plans to be merged * @throws PlanException */ private <O extends Operator<?>, E extends OperatorPlan<O>> void merge( E finPlan, List<E> plans) throws PlanException { for (E e : plans) { finPlan.merge(e); } } private void processUDFs(PhysicalPlan plan) throws VisitorException { if (plan != null) { //Process Scalars (UDF with referencedOperators) ScalarPhyFinder scalarPhyFinder = new ScalarPhyFinder(plan); scalarPhyFinder.visit(); curTezOp.scalars.addAll(scalarPhyFinder.getScalars()); //Process UDFs udfFinder.setPlan(plan); udfFinder.visit(); curTezOp.UDFs.addAll(udfFinder.getUDFs()); } } @Override public void visitCollectedGroup(POCollectedGroup op) throws VisitorException { List<PhysicalOperator> roots = curTezOp.plan.getRoots(); if(roots.size() != 1){ int errCode = 2171; String errMsg = "Expected one but found more then one root physical operator in physical plan."; throw new TezCompilerException(errMsg,errCode,PigException.BUG); } PhysicalOperator phyOp = roots.get(0); if(! (phyOp instanceof POLoad)){ int errCode = 2172; String errMsg = "Expected physical operator at root to be POLoad. Found : "+phyOp.getClass().getCanonicalName(); throw new TezCompilerException(errMsg,errCode,PigException.BUG); } LoadFunc loadFunc = ((POLoad)phyOp).getLoadFunc(); try { if(!(CollectableLoadFunc.class.isAssignableFrom(loadFunc.getClass()))){ int errCode = 2249; throw new TezCompilerException("While using 'collected' on group; data must be loaded via loader implementing CollectableLoadFunc.", errCode); } ((CollectableLoadFunc)loadFunc).ensureAllKeyInstancesInSameSplit(); } catch (TezCompilerException e){ throw (e); } catch (IOException e) { int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } try{ nonBlocking(op); phyToTezOpMap.put(op, curTezOp); }catch(Exception e){ int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } @Override public void visitCounter(POCounter op) throws VisitorException { // Refer visitRank(PORank) for more details try{ curTezOp.markRankCounter(); POCounterTez counterTez = new POCounterTez(op); nonBlocking(counterTez); phyToTezOpMap.put(op, curTezOp); } catch (Exception e) { int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } @Override public void visitCross(POCross op) throws VisitorException { try{ nonBlocking(op); phyToTezOpMap.put(op, curTezOp); } catch (Exception e) { int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } @Override public void visitDistinct(PODistinct op) throws VisitorException { try { POLocalRearrangeTez lr = localRearrangeFactory.create(); lr.setDistinct(true); lr.copyAliasFrom(op); curTezOp.plan.addAsLeaf(lr); TezOperator lastOp = curTezOp; // Mark the start of a new TezOperator, connecting the inputs. blocking(); TezCompilerUtil.setCustomPartitioner(op.getCustomPartitioner(), curTezOp); TezEdgeDescriptor edge = curTezOp.inEdges.get(lastOp.getOperatorKey()); edge.setNeedsDistinctCombiner(true); curTezOp.markDistinct(); addDistinctPlan(curTezOp.plan, op.getRequestedParallelism()); curTezOp.setRequestedParallelism(op.getRequestedParallelism()); phyToTezOpMap.put(op, curTezOp); } catch (Exception e) { int errCode = 2034; String msg = "Cannot compile " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG); } } // Adds the plan for DISTINCT. Note that the PODistinct is not actually added to the plan, but // rather is implemented by the action of the local rearrange, shuffle and project operations. private void addDistinctPlan(PhysicalPlan plan, int rp) throws PlanException { POPackage pkg = getPackage(1, DataType.TUPLE); pkg.getPkgr().setDistinct(true); plan.addAsLeaf(pkg); POProject project = new POProject(OperatorKey.genOpKey(scope)); project.setResultType(DataType.TUPLE); project.setStar(false); project.setColumn(0); project.setOverloaded(false); POForEach forEach = TezCompilerUtil.getForEach(project, rp, scope, nig); plan.addAsLeaf(forEach); } @Override public void visitFilter(POFilter op) throws VisitorException { try { nonBlocking(op); processUDFs(op.getPlan()); phyToTezOpMap.put(op, curTezOp); } catch (Exception e) { int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } @Override public void visitFRJoin(POFRJoin op) throws VisitorException { try { List<String> inputKeys = Lists.newArrayList(); curTezOp = phyToTezOpMap.get(op.getInputs().get(op.getFragment())); for (int i = 0; i < compiledInputs.length; i++) { TezOperator tezOp = compiledInputs[i]; if (curTezOp.equals(tezOp)) { continue; } if (!tezOp.isClosed()) { POLocalRearrangeTez lr = new POLocalRearrangeTez(op.getLRs()[i]); lr.setOutputKey(curTezOp.getOperatorKey().toString()); lr.setConnectedToPackage(false); tezOp.plan.addAsLeaf(lr); TezEdgeDescriptor edge = TezCompilerUtil.connect(tezPlan, tezOp, curTezOp); inputKeys.add(tezOp.getOperatorKey().toString()); // Configure broadcast edges for replicated tables edge.dataMovementType = DataMovementType.BROADCAST; edge.outputClassName = UnorderedKVOutput.class.getName(); edge.inputClassName = UnorderedKVInput.class.getName(); } else { int errCode = 2022; String msg = "The current operator is closed. This is unexpected while compiling."; throw new TezCompilerException(msg, errCode, PigException.BUG); } } if (!curTezOp.isClosed()) { curTezOp.plan.addAsLeaf(new POFRJoinTez(op, inputKeys)); } else { int errCode = 2022; String msg = "The current operator is closed. This is unexpected while compiling."; throw new TezCompilerException(msg, errCode, PigException.BUG); } List<List<PhysicalPlan>> joinPlans = op.getJoinPlans(); if (joinPlans != null) { for (List<PhysicalPlan> joinPlan : joinPlans) { if (joinPlan != null) { for (PhysicalPlan plan : joinPlan) { processUDFs(plan); } } } } phyToTezOpMap.put(op, curTezOp); } catch (Exception e) { int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } @Override public void visitLimit(POLimit op) throws VisitorException { try { if (op.getLimitPlan() != null) { processUDFs(op.getLimitPlan()); } // As an optimization, we'll add a limit to the end of the last tezOp. Note that in // some cases, such as when we have ORDER BY followed by LIMIT, the LimitOptimizer has // already removed the POLimit from the physical plan. if (!pigContext.inIllustrator) { nonBlocking(op); phyToTezOpMap.put(op, curTezOp); } boolean limitAfterSort = curTezOp.isGlobalSort(); // If the parallelism of the current vertex is one and it doesn't do a LOAD (whose // parallelism is determined by the InputFormat), we don't need another vertex. if (curTezOp.getRequestedParallelism() == 1 || curTezOp.isUnion()) { boolean canStop = true; for (PhysicalOperator planOp : curTezOp.plan.getRoots()) { if (planOp instanceof POLoad) { canStop = false; break; } } if (canStop) { // Let's piggyback on the Union operator. UnionOptimizer // will skip this union operator as it is a waste to // do a vertex group followed by another limit in this case if (curTezOp.isUnion()) { curTezOp.setRequestedParallelism(1); } curTezOp.setDontEstimateParallelism(true); if (limitAfterSort) { curTezOp.markLimitAfterSort(); } else { curTezOp.markLimit(); } return; } } // Need to add POValueOutputTez to the end of the last tezOp POValueOutputTez output = new POValueOutputTez(OperatorKey.genOpKey(scope)); output.copyAliasFrom(op); curTezOp.plan.addAsLeaf(output); TezOperator prevOp = curTezOp; // Mark the start of a new TezOperator which will do the actual limiting with 1 task. blocking(); // Explicitly set the parallelism for the new vertex to 1. curTezOp.setRequestedParallelism(1); curTezOp.setDontEstimateParallelism(true); output.addOutputKey(curTezOp.getOperatorKey().toString()); // LIMIT does not make any ordering guarantees and this is unsorted shuffle. TezEdgeDescriptor edge = curTezOp.inEdges.get(prevOp.getOperatorKey()); TezCompilerUtil.configureValueOnlyTupleOutput(edge, DataMovementType.SCATTER_GATHER); // Limit after order by with scalar expression if (limitAfterSort) { curTezOp.markLimitAfterSort(); output.setTaskIndexWithRecordIndexAsKey(true); // POValueOutputTez will write key (task index, record index) in // sorted order. So using UnorderedKVOutput instead of OrderedPartitionedKVOutput. // But input needs to be merged in sorter order and requires OrderedGroupedKVInput edge.outputClassName = UnorderedKVOutput.class.getName(); edge.inputClassName = OrderedGroupedKVInput.class.getName(); edge.setIntermediateOutputKeyClass(TezCompilerUtil.TUPLE_CLASS); edge.setIntermediateOutputKeyComparatorClass(PigTupleWritableComparator.class.getName()); } else { curTezOp.markLimit(); } // Then add a POValueInputTez to the start of the new tezOp. POValueInputTez input = new POValueInputTez(OperatorKey.genOpKey(scope)); input.copyAliasFrom(op); input.setInputKey(prevOp.getOperatorKey().toString()); curTezOp.plan.addAsLeaf(input); if (!pigContext.inIllustrator) { POLimit limitCopy = new POLimit(OperatorKey.genOpKey(scope)); limitCopy.copyAliasFrom(op); limitCopy.setLimit(op.getLimit()); limitCopy.setLimitPlan(op.getLimitPlan()); curTezOp.plan.addAsLeaf(limitCopy); } else { curTezOp.plan.addAsLeaf(op); } phyToTezOpMap.put(op, curTezOp); } catch (Exception e) { int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } @Override public void visitLoad(POLoad op) throws VisitorException { try { nonBlocking(op); curTezOp.setUseMRMapSettings(true); phyToTezOpMap.put(op, curTezOp); } catch (Exception e) { int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } @Override public void visitLocalRearrange(POLocalRearrange op) throws VisitorException { try{ POLocalRearrange opTez = new POLocalRearrangeTez(op); nonBlocking(opTez); List<PhysicalPlan> plans = opTez.getPlans(); if (plans != null) { for (PhysicalPlan ep : plans) { processUDFs(ep); } } phyToTezOpMap.put(opTez, curTezOp); } catch (Exception e) { int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } @Override public void visitGlobalRearrange(POGlobalRearrange op) throws VisitorException { try { blocking(); inputsMap.put(curTezOp, new ArrayList<>(Arrays.asList(compiledInputs))); TezCompilerUtil.setCustomPartitioner(op.getCustomPartitioner(), curTezOp); curTezOp.setRequestedParallelism(op.getRequestedParallelism()); if (op.isCross()) { curTezOp.addCrossKey(op.getOperatorKey().toString()); } phyToTezOpMap.put(op, curTezOp); } catch (Exception e) { int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } @Override public void visitMergeCoGroup(POMergeCogroup poCoGrp) throws VisitorException { if(compiledInputs.length < 2){ int errCode=2251; String errMsg = "Merge Cogroup work on two or more relations." + "To use map-side group-by on single relation, use 'collected' qualifier."; throw new TezCompilerException(errMsg, errCode); } List<FuncSpec> funcSpecs = new ArrayList<FuncSpec>(compiledInputs.length-1); List<String> fileSpecs = new ArrayList<String>(compiledInputs.length-1); List<String> loaderSigns = new ArrayList<String>(compiledInputs.length-1); try{ poCoGrp.setEndOfRecordMark(POStatus.STATUS_NULL); // Iterate through all the TezOpererators, disconnect side TezOperators from // TezOperator and collect all the information needed in different lists. for(int i=0 ; i < compiledInputs.length; i++){ TezOperator tezOper = compiledInputs[i]; PhysicalPlan plan = tezOper.plan; if(plan.getRoots().size() != 1){ int errCode = 2171; String errMsg = "Expected one but found more then one root physical operator in physical plan."; throw new TezCompilerException(errMsg,errCode,PigException.BUG); } PhysicalOperator rootPOOp = plan.getRoots().get(0); if(! (rootPOOp instanceof POLoad)){ int errCode = 2172; String errMsg = "Expected physical operator at root to be POLoad. Found : "+rootPOOp.getClass().getCanonicalName(); throw new TezCompilerException(errMsg,errCode); } POLoad sideLoader = (POLoad)rootPOOp; FileSpec loadFileSpec = sideLoader.getLFile(); FuncSpec funcSpec = loadFileSpec.getFuncSpec(); LoadFunc loadfunc = sideLoader.getLoadFunc(); if(i == 0){ if(!(CollectableLoadFunc.class.isAssignableFrom(loadfunc.getClass()))){ int errCode = 2252; throw new TezCompilerException("Base loader in Cogroup must implement CollectableLoadFunc.", errCode); } ((CollectableLoadFunc)loadfunc).ensureAllKeyInstancesInSameSplit(); continue; } if(!(IndexableLoadFunc.class.isAssignableFrom(loadfunc.getClass()))){ int errCode = 2253; throw new TezCompilerException("Side loaders in cogroup must implement IndexableLoadFunc.", errCode); } funcSpecs.add(funcSpec); fileSpecs.add(loadFileSpec.getFileName()); loaderSigns.add(sideLoader.getSignature()); tezPlan.remove(tezOper); } poCoGrp.setSideLoadFuncs(funcSpecs); poCoGrp.setSideFileSpecs(fileSpecs); poCoGrp.setLoaderSignatures(loaderSigns); // Use tez operator of base relation for the cogroup operation. TezOperator baseMROp = phyToTezOpMap.get(poCoGrp.getInputs().get(0)); if(baseMROp.isClosed()){ int errCode = 2254; throw new TezCompilerException("Currently merged cogroup is not supported after blocking operators.", errCode); } // Create new map-reduce operator for indexing job and then configure it. TezOperator indexerTezOp = getTezOp(); FileSpec idxFileSpec = getIndexingJob(indexerTezOp, baseMROp, poCoGrp.getLRInnerPlansOf(0)); poCoGrp.setIdxFuncSpec(idxFileSpec.getFuncSpec()); poCoGrp.setIndexFileName(idxFileSpec.getFileName()); baseMROp.plan.addAsLeaf(poCoGrp); for (FuncSpec funcSpec : funcSpecs) baseMROp.UDFs.add(funcSpec.toString()); phyToTezOpMap.put(poCoGrp,baseMROp); // Going forward, new operators should be added in baseMRop. To make // sure, reset curMROp. curTezOp = baseMROp; } catch (ExecException e){ throw new TezCompilerException(e.getDetailedMessage(),e.getErrorCode(),e.getErrorSource(),e); } catch (TezCompilerException mrce){ throw(mrce); } catch (CloneNotSupportedException e) { throw new TezCompilerException(e); } catch(PlanException e){ int errCode = 2034; String msg = "Error compiling operator " + poCoGrp.getClass().getCanonicalName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } catch (IOException e){ int errCode = 3000; String errMsg = "IOException caught while compiling POMergeCoGroup"; throw new TezCompilerException(errMsg, errCode,e); } } // Sets up the indexing job for map-side cogroups. private FileSpec getIndexingJob(TezOperator indexerTezOp, final TezOperator baseTezOp, final List<PhysicalPlan> mapperLRInnerPlans) throws TezCompilerException, PlanException, ExecException, IOException, CloneNotSupportedException { // First replace loader with MergeJoinIndexer. PhysicalPlan basePlan = baseTezOp.plan; POLoad baseLoader = (POLoad)basePlan.getRoots().get(0); FileSpec origLoaderFileSpec = baseLoader.getLFile(); FuncSpec funcSpec = origLoaderFileSpec.getFuncSpec(); LoadFunc loadFunc = baseLoader.getLoadFunc(); if (! (OrderedLoadFunc.class.isAssignableFrom(loadFunc.getClass()))){ int errCode = 1104; String errMsg = "Base relation of merge-coGroup must implement " + "OrderedLoadFunc interface. The specified loader " + funcSpec + " doesn't implement it"; throw new TezCompilerException(errMsg,errCode); } String[] indexerArgs = new String[6]; indexerArgs[0] = funcSpec.toString(); indexerArgs[1] = ObjectSerializer.serialize((Serializable)mapperLRInnerPlans); indexerArgs[3] = baseLoader.getSignature(); indexerArgs[4] = baseLoader.getOperatorKey().scope; indexerArgs[5] = Boolean.toString(false); // we care for nulls. PhysicalPlan phyPlan; if (basePlan.getSuccessors(baseLoader) == null || basePlan.getSuccessors(baseLoader).isEmpty()){ // Load-Load-Cogroup case. phyPlan = null; } else{ // We got something. Yank it and set it as inner plan. phyPlan = basePlan.clone(); PhysicalOperator root = phyPlan.getRoots().get(0); phyPlan.disconnect(root, phyPlan.getSuccessors(root).get(0)); phyPlan.remove(root); } indexerArgs[2] = ObjectSerializer.serialize(phyPlan); POLoad idxJobLoader = new POLoad(new OperatorKey(scope,nig.getNextNodeId(scope))); idxJobLoader.setPc(pigContext); idxJobLoader.setIsTmpLoad(true); idxJobLoader.setLFile(new FileSpec(origLoaderFileSpec.getFileName(), new FuncSpec(MergeJoinIndexer.class.getName(), indexerArgs))); indexerTezOp.plan.add(idxJobLoader); indexerTezOp.UDFs.add(baseLoader.getLFile().getFuncSpec().toString()); // Loader of mro will return a tuple of form - // (key1, key2, .. , WritableComparable, splitIndex). See MergeJoinIndexer for details. // After getting an index entry in each mapper, send all of them to one // vertex where they will be sorted on the way by Hadoop. TezOperator indexAggrOper = getTezOp(); tezPlan.add(indexAggrOper); tezPlan.add(indexerTezOp); TezCompilerUtil.simpleConnectTwoVertex(tezPlan, indexerTezOp, indexAggrOper, scope, nig); TezCompilerUtil.connect(tezPlan, indexAggrOper, baseTezOp); indexAggrOper.segmentBelow = true; indexerTezOp.setRequestedParallelism(1); // we need exactly one reducer for indexing job. indexerTezOp.setDontEstimateParallelism(true); POStore st = TezCompilerUtil.getStore(scope, nig); FileSpec strFile = getTempFileSpec(pigContext); st.setSFile(strFile); indexAggrOper.plan.addAsLeaf(st); indexAggrOper.setClosed(true); return strFile; } /** Since merge-join works on two inputs there are exactly two TezOper predecessors identified as left and right. * Right input generates index on-the-fly. This consists of two Tez vertexes. The first vertex generates index, * and the second vertex sort them. * Left input contains POMergeJoin which do the actual join. * First right Tez oper is identified as rightTezOpr, second is identified as rightTezOpr2 * Left Tez oper is identified as curTezOper. * 1) RightTezOpr: It can be preceded only by POLoad. If there is anything else * in physical plan, that is yanked and set as inner plans of joinOp. * 2) LeftTezOper: add the Join operator in it. * * We also need to segment the DAG into two, because POMergeJoin depends on the index file which loads with * DefaultIndexableLoader. It is possible to convert the index as a broadcast input, but that is costly * because a lot of logic is built into DefaultIndexableLoader. We can revisit it later. */ @Override public void visitMergeJoin(POMergeJoin joinOp) throws VisitorException { try{ if(compiledInputs.length != 2 || joinOp.getInputs().size() != 2){ int errCode=1101; throw new TezCompilerException("Merge Join must have exactly two inputs. Found : "+compiledInputs.length, errCode); } curTezOp = phyToTezOpMap.get(joinOp.getInputs().get(0)); TezOperator rightTezOpr = null; TezOperator rightTezOprAggr = null; if(curTezOp.equals(compiledInputs[0])) rightTezOpr = compiledInputs[1]; else rightTezOpr = compiledInputs[0]; // We will first operate on right side which is indexer job. // First yank plan of the compiled right input and set that as an inner plan of right operator. PhysicalPlan rightPipelinePlan; if(!rightTezOpr.closed){ PhysicalPlan rightPlan = rightTezOpr.plan; if(rightPlan.getRoots().size() != 1){ int errCode = 2171; String errMsg = "Expected one but found more then one root physical operator in physical plan."; throw new TezCompilerException(errMsg,errCode,PigException.BUG); } PhysicalOperator rightLoader = rightPlan.getRoots().get(0); if(! (rightLoader instanceof POLoad)){ int errCode = 2172; String errMsg = "Expected physical operator at root to be POLoad. Found : "+rightLoader.getClass().getCanonicalName(); throw new TezCompilerException(errMsg,errCode); } if (rightPlan.getSuccessors(rightLoader) == null || rightPlan.getSuccessors(rightLoader).isEmpty()) // Load - Join case. rightPipelinePlan = null; else{ // We got something on right side. Yank it and set it as inner plan of right input. rightPipelinePlan = rightPlan.clone(); PhysicalOperator root = rightPipelinePlan.getRoots().get(0); rightPipelinePlan.disconnect(root, rightPipelinePlan.getSuccessors(root).get(0)); rightPipelinePlan.remove(root); rightPlan.trimBelow(rightLoader); } } else{ int errCode = 2022; String msg = "Right input plan have been closed. This is unexpected while compiling."; throw new PlanException(msg, errCode, PigException.BUG); } joinOp.setupRightPipeline(rightPipelinePlan); // At this point, we must be operating on input plan of right input and it would contain nothing else other then a POLoad. POLoad rightLoader = (POLoad)rightTezOpr.plan.getRoots().get(0); joinOp.setSignature(rightLoader.getSignature()); LoadFunc rightLoadFunc = rightLoader.getLoadFunc(); List<String> udfs = new ArrayList<String>(); if(IndexableLoadFunc.class.isAssignableFrom(rightLoadFunc.getClass())) { joinOp.setRightLoaderFuncSpec(rightLoader.getLFile().getFuncSpec()); joinOp.setRightInputFileName(rightLoader.getLFile().getFileName()); udfs.add(rightLoader.getLFile().getFuncSpec().toString()); // we don't need the right TezOper since // the right loader is an IndexableLoadFunc which can handle the index // itself tezPlan.remove(rightTezOpr); if(rightTezOpr == compiledInputs[0]) { compiledInputs[0] = null; } else if(rightTezOpr == compiledInputs[1]) { compiledInputs[1] = null; } rightTezOpr = null; // validate that the join keys in merge join are only // simple column projections or '*' and not expression - expressions // cannot be handled when the index is built by the storage layer on the sorted // data when the sorted data (and corresponding index) is written. // So merge join will be restricted not have expressions as // join keys int numInputs = mPlan.getPredecessors(joinOp).size(); // should be 2 for(int i = 0; i < numInputs; i++) { List<PhysicalPlan> keyPlans = joinOp.getInnerPlansOf(i); for (PhysicalPlan keyPlan : keyPlans) { for(PhysicalOperator op : keyPlan) { if(!(op instanceof POProject)) { int errCode = 1106; String errMsg = "Merge join is possible only for simple column or '*' join keys when using " + rightLoader.getLFile().getFuncSpec() + " as the loader"; throw new TezCompilerException(errMsg, errCode, PigException.INPUT); } } } } } else { LoadFunc loadFunc = rightLoader.getLoadFunc(); //Replacing POLoad with indexer is disabled for 'merge-sparse' joins. While //this feature would be useful, the current implementation of DefaultIndexableLoader //is not designed to handle multiple calls to seekNear. Specifically, it rereads the entire index //for each call. Some refactoring of this class is required - and then the check below could be removed. if (joinOp.getJoinType() == LOJoin.JOINTYPE.MERGESPARSE) { int errCode = 1104; String errMsg = "Right input of merge-join must implement IndexableLoadFunc. " + "The specified loader " + loadFunc + " doesn't implement it"; throw new TezCompilerException(errMsg,errCode); } // Replace POLoad with indexer. if (! (OrderedLoadFunc.class.isAssignableFrom(loadFunc.getClass()))){ int errCode = 1104; String errMsg = "Right input of merge-join must implement " + "OrderedLoadFunc interface. The specified loader " + loadFunc + " doesn't implement it"; throw new TezCompilerException(errMsg,errCode); } String[] indexerArgs = new String[6]; List<PhysicalPlan> rightInpPlans = joinOp.getInnerPlansOf(1); FileSpec origRightLoaderFileSpec = rightLoader.getLFile(); indexerArgs[0] = origRightLoaderFileSpec.getFuncSpec().toString(); indexerArgs[1] = ObjectSerializer.serialize((Serializable)rightInpPlans); indexerArgs[2] = ObjectSerializer.serialize(rightPipelinePlan); indexerArgs[3] = rightLoader.getSignature(); indexerArgs[4] = rightLoader.getOperatorKey().scope; indexerArgs[5] = Boolean.toString(true); FileSpec lFile = new FileSpec(rightLoader.getLFile().getFileName(),new FuncSpec(MergeJoinIndexer.class.getName(), indexerArgs)); rightLoader.setLFile(lFile); // Loader of operator will return a tuple of form - // (keyFirst1, keyFirst2, .. , position, splitIndex) See MergeJoinIndexer rightTezOprAggr = getTezOp(); tezPlan.add(rightTezOprAggr); TezCompilerUtil.simpleConnectTwoVertex(tezPlan, rightTezOpr, rightTezOprAggr, scope, nig); rightTezOprAggr.setRequestedParallelism(1); // we need exactly one task for indexing job. rightTezOprAggr.setDontEstimateParallelism(true); POStore st = TezCompilerUtil.getStore(scope, nig); FileSpec strFile = getTempFileSpec(pigContext); st.setSFile(strFile); rightTezOprAggr.plan.addAsLeaf(st); rightTezOprAggr.setClosed(true); rightTezOprAggr.segmentBelow = true; // set up the DefaultIndexableLoader for the join operator String[] defaultIndexableLoaderArgs = new String[5]; defaultIndexableLoaderArgs[0] = origRightLoaderFileSpec.getFuncSpec().toString(); defaultIndexableLoaderArgs[1] = strFile.getFileName(); defaultIndexableLoaderArgs[2] = strFile.getFuncSpec().toString(); defaultIndexableLoaderArgs[3] = joinOp.getOperatorKey().scope; defaultIndexableLoaderArgs[4] = origRightLoaderFileSpec.getFileName(); joinOp.setRightLoaderFuncSpec((new FuncSpec(DefaultIndexableLoader.class.getName(), defaultIndexableLoaderArgs))); joinOp.setRightInputFileName(origRightLoaderFileSpec.getFileName()); joinOp.setIndexFile(strFile.getFileName()); udfs.add(origRightLoaderFileSpec.getFuncSpec().toString()); } // We are done with right side. Lets work on left now. // Join will be materialized in leftTezOper. if(!curTezOp.isClosed()) // Life is easy curTezOp.plan.addAsLeaf(joinOp); else{ int errCode = 2022; String msg = "Input plan has been closed. This is unexpected while compiling."; throw new PlanException(msg, errCode, PigException.BUG); } if(rightTezOprAggr != null) { rightTezOprAggr.markIndexer(); // We want to ensure indexing job runs prior to actual join job. So, connect them in order. TezCompilerUtil.connect(tezPlan, rightTezOprAggr, curTezOp); } phyToTezOpMap.put(joinOp, curTezOp); // no combination of small splits as there is currently no way to guarantee the sortness // of the combined splits. curTezOp.noCombineSmallSplits(); curTezOp.UDFs.addAll(udfs); } catch(PlanException e){ int errCode = 2034; String msg = "Error compiling operator " + joinOp.getClass().getCanonicalName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } catch (IOException e){ int errCode = 3000; String errMsg = "IOException caught while compiling POMergeJoin"; throw new TezCompilerException(errMsg, errCode,e); } catch(CloneNotSupportedException e){ int errCode = 2127; String errMsg = "Cloning exception caught while compiling POMergeJoin"; throw new TezCompilerException(errMsg, errCode, PigException.BUG, e); } } @Override public void visitNative(PONative op) throws VisitorException { try{ curTezOp.segmentBelow = true; TezOperator nativeTezOper = new NativeTezOper(new OperatorKey(scope,nig.getNextNodeId(scope)), op.getNativeMRjar(), op.getParams()); tezPlan.add(nativeTezOper); curTezOp.setClosed(true); tezPlan.connect(curTezOp, nativeTezOper); phyToTezOpMap.put(op, nativeTezOper); nativeTezOper.setClosed(true); nativeTezOper.segmentBelow = true; nativeTezOper.markNative(); curTezOp = nativeTezOper; }catch(Exception e){ int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } @Override public void visitPackage(POPackage op) throws VisitorException{ try{ nonBlocking(op); phyToTezOpMap.put(op, curTezOp); if (op.getPkgr().getPackageType() == PackageType.JOIN) { curTezOp.markRegularJoin(); } else if (op.getPkgr().getPackageType() == PackageType.GROUP) { if (op.getNumInps() == 1) { curTezOp.markGroupBy(); } else if (op.getNumInps() > 1) { curTezOp.markCogroup(); } } else if (op.getPkgr().getPackageType() == PackageType.BLOOMJOIN) { curTezOp.markRegularJoin(); addBloomToJoin(op, curTezOp); } } catch (Exception e) { int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } private void addBloomToJoin(POPackage op, TezOperator curTezOp) throws PlanException { List<TezOperator> inputs = inputsMap.get(curTezOp); TezOperator buildBloomOp; List<TezOperator> applyBloomOps = new ArrayList<>(); String strategy = conf.get(PigConfiguration.PIG_BLOOMJOIN_STRATEGY, POBuildBloomRearrangeTez.DEFAULT_BLOOM_STRATEGY); boolean createBloomInMap = "map".equals(strategy); if (!createBloomInMap && !strategy.equals("reduce")) { throw new PlanException(new IllegalArgumentException( "Invalid value for " + PigConfiguration.PIG_BLOOMJOIN_STRATEGY + " - " + strategy + ". Valid values are map and reduce")); } int numHash = conf.getInt(PigConfiguration.PIG_BLOOMJOIN_HASH_FUNCTIONS, POBuildBloomRearrangeTez.DEFAULT_NUM_BLOOM_HASH_FUNCTIONS); int vectorSizeBytes = conf.getInt(PigConfiguration.PIG_BLOOMJOIN_VECTORSIZE_BYTES, POBuildBloomRearrangeTez.DEFAULT_BLOOM_VECTOR_SIZE_BYTES); int numBloomFilters = POBuildBloomRearrangeTez.getNumBloomFilters(conf); int hashType = Hash.parseHashType(conf.get(PigConfiguration.PIG_BLOOMJOIN_HASH_TYPE, POBuildBloomRearrangeTez.DEFAULT_BLOOM_HASH_TYPE)); // We build bloom of the right most input and apply the bloom filter on the left inputs by default. // But in case of left outer join we build bloom of the left input and use it on the right input boolean[] inner = op.getPkgr().getInner(); boolean skipNullKeys = true; if (inner[inner.length - 1]) { // inner has from right to left while inputs has from left to right buildBloomOp = inputs.get(inputs.size() - 1); // Bloom filter is built from right most input for (int i = 0; i < (inner.length - 1); i++) { applyBloomOps.add(inputs.get(i)); } skipNullKeys = inner[0]; } else { // Left outer join skipNullKeys = false; buildBloomOp = inputs.get(0); // Bloom filter is built from left most input for (int i = 1; i < inner.length; i++) { applyBloomOps.add(inputs.get(i)); } } // Add BuildBloom operator to the input POLocalRearrangeTez lr = (POLocalRearrangeTez) buildBloomOp.plan.getLeaves().get(0); POBuildBloomRearrangeTez bbr = new POBuildBloomRearrangeTez(lr, createBloomInMap, numBloomFilters, vectorSizeBytes, numHash, hashType); bbr.setSkipNullKeys(skipNullKeys); buildBloomOp.plan.remove(lr); buildBloomOp.plan.addAsLeaf(bbr); // Add a new reduce vertex that will construct the final bloom filter // - by combining the bloom filters from the buildBloomOp input tasks in the map strategy // - or directly from the keys from the buildBloomOp input tasks in the reduce strategy TezOperator combineBloomOp = getTezOp(); tezPlan.add(combineBloomOp); combineBloomOp.markBuildBloom(); // Explicitly set the parallelism for the new vertex to number of bloom filters. // Auto parallelism will bring it down based on the actual output size combineBloomOp.setEstimatedParallelism(numBloomFilters); // We don't want parallelism to be changed during the run by grace auto parallelism // It will take the whole input size and estimate way higher combineBloomOp.setDontEstimateParallelism(true); String combineBloomOpKey = combineBloomOp.getOperatorKey().toString(); TezEdgeDescriptor edge = new TezEdgeDescriptor(); TezCompilerUtil.connect(tezPlan, buildBloomOp, combineBloomOp, edge); bbr.setBloomOutputKey(combineBloomOpKey); POPackage pkg = new POPackage(OperatorKey.genOpKey(scope)); pkg.setNumInps(1); BloomPackager pkgr = new BloomPackager(createBloomInMap, vectorSizeBytes, numHash, hashType);; pkgr.setKeyType(DataType.INTEGER); pkg.setPkgr(pkgr); POValueOutputTez combineBloomOutput = new POValueOutputTez(OperatorKey.genOpKey(scope)); combineBloomOp.plan.addAsLeaf(pkg); combineBloomOp.plan.addAsLeaf(combineBloomOutput); edge.setIntermediateOutputKeyClass(NullableIntWritable.class.getName()); edge.setIntermediateOutputKeyComparatorClass(PigWritableComparators.PigIntRawBytesComparator.class.getName()); // Add combiner as well. POPackage pkg_c = new POPackage(OperatorKey.genOpKey(scope)); BloomPackager combinerPkgr = new BloomPackager(createBloomInMap, vectorSizeBytes, numHash, hashType); combinerPkgr.setCombiner(true); combinerPkgr.setKeyType(DataType.INTEGER); pkg_c.setPkgr(combinerPkgr); pkg_c.setNumInps(1); edge.combinePlan.addAsLeaf(pkg_c); POProject prjKey = new POProject(OperatorKey.genOpKey(scope)); prjKey.setResultType(DataType.INTEGER); List<PhysicalPlan> clrInps = new ArrayList<PhysicalPlan>(); PhysicalPlan pp = new PhysicalPlan(); pp.add(prjKey); clrInps.add(pp); POLocalRearrangeTez clr = localRearrangeFactory.create(0, LocalRearrangeType.WITHPLAN, clrInps, DataType.INTEGER); clr.setOutputKey(combineBloomOpKey); edge.combinePlan.addAsLeaf(clr); if (createBloomInMap) { // No combiner needed on map as there will be only one bloom filter per map for each partition // In the reducer, the bloom filters will be combined with same logic of reduce in BloomPackager edge.setCombinerInMap(false); edge.setCombinerInReducer(true); } else { pkgr.setBloomKeyType(op.getPkgr().getKeyType()); // Do distinct of the keys on the map side to reduce data sent to reducers. // In case of reduce, not adding a combiner and doing the distinct during reduce itself. // If needed one can be added later edge.setCombinerInMap(true); edge.setCombinerInReducer(false); } // Broadcast the final bloom filter to other inputs for (TezOperator applyBloomOp : applyBloomOps) { applyBloomOp.markFilterBloom(); lr = (POLocalRearrangeTez) applyBloomOp.plan.getLeaves().get(0); POBloomFilterRearrangeTez bfr = new POBloomFilterRearrangeTez(lr, numBloomFilters); applyBloomOp.plan.remove(lr); applyBloomOp.plan.addAsLeaf(bfr); bfr.setInputKey(combineBloomOpKey); edge = new TezEdgeDescriptor(); edge.setIntermediateOutputKeyClass(NullableIntWritable.class.getName()); edge.setIntermediateOutputKeyComparatorClass(PigWritableComparators.PigIntRawBytesComparator.class.getName()); TezCompilerUtil.configureValueOnlyTupleOutput(edge, DataMovementType.BROADCAST); TezCompilerUtil.connect(tezPlan, combineBloomOp, applyBloomOp, edge); combineBloomOutput.addOutputKey(applyBloomOp.getOperatorKey().toString()); } } @Override public void visitPOForEach(POForEach op) throws VisitorException{ try{ nonBlocking(op); List<PhysicalPlan> plans = op.getInputPlans(); if (plans != null) { for (PhysicalPlan ep : plans) { processUDFs(ep); } } phyToTezOpMap.put(op, curTezOp); } catch (Exception e) { int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } @Override public void visitRank(PORank op) throws VisitorException { try{ // Rank implementation has 3 vertices // Vertex 1 has POCounterTez produce output tuples and send to Vertex 3 via 1-1 edge. // Vertex 1 also sends the count of tuples of each task in Vertex 1 to Vertex 2 which is a single reducer. // Vertex 3 has PORankTez which consumes from Vertex 2 as broadcast input and also tuples from Vertex 1 and // produces tuples with updated ranks based on the count of tuples from Vertex 2. // This is different from MR implementation where POCounter updates job counters, and that is // copied by JobControlCompiler into the PORank job's jobconf. // Previous operator is always POCounterTez (Vertex 1) TezOperator counterOper = curTezOp; POCounterTez counterTez = (POCounterTez) counterOper.plan.getLeaves().get(0); //Construct Vertex 2 TezOperator statsOper = getTezOp(); tezPlan.add(statsOper); POCounterStatsTez counterStatsTez = new POCounterStatsTez(OperatorKey.genOpKey(scope)); statsOper.plan.addAsLeaf(counterStatsTez); statsOper.setRequestedParallelism(1); statsOper.setDontEstimateParallelism(true); //Construct Vertex 3 TezOperator rankOper = getTezOp(); tezPlan.add(rankOper); PORankTez rankTez = new PORankTez(op); rankOper.plan.addAsLeaf(rankTez); curTezOp = rankOper; // Connect counterOper vertex to rankOper vertex by 1-1 edge rankOper.setRequestedParallelismByReference(counterOper); TezEdgeDescriptor edge = TezCompilerUtil.connect(tezPlan, counterOper, rankOper); rankOper.setUseMRMapSettings(counterOper.isUseMRMapSettings()); TezCompilerUtil.configureValueOnlyTupleOutput(edge, DataMovementType.ONE_TO_ONE); counterTez.setTuplesOutputKey(rankOper.getOperatorKey().toString()); rankTez.setTuplesInputKey(counterOper.getOperatorKey().toString()); // Connect counterOper vertex to statsOper vertex by Shuffle edge edge = TezCompilerUtil.connect(tezPlan, counterOper, statsOper); // Task id edge.setIntermediateOutputKeyClass(IntWritable.class.getName()); edge.partitionerClass = HashPartitioner.class; // Number of records in that task edge.setIntermediateOutputValueClass(LongWritable.class.getName()); counterTez.setStatsOutputKey(statsOper.getOperatorKey().toString()); counterStatsTez.setInputKey(counterOper.getOperatorKey().toString()); // Connect statsOper vertex to rankOper vertex by Broadcast edge edge = TezCompilerUtil.connect(tezPlan, statsOper, rankOper); // Map of task id, offset count based on total number of records is in the value TezCompilerUtil.configureValueOnlyTupleOutput(edge, DataMovementType.BROADCAST); counterStatsTez.setOutputKey(rankOper.getOperatorKey().toString()); rankTez.setStatsInputKey(statsOper.getOperatorKey().toString()); phyToTezOpMap.put(op, rankOper); } catch (Exception e) { int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } @Override public void visitSkewedJoin(POSkewedJoin op) throws VisitorException { //TODO: handle split and connect try { // The first vertex (prevOp) loads the left table and sends sample of join keys to // vertex 2 (sampler vertex) and all data to vertex 3 (partition vertex) via 1-1 edge // LR that transfers loaded input to partition vertex POLocalRearrangeTez lrTez = new POLocalRearrangeTez(OperatorKey.genOpKey(scope)); // LR that broadcasts sampled input to sampling aggregation vertex POLocalRearrangeTez lrTezSample = localRearrangeFactory.create(LocalRearrangeType.NULL); int sampleRate = POPoissonSample.DEFAULT_SAMPLE_RATE; if (pigProperties.containsKey(PigConfiguration.PIG_POISSON_SAMPLER_SAMPLE_RATE)) { sampleRate = Integer.valueOf(pigProperties.getProperty(PigConfiguration.PIG_POISSON_SAMPLER_SAMPLE_RATE)); } float heapPerc = PartitionSkewedKeys.DEFAULT_PERCENT_MEMUSAGE; if (pigProperties.containsKey(PigConfiguration.PIG_SKEWEDJOIN_REDUCE_MEMUSAGE)) { heapPerc = Float.valueOf(pigProperties.getProperty(PigConfiguration.PIG_SKEWEDJOIN_REDUCE_MEMUSAGE)); } long totalMemory = -1; if (pigProperties.containsKey(PigConfiguration.PIG_SKEWEDJOIN_REDUCE_MEM)) { totalMemory = Long.valueOf(pigProperties.getProperty(PigConfiguration.PIG_SKEWEDJOIN_REDUCE_MEM)); } POPoissonSample poSample = new POPoissonSample(new OperatorKey(scope,nig.getNextNodeId(scope)), -1, sampleRate, heapPerc, totalMemory); TezOperator samplerOper = compiledInputs[0]; boolean writeDataForPartitioner = shouldWriteDataForPartitioner(samplerOper); PhysicalPlan partitionerPlan = null; if (writeDataForPartitioner) { samplerOper.plan.addAsLeaf(lrTez); } else { partitionerPlan = samplerOper.plan.clone(); partitionerPlan.addAsLeaf(lrTez); } samplerOper.plan.addAsLeaf(poSample); samplerOper.markSampler(); MultiMap<PhysicalOperator, PhysicalPlan> joinPlans = op.getJoinPlans(); List<PhysicalOperator> l = plan.getPredecessors(op); List<PhysicalPlan> groups = joinPlans.get(l.get(0)); List<Boolean> ascCol = new ArrayList<Boolean>(); for (int i=0; i< groups.size(); i++) { ascCol.add(false); } // Set up transform plan to get keys and memory size of input // tuples. It first adds all the plans to get key columns. List<PhysicalPlan> transformPlans = new ArrayList<PhysicalPlan>(); transformPlans.addAll(groups); // Then it adds a column for memory size POProject prjStar = new POProject(new OperatorKey(scope,nig.getNextNodeId(scope))); prjStar.setResultType(DataType.TUPLE); prjStar.setStar(true); List<PhysicalOperator> ufInps = new ArrayList<PhysicalOperator>(); ufInps.add(prjStar); PhysicalPlan ep = new PhysicalPlan(); POUserFunc uf = new POUserFunc(new OperatorKey(scope,nig.getNextNodeId(scope)), -1, ufInps, new FuncSpec(GetMemNumRows.class.getName(), (String[])null)); uf.setResultType(DataType.TUPLE); ep.add(uf); ep.add(prjStar); ep.connect(prjStar, uf); transformPlans.add(ep); List<Boolean> flat1 = new ArrayList<Boolean>(); List<PhysicalPlan> eps1 = new ArrayList<PhysicalPlan>(); for (int i=0; i<transformPlans.size(); i++) { eps1.add(transformPlans.get(i)); flat1.add(i == transformPlans.size() - 1 ? true : false); } // This foreach will pick the sort key columns from the POPoissonSample output POForEach nfe1 = new POForEach(new OperatorKey(scope,nig.getNextNodeId(scope)), -1, eps1, flat1); samplerOper.plan.addAsLeaf(nfe1); samplerOper.plan.addAsLeaf(lrTezSample); samplerOper.setClosed(true); int rp = op.getRequestedParallelism(); if (rp == -1) { rp = pigContext.defaultParallel; } POSort sort = new POSort(op.getOperatorKey(), rp, null, groups, ascCol, null); String per = pigProperties.getProperty("pig.skewedjoin.reduce.memusage", String.valueOf(PartitionSkewedKeys.DEFAULT_PERCENT_MEMUSAGE)); String mc = pigProperties.getProperty("pig.skewedjoin.reduce.maxtuple", "0"); Pair<TezOperator, Integer> sampleJobPair = getSamplingAggregationJob(sort, rp, null, PartitionSkewedKeysTez.class.getName(), new String[]{per, mc}); rp = sampleJobPair.second; TezOperator[] joinJobs = new TezOperator[] {null, compiledInputs[1], null}; TezOperator[] joinInputs = new TezOperator[] {compiledInputs[0], compiledInputs[1]}; TezOperator[] rearrangeOutputs = new TezOperator[2]; compiledInputs = new TezOperator[] {joinInputs[0]}; // Add a partitioner vertex that partitions the data based on the quantiles from sample vertex. curTezOp = getTezOp(); tezPlan.add(curTezOp); joinJobs[0] = curTezOp; try { lrTez.setIndex(0); } catch (ExecException e) { int errCode = 2058; String msg = "Unable to set index on newly created POLocalRearrange."; throw new PlanException(msg, errCode, PigException.BUG, e); } // Check the type of group keys, if there are more than one field, the key is TUPLE. byte type = DataType.TUPLE; if (groups.size() == 1) { type = groups.get(0).getLeaves().get(0).getResultType(); } lrTez.setKeyType(type); lrTez.setPlans(groups); lrTez.setResultType(DataType.TUPLE); POLocalRearrangeTez partitionerLR = null; if (!writeDataForPartitioner) { // Read input from hdfs again joinJobs[0].plan = partitionerPlan; partitionerLR = lrTez; lrTez.setSkewedJoin(true); } else { // Add a POIdentityInOutTez which just passes data through from sampler vertex partitionerLR = new POIdentityInOutTez( OperatorKey.genOpKey(scope), lrTez, samplerOper.getOperatorKey().toString()); partitionerLR.setSkewedJoin(true); joinJobs[0].plan.addAsLeaf(partitionerLR); // Connect the sampler vertex to the partitioner vertex lrTez.setOutputKey(joinJobs[0].getOperatorKey().toString()); TezEdgeDescriptor edge = TezCompilerUtil.connect(tezPlan, samplerOper, joinJobs[0]); // TODO: PIG-3775 unsorted shuffle edge.dataMovementType = DataMovementType.ONE_TO_ONE; edge.outputClassName = UnorderedKVOutput.class.getName(); edge.inputClassName = UnorderedKVInput.class.getName(); // If prevOp.requestedParallelism changes based on no. of input splits // it will reflect for joinJobs[0] so that 1-1 edge will work. joinJobs[0].setRequestedParallelismByReference(samplerOper); } joinJobs[0].setClosed(true); joinJobs[0].markSampleBasedPartitioner(); joinJobs[0].setUseMRMapSettings(samplerOper.isUseMRMapSettings()); rearrangeOutputs[0] = joinJobs[0]; compiledInputs = new TezOperator[] {joinInputs[1]}; // Run POPartitionRearrange for second join table. Note we set the // parallelism of POPartitionRearrange to -1, so its parallelism // will be determined by the size of streaming table. POPartitionRearrangeTez pr = new POPartitionRearrangeTez(OperatorKey.genOpKey(scope)); try { pr.setIndex(1); } catch (ExecException e) { int errCode = 2058; String msg = "Unable to set index on newly created POPartitionRearrange."; throw new PlanException(msg, errCode, PigException.BUG, e); } groups = joinPlans.get(l.get(1)); pr.setPlans(groups); pr.setKeyType(type); pr.setSkewedJoin(true); pr.setResultType(DataType.TUPLE); joinJobs[1].plan.addAsLeaf(pr); joinJobs[1].setClosed(true); rearrangeOutputs[1] = joinJobs[1]; compiledInputs = rearrangeOutputs; // Create POGlobalRearrange POGlobalRearrange gr = new POGlobalRearrange(OperatorKey.genOpKey(scope), rp); // Skewed join has its own special partitioner gr.setResultType(DataType.TUPLE); gr.visit(this); joinJobs[2] = curTezOp; joinJobs[2].setRequestedParallelism(rp); compiledInputs = new TezOperator[] {joinJobs[2]}; // Create POPakcage POPackage pkg = getPackage(2, type); pkg.setResultType(DataType.TUPLE); boolean [] inner = op.getInnerFlags(); pkg.getPkgr().setInner(inner); pkg.visit(this); compiledInputs = new TezOperator[] {curTezOp}; // Create POForEach List<PhysicalPlan> eps = new ArrayList<PhysicalPlan>(); List<Boolean> flat = new ArrayList<Boolean>(); boolean containsRightOuter = false; // Add corresponding POProjects for (int i=0; i < 2; i++) { ep = new PhysicalPlan(); POProject prj = new POProject(OperatorKey.genOpKey(scope)); prj.setColumn(i+1); prj.setOverloaded(false); prj.setResultType(DataType.BAG); ep.add(prj); eps.add(ep); if (!inner[i]) { // Add an empty bag for outer join. For right outer, add IsFirstReduceOfKeyTez UDF as well if (i == 0) { containsRightOuter = true; CompilerUtils.addEmptyBagOuterJoin(ep, op.getSchema(i), true, IsFirstReduceOfKeyTez.class.getName()); } else { CompilerUtils.addEmptyBagOuterJoin(ep, op.getSchema(i), false, IsFirstReduceOfKeyTez.class.getName()); } } flat.add(true); } POForEach fe = new POForEach(OperatorKey.genOpKey(scope), -1, eps, flat); fe.setResultType(DataType.TUPLE); fe.visit(this); // Connect vertices lrTezSample.setOutputKey(sampleJobPair.first.getOperatorKey().toString()); partitionerLR.setOutputKey(joinJobs[2].getOperatorKey().toString()); pr.setOutputKey(joinJobs[2].getOperatorKey().toString()); TezCompilerUtil.connect(tezPlan, samplerOper, sampleJobPair.first); POValueOutputTez sampleOut = (POValueOutputTez) sampleJobPair.first.plan.getLeaves().get(0); for (int i = 0; i <= 2; i++) { if (i != 2 || containsRightOuter) { // We need to send sample to left relation partitioner vertex, right relation load vertex, // and join vertex (IsFirstReduceOfKey in join vertex need sample file as well) joinJobs[i].setSampleOperator(sampleJobPair.first); // Configure broadcast edges for distribution map TezEdgeDescriptor edge = TezCompilerUtil.connect(tezPlan, sampleJobPair.first, joinJobs[i]); TezCompilerUtil.configureValueOnlyTupleOutput(edge, DataMovementType.BROADCAST); sampleOut.addOutputKey(joinJobs[i].getOperatorKey().toString()); } // Configure skewed partitioner for join if (i != 2) { TezEdgeDescriptor edge = joinJobs[2].inEdges.get(joinJobs[i].getOperatorKey()); edge.partitionerClass = SkewedPartitionerTez.class; } } joinJobs[2].markSkewedJoin(); sampleJobPair.first.setSortOperator(joinJobs[2]); if (rp == -1) { sampleJobPair.first.setNeedEstimatedQuantile(true); } phyToTezOpMap.put(op, curTezOp); } catch (Exception e) { int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } /** * Returns a temporary DFS Path * @return * @throws IOException */ public static FileSpec getTempFileSpec(PigContext pigContext) throws IOException { return new FileSpec(FileLocalizer.getTemporaryPath(pigContext).toString(), new FuncSpec(Utils.getTmpFileCompressorName(pigContext))); } private boolean shouldWriteDataForPartitioner(TezOperator samplerOper) { // If there are operators other than load and foreach (like filter // split, etc) in the plan, then process and write the data out // and save the cost of processing in the partitioner vertex // Else read from hdfs again save the IO cost of the extra write boolean writeDataForPartitioner = false; if (samplerOper.plan.getRoots().get(0) instanceof POLoad) { for (PhysicalOperator oper : samplerOper.plan) { if (oper instanceof POForEach) { continue; } else if (oper instanceof POLoad && oper.getInputs() == null) { // TODO: oper.getInputs() is not null in case of PONative and // clone needs to be fixed in that case. e2e test - Native_2. String loadFunc = ((POLoad) oper).getLoadFunc().getClass().getName(); // We do not want to read all data again from hbase/accumulo for sampling. if (readOnceLoadFuncs == null || !readOnceLoadFuncs.contains(loadFunc)) { continue; } } writeDataForPartitioner = true; break; } } else { writeDataForPartitioner = true; } return writeDataForPartitioner; } /** * Get LocalRearrange for POSort input */ private POLocalRearrangeTez getLocalRearrangeForSortInput(POSort sort, byte keyType, Pair<POProject, Byte>[] fields) throws PlanException { POLocalRearrangeTez lr = new POLocalRearrangeTez(OperatorKey.genOpKey(scope)); List<PhysicalPlan> eps = new ArrayList<PhysicalPlan>(); if (fields == null) { // This is project * PhysicalPlan ep = new PhysicalPlan(); POProject prj = new POProject(new OperatorKey(scope,nig.getNextNodeId(scope))); prj.setStar(true); prj.setOverloaded(false); prj.setResultType(DataType.TUPLE); ep.add(prj); eps.add(ep); } else { // Attach the sort plans to the local rearrange to get the // projection. eps.addAll(sort.getSortPlans()); } try { lr.setIndex(0); } catch (ExecException e) { int errCode = 2058; String msg = "Unable to set index on newly created POLocalRearrange."; throw new PlanException(msg, errCode, PigException.BUG, e); } lr.setKeyType((fields == null || fields.length>1) ? DataType.TUPLE : keyType); lr.setPlans(eps); lr.setResultType(DataType.TUPLE); lr.addOriginalLocation(sort.getAlias(), sort.getOriginalLocations()); return lr; } /** * Add a sampler to the sort input */ private POLocalRearrangeTez addSamplingToSortInput(POSort sort, TezOperator oper, byte keyType, Pair<POProject, Byte>[] fields) throws PlanException { POLocalRearrangeTez lrSample = localRearrangeFactory.create(LocalRearrangeType.NULL); if (!oper.isClosed()) { List<Boolean> flat1 = new ArrayList<Boolean>(); List<PhysicalPlan> eps1 = new ArrayList<PhysicalPlan>(); Pair<POProject, Byte>[] sortProjs = null; try{ sortProjs = getSortCols(sort.getSortPlans()); }catch(Exception e) { throw new RuntimeException(e); } // Set up the projections of the key columns if (sortProjs == null) { PhysicalPlan ep = new PhysicalPlan(); POProject prj = new POProject(new OperatorKey(scope, nig.getNextNodeId(scope))); prj.setStar(true); prj.setOverloaded(false); prj.setResultType(DataType.TUPLE); ep.add(prj); eps1.add(ep); flat1.add(false); } else { for (Pair<POProject, Byte> sortProj : sortProjs) { // Check for proj being null, null is used by getSortCols for a non POProject // operator. Since Order by does not allow expression operators, //it should never be set to null if(sortProj == null){ int errCode = 2174; String msg = "Internal exception. Could not create a sampler job"; throw new PlanException(msg, errCode, PigException.BUG); } PhysicalPlan ep = new PhysicalPlan(); POProject prj; try { prj = sortProj.first.clone(); } catch (CloneNotSupportedException e) { //should not get here throw new AssertionError( "Error cloning project caught exception" + e ); } ep.add(prj); eps1.add(ep); flat1.add(false); } } String numSamples = pigContext.getProperties().getProperty(PigConfiguration.PIG_RANDOM_SAMPLER_SAMPLE_SIZE, "100"); POReservoirSample poSample = new POReservoirSample(new OperatorKey(scope,nig.getNextNodeId(scope)), -1, null, Integer.parseInt(numSamples)); oper.plan.addAsLeaf(poSample); List<PhysicalPlan> sortPlans = sort.getSortPlans(); // Set up transform plan to get keys and memory size of input // tuples. It first adds all the plans to get key columns. List<PhysicalPlan> transformPlans = new ArrayList<PhysicalPlan>(); transformPlans.addAll(sortPlans); // Then it adds a column for memory size POProject prjStar = new POProject(new OperatorKey(scope,nig.getNextNodeId(scope))); prjStar.setResultType(DataType.TUPLE); prjStar.setStar(true); List<PhysicalOperator> ufInps = new ArrayList<PhysicalOperator>(); ufInps.add(prjStar); PhysicalPlan ep = new PhysicalPlan(); POUserFunc uf = new POUserFunc(new OperatorKey(scope,nig.getNextNodeId(scope)), -1, ufInps, new FuncSpec(GetMemNumRows.class.getName(), (String[])null)); uf.setResultType(DataType.TUPLE); ep.add(uf); ep.add(prjStar); ep.connect(prjStar, uf); transformPlans.add(ep); flat1 = new ArrayList<Boolean>(); eps1 = new ArrayList<PhysicalPlan>(); for (int i=0; i<transformPlans.size(); i++) { eps1.add(transformPlans.get(i)); if (i<sortPlans.size()) { flat1.add(false); } else { flat1.add(true); } } // This foreach will pick the sort key columns from the POPoissonSample output POForEach nfe1 = new POForEach(new OperatorKey(scope,nig.getNextNodeId(scope)), -1, eps1, flat1); oper.plan.addAsLeaf(nfe1); lrSample.setOutputKey(curTezOp.getOperatorKey().toString()); oper.plan.addAsLeaf(lrSample); } else { int errCode = 2022; String msg = "The current operator is closed. This is unexpected while compiling."; throw new PlanException(msg, errCode, PigException.BUG); } oper.markSampler(); return lrSample; } private Pair<TezOperator,Integer> getOrderbySamplingAggregationJob( POSort inpSort, int rp) throws PlanException, VisitorException, ExecException { POSort sort = new POSort(inpSort.getOperatorKey(), inpSort .getRequestedParallelism(), null, inpSort.getSortPlans(), inpSort.getMAscCols(), inpSort.getMSortFunc()); sort.addOriginalLocation(inpSort.getAlias(), inpSort.getOriginalLocations()); // Turn the asc/desc array into an array of strings so that we can pass it // to the FindQuantiles function. List<Boolean> ascCols = inpSort.getMAscCols(); String[] ascs = new String[ascCols.size()]; for (int i = 0; i < ascCols.size(); i++) ascs[i] = ascCols.get(i).toString(); // check if user defined comparator is used in the sort, if so // prepend the name of the comparator as the first fields in the // constructor args array to the FindQuantiles udf String[] ctorArgs = ascs; if(sort.isUDFComparatorUsed) { String userComparatorFuncSpec = sort.getMSortFunc().getFuncSpec().toString(); ctorArgs = new String[ascs.length + 1]; ctorArgs[0] = USER_COMPARATOR_MARKER + userComparatorFuncSpec; for(int j = 0; j < ascs.length; j++) { ctorArgs[j+1] = ascs[j]; } } return getSamplingAggregationJob(sort, rp, null, FindQuantilesTez.class.getName(), ctorArgs); } /** * Create a sampling job to collect statistics by sampling input data. The * sequence of operations is as following: * <li>Add an extra field "all" into the tuple </li> * <li>Package all tuples into one bag </li> * <li>Add constant field for number of reducers. </li> * <li>Sorting the bag </li> * <li>Invoke UDF with the number of reducers and the sorted bag.</li> * <li>Data generated by UDF is transferred via a broadcast edge.</li> * * @param sort the POSort operator used to sort the bag * @param rp configured parallemism * @param sortKeyPlans PhysicalPlans to be set into POSort operator to get sorting keys * @param udfClassName the class name of UDF * @param udfArgs the arguments of UDF * @return pair<tezoperator[],integer> * @throws PlanException * @throws VisitorException * @throws ExecException */ private Pair<TezOperator,Integer> getSamplingAggregationJob(POSort sort, int rp, List<PhysicalPlan> sortKeyPlans, String udfClassName, String[] udfArgs) throws PlanException, VisitorException, ExecException { TezOperator oper = getTezOp(); tezPlan.add(oper); POPackage pkg = getPackage(1, DataType.BYTEARRAY); oper.plan.add(pkg); // Lets start building the plan which will have the sort // for the foreach PhysicalPlan fe2Plan = new PhysicalPlan(); // Top level project which just projects the tuple which is coming // from the foreach after the package POProject topPrj = new POProject(new OperatorKey(scope,nig.getNextNodeId(scope))); topPrj.setColumn(1); topPrj.setResultType(DataType.BAG); topPrj.setOverloaded(true); fe2Plan.add(topPrj); // the projections which will form sort plans List<PhysicalPlan> nesSortPlanLst = new ArrayList<PhysicalPlan>(); if (sortKeyPlans != null) { for(int i=0; i<sortKeyPlans.size(); i++) { nesSortPlanLst.add(sortKeyPlans.get(i)); } }else{ Pair<POProject, Byte>[] sortProjs = null; try{ sortProjs = getSortCols(sort.getSortPlans()); }catch(Exception e) { throw new RuntimeException(e); } // Set up the projections of the key columns if (sortProjs == null) { PhysicalPlan ep = new PhysicalPlan(); POProject prj = new POProject(new OperatorKey(scope, nig.getNextNodeId(scope))); prj.setStar(true); prj.setOverloaded(false); prj.setResultType(DataType.TUPLE); ep.add(prj); nesSortPlanLst.add(ep); } else { for (int i=0; i<sortProjs.length; i++) { POProject prj = new POProject(new OperatorKey(scope,nig.getNextNodeId(scope))); prj.setResultType(sortProjs[i].second); if(sortProjs[i].first != null && sortProjs[i].first.isProjectToEnd()){ if(i != sortProjs.length -1){ //project to end has to be the last sort column throw new AssertionError("Project-range to end (x..)" + " is supported in order-by only as last sort column"); } prj.setProjectToEnd(i); break; } else{ prj.setColumn(i); } prj.setOverloaded(false); PhysicalPlan ep = new PhysicalPlan(); ep.add(prj); nesSortPlanLst.add(ep); } } } sort.setSortPlans(nesSortPlanLst); sort.setResultType(DataType.BAG); fe2Plan.add(sort); fe2Plan.connect(topPrj, sort); // The plan which will have a constant representing the // degree of parallelism for the final order by map-reduce job // this will either come from a "order by parallel x" in the script // or will be the default number of reducers for the cluster if // "parallel x" is not used in the script PhysicalPlan rpep = new PhysicalPlan(); ConstantExpression rpce = new ConstantExpression(new OperatorKey(scope,nig.getNextNodeId(scope))); rpce.setRequestedParallelism(rp); // We temporarily set it to rp and will adjust it at runtime, because the final degree of parallelism // is unknown until we are ready to submit it. See PIG-2779. rpce.setValue(rp); rpce.setResultType(DataType.INTEGER); rpep.add(rpce); List<PhysicalPlan> genEps = new ArrayList<PhysicalPlan>(); genEps.add(rpep); genEps.add(fe2Plan); List<Boolean> flattened2 = new ArrayList<Boolean>(); flattened2.add(false); flattened2.add(false); POForEach nfe2 = new POForEach(new OperatorKey(scope,nig.getNextNodeId(scope)),-1, genEps, flattened2); oper.plan.add(nfe2); oper.plan.connect(pkg, nfe2); // Let's connect the output from the foreach containing // number of quantiles and the sorted bag of samples to // another foreach with the FindQuantiles udf. The input // to the FindQuantiles udf is a project(*) which takes the // foreach input and gives it to the udf PhysicalPlan ep4 = new PhysicalPlan(); POProject prjStar4 = new POProject(new OperatorKey(scope,nig.getNextNodeId(scope))); prjStar4.setResultType(DataType.TUPLE); prjStar4.setStar(true); ep4.add(prjStar4); List<PhysicalOperator> ufInps = new ArrayList<PhysicalOperator>(); ufInps.add(prjStar4); POUserFunc uf = new POUserFunc(new OperatorKey(scope,nig.getNextNodeId(scope)), -1, ufInps, new FuncSpec(udfClassName, udfArgs)); ep4.add(uf); ep4.connect(prjStar4, uf); List<PhysicalPlan> ep4s = new ArrayList<PhysicalPlan>(); ep4s.add(ep4); List<Boolean> flattened3 = new ArrayList<Boolean>(); flattened3.add(false); POForEach nfe3 = new POForEach(new OperatorKey(scope,nig.getNextNodeId(scope)), -1, ep4s, flattened3); oper.plan.add(nfe3); oper.plan.connect(nfe2, nfe3); POValueOutputTez sampleOut = new POValueOutputTez(OperatorKey.genOpKey(scope)); oper.plan.add(sampleOut); oper.plan.connect(nfe3, sampleOut); oper.setClosed(true); oper.setRequestedParallelism(1); oper.setDontEstimateParallelism(true); oper.markSampleAggregation(); return new Pair<TezOperator, Integer>(oper, rp); } private static class FindKeyTypeVisitor extends PhyPlanVisitor { byte keyType = DataType.UNKNOWN; FindKeyTypeVisitor(PhysicalPlan plan) { super(plan, new DepthFirstWalker<PhysicalOperator, PhysicalPlan>(plan)); } @Override public void visitProject(POProject p) throws VisitorException { keyType = p.getResultType(); } } public static Pair<POProject,Byte> [] getSortCols(List<PhysicalPlan> plans) throws PlanException { if(plans!=null){ @SuppressWarnings("unchecked") Pair<POProject,Byte>[] ret = new Pair[plans.size()]; int i=-1; for (PhysicalPlan plan : plans) { PhysicalOperator op = plan.getLeaves().get(0); POProject proj; if (op instanceof POProject) { if (((POProject)op).isStar()) return null; proj = (POProject)op; } else { proj = null; } byte type = op.getResultType(); ret[++i] = new Pair<POProject, Byte>(proj, type); } return ret; } int errCode = 2026; String msg = "No expression plan found in POSort."; throw new PlanException(msg, errCode, PigException.BUG); } private TezOperator[] getSortJobs( TezOperator inputOper, PhysicalPlan partitionerPlan, POLocalRearrangeTez inputOperRearrange, POSort sort, byte keyType, Pair<POProject, Byte>[] fields) throws PlanException{ TezOperator[] opers = new TezOperator[2]; // Partitioner Vertex TezOperator oper1 = getTezOp(); tezPlan.add(oper1); opers[0] = oper1; POLocalRearrangeTez partitionerLR = null; if (partitionerPlan != null) { // Read from hdfs again oper1.plan = partitionerPlan; partitionerLR = inputOperRearrange; } else { partitionerLR = new POIdentityInOutTez( OperatorKey.genOpKey(scope), inputOperRearrange, inputOper.getOperatorKey().toString()); oper1.plan.addAsLeaf(partitionerLR); } oper1.setClosed(true); oper1.markSampleBasedPartitioner(); // Global Sort Vertex TezOperator oper2 = getTezOp(); partitionerLR.setOutputKey(oper2.getOperatorKey().toString()); oper2.markGlobalSort(); opers[1] = oper2; tezPlan.add(oper2); long limit = sort.getLimit(); boolean[] sortOrder; List<Boolean> sortOrderList = sort.getMAscCols(); if(sortOrderList != null) { sortOrder = new boolean[sortOrderList.size()]; for(int i = 0; i < sortOrderList.size(); ++i) { sortOrder[i] = sortOrderList.get(i); } oper2.setSortOrder(sortOrder); } if (limit!=-1) { POPackage pkg_c = new POPackage(OperatorKey.genOpKey(scope)); pkg_c.setPkgr(new LitePackager()); pkg_c.getPkgr().setKeyType((fields == null || fields.length > 1) ? DataType.TUPLE : keyType); pkg_c.setNumInps(1); oper2.inEdges.put(oper1.getOperatorKey(), new TezEdgeDescriptor()); PhysicalPlan combinePlan = oper2.inEdges.get(oper1.getOperatorKey()).combinePlan; combinePlan.add(pkg_c); List<PhysicalPlan> eps_c1 = new ArrayList<PhysicalPlan>(); List<Boolean> flat_c1 = new ArrayList<Boolean>(); PhysicalPlan ep_c1 = new PhysicalPlan(); POProject prj_c1 = new POProject(new OperatorKey(scope,nig.getNextNodeId(scope))); prj_c1.setColumn(1); prj_c1.setOverloaded(false); prj_c1.setResultType(DataType.BAG); ep_c1.add(prj_c1); eps_c1.add(ep_c1); flat_c1.add(true); POForEach fe_c1 = new POForEach(new OperatorKey(scope,nig.getNextNodeId(scope)), -1, eps_c1, flat_c1); fe_c1.setResultType(DataType.TUPLE); combinePlan.addAsLeaf(fe_c1); POLimit pLimit = new POLimit(new OperatorKey(scope,nig.getNextNodeId(scope))); pLimit.setLimit(limit); combinePlan.addAsLeaf(pLimit); List<PhysicalPlan> eps_c2 = new ArrayList<PhysicalPlan>(); eps_c2.addAll(sort.getSortPlans()); POLocalRearrangeTez lr_c2 = new POLocalRearrangeTez(new OperatorKey(scope,nig.getNextNodeId(scope))); lr_c2.setOutputKey(oper2.getOperatorKey().toString()); try { lr_c2.setIndex(0); } catch (ExecException e) { int errCode = 2058; String msg = "Unable to set index on newly created POLocalRearrange."; throw new PlanException(msg, errCode, PigException.BUG, e); } lr_c2.setKeyType((fields.length>1) ? DataType.TUPLE : keyType); lr_c2.setPlans(eps_c2); lr_c2.setResultType(DataType.TUPLE); combinePlan.addAsLeaf(lr_c2); } POPackage pkg = new POPackage(OperatorKey.genOpKey(scope)); pkg.setPkgr(new LitePackager()); pkg.getPkgr().setKeyType((fields == null || fields.length > 1) ? DataType.TUPLE : keyType); pkg.setNumInps(1); oper2.plan.add(pkg); PhysicalPlan ep = new PhysicalPlan(); POProject prj = new POProject(new OperatorKey(scope,nig.getNextNodeId(scope))); prj.setColumn(1); prj.setOverloaded(false); prj.setResultType(DataType.BAG); ep.add(prj); List<PhysicalPlan> eps2 = new ArrayList<PhysicalPlan>(); eps2.add(ep); List<Boolean> flattened = new ArrayList<Boolean>(); flattened.add(true); POForEach nfe1 = new POForEach(new OperatorKey(scope,nig.getNextNodeId(scope)),-1,eps2,flattened); oper2.plan.add(nfe1); oper2.plan.connect(pkg, nfe1); if (limit!=-1) { POLimit pLimit2 = new POLimit(new OperatorKey(scope,nig.getNextNodeId(scope))); pLimit2.setLimit(limit); oper2.plan.addAsLeaf(pLimit2); } return opers; } @Override public void visitSort(POSort op) throws VisitorException { try{ if (compiledInputs.length > 1) { int errCode = 2023; String msg = "Received a multi input plan when expecting only a single input one."; throw new PlanException(msg, errCode, PigException.BUG); } Pair<POProject, Byte>[] fields = getSortCols(op.getSortPlans()); byte keyType = DataType.UNKNOWN; try { FindKeyTypeVisitor fktv = new FindKeyTypeVisitor(op.getSortPlans().get(0)); fktv.visit(); keyType = fktv.keyType; } catch (VisitorException ve) { int errCode = 2035; String msg = "Internal error. Could not compute key type of sort operator."; throw new PlanException(msg, errCode, PigException.BUG, ve); } TezOperator samplerOper = compiledInputs[0]; boolean writeDataForPartitioner = shouldWriteDataForPartitioner(samplerOper); POLocalRearrangeTez lr = getLocalRearrangeForSortInput(op, keyType, fields); PhysicalPlan partitionerPlan = null; if (writeDataForPartitioner) { samplerOper.plan.addAsLeaf(lr); } else { partitionerPlan = samplerOper.plan.clone(); partitionerPlan.addAsLeaf(lr); } // if rp is still -1, let it be, TezParallelismEstimator will set it to an estimated rp int rp = op.getRequestedParallelism(); if (rp == -1) { rp = pigContext.defaultParallel; } // Add sampling to sort input. Create a sample aggregation operator and connect both POLocalRearrangeTez lrSample = addSamplingToSortInput(op, samplerOper, keyType, fields); Pair<TezOperator, Integer> quantJobParallelismPair = getOrderbySamplingAggregationJob(op, rp); TezCompilerUtil.connect(tezPlan, samplerOper, quantJobParallelismPair.first); // Create the partitioner and the global sort vertices TezOperator[] sortOpers = getSortJobs(samplerOper, partitionerPlan, lr, op, keyType, fields); sortOpers[0].setUseMRMapSettings(samplerOper.isUseMRMapSettings()); if (writeDataForPartitioner) { // Connect the sampler and partitioner vertex lr.setOutputKey(sortOpers[0].getOperatorKey().toString()); TezEdgeDescriptor edge = TezCompilerUtil.connect(tezPlan, samplerOper, sortOpers[0]); // Use 1-1 edge edge.dataMovementType = DataMovementType.ONE_TO_ONE; edge.outputClassName = UnorderedKVOutput.class.getName(); edge.inputClassName = UnorderedKVInput.class.getName(); // If prevOper.requestedParallelism changes based on no. of input splits // it will reflect for sortOpers[0] so that 1-1 edge will work. sortOpers[0].setRequestedParallelismByReference(samplerOper); } if (rp == -1) { quantJobParallelismPair.first.setNeedEstimatedQuantile(true); } quantJobParallelismPair.first.setSortOperator(sortOpers[1]); sortOpers[1].setRequestedParallelism(quantJobParallelismPair.second); // Broadcast the sample to Partitioner vertex TezEdgeDescriptor edge = TezCompilerUtil.connect(tezPlan, quantJobParallelismPair.first, sortOpers[0]); TezCompilerUtil.configureValueOnlyTupleOutput(edge, DataMovementType.BROADCAST); POValueOutputTez sampleOut = (POValueOutputTez)quantJobParallelismPair.first.plan.getLeaves().get(0); sampleOut.addOutputKey(sortOpers[0].getOperatorKey().toString()); sortOpers[0].setSampleOperator(quantJobParallelismPair.first); lrSample.setOutputKey(quantJobParallelismPair.first.getOperatorKey().toString()); // Connect the Partitioner and Global Sort vertex edge = TezCompilerUtil.connect(tezPlan, sortOpers[0], sortOpers[1]); edge.partitionerClass = WeightedRangePartitionerTez.class; curTezOp = sortOpers[1]; // TODO: Review sort udf // if(op.isUDFComparatorUsed){ // curTezOp.UDFs.add(op.getMSortFunc().getFuncSpec().toString()); // curTezOp.isUDFComparatorUsed = true; // } // If Order by followed by Limit and parallelism of order by is not 1 // add a new vertex for Limit with parallelism 1. // Equivalent of LimitAdjuster.java in MR if (op.isLimited() && rp != 1) { POValueOutputTez output = new POValueOutputTez(OperatorKey.genOpKey(scope)); output.copyAliasFrom(op); sortOpers[1].plan.addAsLeaf(output); TezOperator limitOper = getTezOp(); tezPlan.add(limitOper); curTezOp = limitOper; // Explicitly set the parallelism for the new vertex to 1. limitOper.setRequestedParallelism(1); limitOper.setDontEstimateParallelism(true); limitOper.markLimitAfterSort(); edge = TezCompilerUtil.connect(tezPlan, sortOpers[1], limitOper); // LIMIT in this case should be ordered. So we output unordered with key as task index // and on the input we use OrderedGroupedKVInput to do ordered merge to retain sorted order. output.addOutputKey(limitOper.getOperatorKey().toString()); output.setTaskIndexWithRecordIndexAsKey(true); edge = curTezOp.inEdges.get(sortOpers[1].getOperatorKey()); TezCompilerUtil.configureValueOnlyTupleOutput(edge, DataMovementType.SCATTER_GATHER); // POValueOutputTez will write key (task index, record index) in // sorted order. So using UnorderedKVOutput instead of OrderedPartitionedKVOutput. // But input needs to be merged in sorter order and requires OrderedGroupedKVInput edge.outputClassName = UnorderedKVOutput.class.getName(); edge.inputClassName = OrderedGroupedKVInput.class.getName(); edge.setIntermediateOutputKeyClass(TezCompilerUtil.TUPLE_CLASS); edge.setIntermediateOutputKeyComparatorClass(PigTupleWritableComparator.class.getName()); // Then add a POValueInputTez to the start of the new tezOp followed by a LIMIT POValueInputTez input = new POValueInputTez(OperatorKey.genOpKey(scope)); input.copyAliasFrom(op); input.setInputKey(sortOpers[1].getOperatorKey().toString()); curTezOp.plan.addAsLeaf(input); POLimit limit = new POLimit(OperatorKey.genOpKey(scope)); limit.setLimit(op.getLimit()); curTezOp.plan.addAsLeaf(limit); } phyToTezOpMap.put(op, curTezOp); }catch(Exception e){ int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } @Override public void visitSplit(POSplit op) throws VisitorException { try { TezOperator splitOp = curTezOp; POValueOutputTez output = null; if (splitsSeen.containsKey(op.getOperatorKey())) { splitOp = splitsSeen.get(op.getOperatorKey()); output = (POValueOutputTez)splitOp.plan.getLeaves().get(0); } else { splitsSeen.put(op.getOperatorKey(), splitOp); splitOp.setSplitter(true); phyToTezOpMap.put(op, splitOp); output = new POValueOutputTez(OperatorKey.genOpKey(scope)); output.copyAliasFrom(op); splitOp.plan.addAsLeaf(output); } curTezOp = getTezOp(); curTezOp.setSplitParent(splitOp.getOperatorKey()); tezPlan.add(curTezOp); output.addOutputKey(curTezOp.getOperatorKey().toString()); TezEdgeDescriptor edge = TezCompilerUtil.connect(tezPlan, splitOp, curTezOp); //TODO shared edge once support is available in Tez TezCompilerUtil.configureValueOnlyTupleOutput(edge, DataMovementType.ONE_TO_ONE); curTezOp.setRequestedParallelismByReference(splitOp); POValueInputTez input = new POValueInputTez(OperatorKey.genOpKey(scope)); input.copyAliasFrom(op); input.setInputKey(splitOp.getOperatorKey().toString()); curTezOp.plan.addAsLeaf(input); } catch (Exception e) { int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } @Override public void visitStore(POStore op) throws VisitorException { try { POStoreTez store = new POStoreTez(op); nonBlocking(store); phyToTezOpMap.put(op, curTezOp); if (store.getSFile()!=null && store.getSFile().getFuncSpec()!=null) curTezOp.UDFs.add(store.getSFile().getFuncSpec().toString()); } catch (Exception e) { int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } @Override public void visitStream(POStream op) throws VisitorException { try { nonBlocking(op); phyToTezOpMap.put(op, curTezOp); } catch(Exception e) { int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } @Override public void visitUnion(POUnion op) throws VisitorException { try { // Without VertexGroup (UnionOptimizer), there is an extra union vertex // which unions input from the two predecessor vertices TezOperator unionTezOp = getTezOp(); tezPlan.add(unionTezOp); unionTezOp.markUnion(); unionTezOp.setRequestedParallelism(op.getRequestedParallelism()); POShuffledValueInputTez unionInput = new POShuffledValueInputTez(OperatorKey.genOpKey(scope)); unionTezOp.plan.addAsLeaf(unionInput); POValueOutputTez[] outputs = new POValueOutputTez[compiledInputs.length]; for (int i = 0; i < compiledInputs.length; i++) { TezOperator prevTezOp = compiledInputs[i]; // Some predecessors of union need not be part of the union (For eg: replicated join). // So mark predecessors that are input to the union operation. unionTezOp.addUnionPredecessor(prevTezOp.getOperatorKey()); TezEdgeDescriptor edge = TezCompilerUtil.connect(tezPlan, prevTezOp, unionTezOp); TezCompilerUtil.configureValueOnlyTupleOutput(edge, DataMovementType.SCATTER_GATHER); outputs[i] = new POValueOutputTez(OperatorKey.genOpKey(scope)); outputs[i].addOutputKey(unionTezOp.getOperatorKey().toString()); unionInput.addInputKey(prevTezOp.getOperatorKey().toString()); prevTezOp.plan.addAsLeaf(outputs[i]); prevTezOp.setClosed(true); if (prevTezOp.isUseMRMapSettings()) { unionTezOp.setUseMRMapSettings(true); } } curTezOp = unionTezOp; phyToTezOpMap.put(op, curTezOp); } catch (Exception e) { int errCode = 2034; String msg = "Error compiling operator " + op.getClass().getSimpleName(); throw new TezCompilerException(msg, errCode, PigException.BUG, e); } } /** * Returns a POPackage with default packager. This method shouldn't be used * if special packager such as LitePackager and CombinerPackager is needed. */ private POPackage getPackage(int numOfInputs, byte keyType) { // The default value of boolean is false boolean[] inner = new boolean[numOfInputs]; POPackage pkg = new POPackage(OperatorKey.genOpKey(scope)); pkg.getPkgr().setInner(inner); pkg.getPkgr().setKeyType(keyType); pkg.setNumInps(numOfInputs); return pkg; } private TezOperator getTezOp() { return new TezOperator(OperatorKey.genOpKey(scope)); } }