/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.hive.hcatalog.streaming.mutate; import java.util.List; import org.apache.hive.hcatalog.streaming.mutate.client.MutatorClient; import org.apache.hive.hcatalog.streaming.mutate.client.MutatorClientBuilder; import org.apache.hive.hcatalog.streaming.mutate.client.AcidTable; import org.apache.hive.hcatalog.streaming.mutate.client.Transaction; import org.apache.hive.hcatalog.streaming.mutate.worker.BucketIdResolver; import org.apache.hive.hcatalog.streaming.mutate.worker.MutatorCoordinator; import org.apache.hive.hcatalog.streaming.mutate.worker.MutatorCoordinatorBuilder; import org.apache.hive.hcatalog.streaming.mutate.worker.MutatorFactory; public class ExampleUseCase { private String metaStoreUri; private String databaseName; private String tableName; private boolean createPartitions = true; private List<String> partitionValues1, partitionValues2, partitionValues3; private Object record1, record2, record3; private MutatorFactory mutatorFactory; /* This is an illustration, not a functioning example. */ public void example() throws Exception { // CLIENT/TOOL END // // Singleton instance in the job client // Create a client to manage our transaction MutatorClient client = new MutatorClientBuilder() .addSinkTable(databaseName, tableName, createPartitions) .metaStoreUri(metaStoreUri) .build(); // Get the transaction Transaction transaction = client.newTransaction(); // Get serializable details of the destination tables List<AcidTable> tables = client.getTables(); transaction.begin(); // CLUSTER / WORKER END // // Job submitted to the cluster // BucketIdResolver bucketIdResolver = mutatorFactory.newBucketIdResolver(tables.get(0).getTotalBuckets()); record1 = bucketIdResolver.attachBucketIdToRecord(record1); // -------------------------------------------------------------- // DATA SHOULD GET SORTED BY YOUR ETL/MERGE PROCESS HERE // // Group the data by (partitionValues, ROW__ID.bucketId) // Order the groups by (ROW__ID.lastTransactionId, ROW__ID.rowId) // -------------------------------------------------------------- // One of these runs at the output of each reducer // MutatorCoordinator coordinator = new MutatorCoordinatorBuilder() .metaStoreUri(metaStoreUri) .table(tables.get(0)) .mutatorFactory(mutatorFactory) .build(); coordinator.insert(partitionValues1, record1); coordinator.update(partitionValues2, record2); coordinator.delete(partitionValues3, record3); coordinator.close(); // CLIENT/TOOL END // // The tasks have completed, control is back at the tool transaction.commit(); client.close(); } }