/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.kafka.connect.transforms; import org.apache.kafka.common.cache.Cache; import org.apache.kafka.common.cache.LRUCache; import org.apache.kafka.common.cache.SynchronizedCache; import org.apache.kafka.common.config.ConfigDef; import org.apache.kafka.common.config.ConfigException; import org.apache.kafka.connect.connector.ConnectRecord; import org.apache.kafka.connect.data.Field; import org.apache.kafka.connect.data.Schema; import org.apache.kafka.connect.data.SchemaBuilder; import org.apache.kafka.connect.data.Struct; import org.apache.kafka.connect.transforms.util.SchemaUtil; import org.apache.kafka.connect.transforms.util.SimpleConfig; import java.util.Collections; import java.util.HashMap; import java.util.List; import java.util.Map; import static org.apache.kafka.connect.transforms.util.Requirements.requireMap; import static org.apache.kafka.connect.transforms.util.Requirements.requireStruct; public abstract class ReplaceField<R extends ConnectRecord<R>> implements Transformation<R> { public static final String OVERVIEW_DOC = "Filter or rename fields." + "<p/>Use the concrete transformation type designed for the record key (<code>" + Key.class.getName() + "</code>) " + "or value (<code>" + Value.class.getName() + "</code>)."; interface ConfigName { String BLACKLIST = "blacklist"; String WHITELIST = "whitelist"; String RENAME = "renames"; } public static final ConfigDef CONFIG_DEF = new ConfigDef() .define(ConfigName.BLACKLIST, ConfigDef.Type.LIST, Collections.emptyList(), ConfigDef.Importance.MEDIUM, "Fields to exclude. This takes precedence over the whitelist.") .define(ConfigName.WHITELIST, ConfigDef.Type.LIST, Collections.emptyList(), ConfigDef.Importance.MEDIUM, "Fields to include. If specified, only these fields will be used.") .define(ConfigName.RENAME, ConfigDef.Type.LIST, Collections.emptyList(), new ConfigDef.Validator() { @Override public void ensureValid(String name, Object value) { parseRenameMappings((List<String>) value); } @Override public String toString() { return "list of colon-delimited pairs, e.g. <code>foo:bar,abc:xyz</code>"; } }, ConfigDef.Importance.MEDIUM, "Field rename mappings."); private static final String PURPOSE = "field replacement"; private List<String> blacklist; private List<String> whitelist; private Map<String, String> renames; private Map<String, String> reverseRenames; private Cache<Schema, Schema> schemaUpdateCache; @Override public void configure(Map<String, ?> configs) { final SimpleConfig config = new SimpleConfig(CONFIG_DEF, configs); blacklist = config.getList(ConfigName.BLACKLIST); whitelist = config.getList(ConfigName.WHITELIST); renames = parseRenameMappings(config.getList(ConfigName.RENAME)); reverseRenames = invert(renames); schemaUpdateCache = new SynchronizedCache<>(new LRUCache<Schema, Schema>(16)); } static Map<String, String> parseRenameMappings(List<String> mappings) { final Map<String, String> m = new HashMap<>(); for (String mapping : mappings) { final String[] parts = mapping.split(":"); if (parts.length != 2) { throw new ConfigException(ConfigName.RENAME, mappings, "Invalid rename mapping: " + mapping); } m.put(parts[0], parts[1]); } return m; } static Map<String, String> invert(Map<String, String> source) { final Map<String, String> m = new HashMap<>(); for (Map.Entry<String, String> e : source.entrySet()) { m.put(e.getValue(), e.getKey()); } return m; } boolean filter(String fieldName) { return !blacklist.contains(fieldName) && (whitelist.isEmpty() || whitelist.contains(fieldName)); } String renamed(String fieldName) { final String mapping = renames.get(fieldName); return mapping == null ? fieldName : mapping; } String reverseRenamed(String fieldName) { final String mapping = reverseRenames.get(fieldName); return mapping == null ? fieldName : mapping; } @Override public R apply(R record) { if (operatingSchema(record) == null) { return applySchemaless(record); } else { return applyWithSchema(record); } } private R applySchemaless(R record) { final Map<String, Object> value = requireMap(operatingValue(record), PURPOSE); final Map<String, Object> updatedValue = new HashMap<>(value.size()); for (Map.Entry<String, Object> e : value.entrySet()) { final String fieldName = e.getKey(); if (filter(fieldName)) { final Object fieldValue = e.getValue(); updatedValue.put(renamed(fieldName), fieldValue); } } return newRecord(record, null, updatedValue); } private R applyWithSchema(R record) { final Struct value = requireStruct(operatingValue(record), PURPOSE); Schema updatedSchema = schemaUpdateCache.get(value.schema()); if (updatedSchema == null) { updatedSchema = makeUpdatedSchema(value.schema()); schemaUpdateCache.put(value.schema(), updatedSchema); } final Struct updatedValue = new Struct(updatedSchema); for (Field field : updatedSchema.fields()) { final Object fieldValue = value.get(reverseRenamed(field.name())); updatedValue.put(field.name(), fieldValue); } return newRecord(record, updatedSchema, updatedValue); } private Schema makeUpdatedSchema(Schema schema) { final SchemaBuilder builder = SchemaUtil.copySchemaBasics(schema, SchemaBuilder.struct()); for (Field field : schema.fields()) { if (filter(field.name())) { builder.field(renamed(field.name()), field.schema()); } } return builder.build(); } @Override public ConfigDef config() { return CONFIG_DEF; } @Override public void close() { schemaUpdateCache = null; } protected abstract Schema operatingSchema(R record); protected abstract Object operatingValue(R record); protected abstract R newRecord(R record, Schema updatedSchema, Object updatedValue); public static class Key<R extends ConnectRecord<R>> extends ReplaceField<R> { @Override protected Schema operatingSchema(R record) { return record.keySchema(); } @Override protected Object operatingValue(R record) { return record.key(); } @Override protected R newRecord(R record, Schema updatedSchema, Object updatedValue) { return record.newRecord(record.topic(), record.kafkaPartition(), updatedSchema, updatedValue, record.valueSchema(), record.value(), record.timestamp()); } } public static class Value<R extends ConnectRecord<R>> extends ReplaceField<R> { @Override protected Schema operatingSchema(R record) { return record.valueSchema(); } @Override protected Object operatingValue(R record) { return record.value(); } @Override protected R newRecord(R record, Schema updatedSchema, Object updatedValue) { return record.newRecord(record.topic(), record.kafkaPartition(), record.keySchema(), record.key(), updatedSchema, updatedValue, record.timestamp()); } } }