/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, * software distributed under the License is distributed on an * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. */ package org.apache.streams.kafka; import org.apache.streams.config.ComponentConfigurator; import org.apache.streams.config.StreamsConfigurator; import org.apache.streams.core.StreamsDatum; import org.apache.streams.core.StreamsPersistWriter; import org.apache.streams.util.GuidUtils; import com.fasterxml.jackson.core.JsonProcessingException; import com.fasterxml.jackson.databind.ObjectMapper; import kafka.javaapi.producer.Producer; import kafka.producer.KeyedMessage; import kafka.producer.ProducerConfig; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import java.io.Serializable; import java.util.Properties; import java.util.Queue; import java.util.concurrent.ConcurrentLinkedQueue; /** * KafkaPersistWriter writes documents to kafka. */ public class KafkaPersistWriter implements StreamsPersistWriter, Serializable, Runnable { public static final String STREAMS_ID = "KafkaPersistWriter"; private static final Logger LOGGER = LoggerFactory.getLogger(KafkaPersistWriter.class); protected volatile Queue<StreamsDatum> persistQueue; private ObjectMapper mapper = new ObjectMapper(); private KafkaConfiguration config; private Producer<String, String> producer; /** * KafkaPersistWriter constructor - resolves KafkaConfiguration from JVM 'kafka'. */ public KafkaPersistWriter() { this.config = new ComponentConfigurator<>(KafkaConfiguration.class) .detectConfiguration(StreamsConfigurator.getConfig().getConfig("kafka")); this.persistQueue = new ConcurrentLinkedQueue<>(); } /** * KafkaPersistWriter constructor - uses supplied persistQueue. */ public KafkaPersistWriter(Queue<StreamsDatum> persistQueue) { this.config = new ComponentConfigurator<>(KafkaConfiguration.class) .detectConfiguration(StreamsConfigurator.getConfig().getConfig("kafka")); this.persistQueue = persistQueue; } public void setConfig(KafkaConfiguration config) { this.config = config; } /** * run persist writer thread */ public void start() { Properties props = new Properties(); props.put("metadata.broker.list", config.getBrokerlist()); props.put("serializer.class", "kafka.serializer.StringEncoder"); props.put("partitioner.class", "org.apache.streams.kafka.StreamsPartitioner"); props.put("request.required.acks", "1"); ProducerConfig config = new ProducerConfig(props); producer = new Producer<>(config); new Thread(new KafkaPersistWriterTask(this)).start(); } public void stop() { producer.close(); } public void setPersistQueue(Queue<StreamsDatum> persistQueue) { this.persistQueue = persistQueue; } public Queue<StreamsDatum> getPersistQueue() { return this.persistQueue; } @Override public String getId() { return STREAMS_ID; } @Override public void write(StreamsDatum entry) { try { String text = mapper.writeValueAsString(entry); String hash = GuidUtils.generateGuid(text); KeyedMessage<String, String> data = new KeyedMessage<>(config.getTopic(), hash, text); producer.send(data); } catch (JsonProcessingException ex) { LOGGER.warn("save: {}", ex); } } @Override public void run() { start(); } @Override public void prepare(Object configurationObject) { start(); } @Override public void cleanUp() { stop(); } }