/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package kafka.examples; import kafka.api.FetchRequest; import kafka.api.FetchRequestBuilder; import kafka.javaapi.FetchResponse; import kafka.javaapi.consumer.SimpleConsumer; import kafka.javaapi.message.ByteBufferMessageSet; import kafka.message.MessageAndOffset; import java.io.UnsupportedEncodingException; import java.nio.ByteBuffer; import java.util.Collections; import java.util.HashMap; import java.util.List; import java.util.Map; public class SimpleConsumerDemo { private static void printMessages(ByteBufferMessageSet messageSet) throws UnsupportedEncodingException { for (MessageAndOffset messageAndOffset : messageSet) { ByteBuffer payload = messageAndOffset.message().payload(); byte[] bytes = new byte[payload.limit()]; payload.get(bytes); System.out.println(new String(bytes, "UTF-8")); } } private static void generateData() { Producer producer2 = new Producer(KafkaProperties.TOPIC2, false); producer2.start(); Producer producer3 = new Producer(KafkaProperties.TOPIC3, false); producer3.start(); try { Thread.sleep(1000); } catch (InterruptedException e) { e.printStackTrace(); } } public static void main(String[] args) throws Exception { generateData(); SimpleConsumer simpleConsumer = new SimpleConsumer(KafkaProperties.KAFKA_SERVER_URL, KafkaProperties.KAFKA_SERVER_PORT, KafkaProperties.CONNECTION_TIMEOUT, KafkaProperties.KAFKA_PRODUCER_BUFFER_SIZE, KafkaProperties.CLIENT_ID); System.out.println("Testing single fetch"); FetchRequest req = new FetchRequestBuilder() .clientId(KafkaProperties.CLIENT_ID) .addFetch(KafkaProperties.TOPIC2, 0, 0L, 100) .build(); FetchResponse fetchResponse = simpleConsumer.fetch(req); printMessages(fetchResponse.messageSet(KafkaProperties.TOPIC2, 0)); System.out.println("Testing single multi-fetch"); Map<String, List<Integer>> topicMap = new HashMap<>(); topicMap.put(KafkaProperties.TOPIC2, Collections.singletonList(0)); topicMap.put(KafkaProperties.TOPIC3, Collections.singletonList(0)); req = new FetchRequestBuilder() .clientId(KafkaProperties.CLIENT_ID) .addFetch(KafkaProperties.TOPIC2, 0, 0L, 100) .addFetch(KafkaProperties.TOPIC3, 0, 0L, 100) .build(); fetchResponse = simpleConsumer.fetch(req); int fetchReq = 0; for (Map.Entry<String, List<Integer>> entry : topicMap.entrySet()) { String topic = entry.getKey(); for (Integer offset : entry.getValue()) { System.out.println("Response from fetch request no: " + ++fetchReq); printMessages(fetchResponse.messageSet(topic, offset)); } } } }