/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.flink.streaming.connectors.elasticsearch.testutils;
import org.apache.flink.api.common.functions.RuntimeContext;
import org.apache.flink.api.java.tuple.Tuple2;
import org.apache.flink.streaming.api.functions.source.SourceFunction;
import org.apache.flink.streaming.connectors.elasticsearch.ElasticsearchSinkFunction;
import org.apache.flink.streaming.connectors.elasticsearch.RequestIndexer;
import org.elasticsearch.action.get.GetRequest;
import org.elasticsearch.action.get.GetResponse;
import org.elasticsearch.action.index.IndexRequest;
import org.elasticsearch.client.Client;
import org.junit.Assert;
import java.util.HashMap;
import java.util.Map;
/**
* This class contains utilities and a pre-defined source function and
* Elasticsearch Sink function used to simulate and verify data used in tests.
*/
public class SourceSinkDataTestKit {
private static final int NUM_ELEMENTS = 20;
private static final String DATA_PREFIX = "message #";
private static final String DATA_FIELD_NAME = "data";
private static final String TYPE_NAME = "flink-es-test-type";
/**
* A {@link SourceFunction} that generates the elements (id, "message #" + id) with id being 0 - 20.
*/
public static class TestDataSourceFunction implements SourceFunction<Tuple2<Integer, String>> {
private static final long serialVersionUID = 1L;
private volatile boolean running = true;
@Override
public void run(SourceFunction.SourceContext<Tuple2<Integer, String>> ctx) throws Exception {
for (int i = 0; i < NUM_ELEMENTS && running; i++) {
ctx.collect(Tuple2.of(i, DATA_PREFIX + i));
}
}
@Override
public void cancel() {
running = false;
}
}
/**
* A {@link ElasticsearchSinkFunction} that indexes each element it receives to a sepecified Elasticsearch index.
*/
public static class TestElasticsearchSinkFunction implements ElasticsearchSinkFunction<Tuple2<Integer, String>> {
private static final long serialVersionUID = 1L;
private final String index;
/**
* Create the sink function, specifying a target Elasticsearch index.
*
* @param index Name of the target Elasticsearch index.
*/
public TestElasticsearchSinkFunction(String index) {
this.index = index;
}
public IndexRequest createIndexRequest(Tuple2<Integer, String> element) {
Map<String, Object> json = new HashMap<>();
json.put(DATA_FIELD_NAME, element.f1);
return new IndexRequest(index, TYPE_NAME, element.f0.toString()).source(json);
}
@Override
public void process(Tuple2<Integer, String> element, RuntimeContext ctx, RequestIndexer indexer) {
indexer.add(createIndexRequest(element));
}
}
/**
* Verify the results in an Elasticsearch index. The results must first be produced into the index
* using a {@link TestElasticsearchSinkFunction};
*
* @param client The client to use to connect to Elasticsearch
* @param index The index to check
*/
public static void verifyProducedSinkData(Client client, String index) {
for (int i = 0; i < NUM_ELEMENTS; i++) {
GetResponse response = client.get(new GetRequest(index, TYPE_NAME, Integer.toString(i))).actionGet();
Assert.assertEquals(DATA_PREFIX + i, response.getSource().get(DATA_FIELD_NAME));
}
}
}