/* * Copyright 1999-2015 dangdang.com. * <p> * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. * </p> */ package com.dangdang.ddframe.job.lite.integrate.fixture.dataflow; import com.dangdang.ddframe.job.api.ShardingContext; import com.dangdang.ddframe.job.api.dataflow.DataflowJob; import java.util.Arrays; import java.util.List; import java.util.Set; import java.util.concurrent.CopyOnWriteArraySet; public final class OneOffDataflowElasticJob implements DataflowJob<String> { private static volatile Set<String> processedData = new CopyOnWriteArraySet<>(); private static volatile List<String> result = Arrays.asList("data0", "data1", "data2", "data3", "data4", "data5", "data6", "data7", "data8", "data9"); @Override public List<String> fetchData(final ShardingContext shardingContext) { return result; } @Override public void processData(final ShardingContext shardingContext, final List<String> data) { for (String each : data) { processedData.add(each); } } public static boolean isCompleted() { return result.size() == processedData.size(); } public static void reset() { processedData.clear(); } }