/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.hadoop.hive.druid.io; import org.apache.hadoop.fs.Path; import org.junit.Assert; import org.junit.Test; import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.DataInputStream; import java.io.DataOutput; import java.io.DataOutputStream; import java.io.IOException; public class TestHiveDruidSplit { @Test public void testSerDeser() throws IOException { HiveDruidSplit hiveDruidSplit = new HiveDruidSplit("query string", new Path("test-path"), new String []{"host:8080", "host2:8090"}); ByteArrayOutputStream byteArrayOutputStream = new ByteArrayOutputStream(); DataOutput dataOutput = new DataOutputStream(byteArrayOutputStream); hiveDruidSplit.write(dataOutput); ByteArrayInputStream byteArrayInputStream = new ByteArrayInputStream(byteArrayOutputStream.toByteArray()); HiveDruidSplit actualHiveDruidSplit = new HiveDruidSplit(); actualHiveDruidSplit.readFields(new DataInputStream(byteArrayInputStream)); Assert.assertEquals(actualHiveDruidSplit.getDruidQuery(), "query string"); Assert.assertArrayEquals(actualHiveDruidSplit.getLocations(), new String []{"host:8080", "host2:8090"}); } }