/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.falcon.regression;
import org.apache.commons.lang.StringUtils;
import org.apache.falcon.entity.v0.EntityType;
import org.apache.falcon.entity.v0.cluster.Interfacetype;
import org.apache.falcon.entity.v0.feed.ActionType;
import org.apache.falcon.entity.v0.feed.ClusterType;
import org.apache.falcon.regression.Entities.FeedMerlin;
import org.apache.falcon.regression.core.bundle.Bundle;
import org.apache.falcon.regression.core.enumsAndConstants.MerlinConstants;
import org.apache.falcon.regression.core.helpers.ColoHelper;
import org.apache.falcon.regression.core.util.AssertUtil;
import org.apache.falcon.regression.core.util.BundleUtil;
import org.apache.falcon.regression.core.util.HadoopUtil;
import org.apache.falcon.regression.core.util.InstanceUtil;
import org.apache.falcon.regression.core.util.MatrixUtil;
import org.apache.falcon.regression.core.util.OSUtil;
import org.apache.falcon.regression.core.util.OozieUtil;
import org.apache.falcon.regression.core.util.TimeUtil;
import org.apache.falcon.regression.core.util.Util;
import org.apache.falcon.regression.testHelper.BaseTestClass;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.ContentSummary;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.log4j.Logger;
import org.apache.oozie.client.CoordinatorAction;
import org.apache.oozie.client.OozieClient;
import org.joda.time.DateTime;
import org.joda.time.DateTimeZone;
import org.joda.time.format.DateTimeFormat;
import org.joda.time.format.DateTimeFormatter;
import org.testng.Assert;
import org.testng.annotations.AfterClass;
import org.testng.annotations.AfterMethod;
import org.testng.annotations.BeforeClass;
import org.testng.annotations.BeforeMethod;
import org.testng.annotations.DataProvider;
import org.testng.annotations.Test;
import javax.xml.bind.JAXBException;
import java.io.IOException;
import java.util.List;
import java.util.UUID;
/**
* Tests for operations with external file systems.
*/
@Test(groups = "embedded")
public class ExternalFSTest extends BaseTestClass{
public static final String WASB_END_POINT =
"wasb://" + MerlinConstants.WASB_CONTAINER + "@" + MerlinConstants.WASB_ACCOUNT;
private ColoHelper cluster = servers.get(0);
private FileSystem clusterFS = serverFS.get(0);
private OozieClient clusterOC = serverOC.get(0);
private FileSystem wasbFS;
private Bundle externalBundle;
private String baseTestDir = cleanAndGetTestDir();
private String sourcePath = baseTestDir + "/source";
private String baseWasbDir = "/falcon-regression/" + UUID.randomUUID().toString().split("-")[0];
private String testWasbTargetDir = baseWasbDir + '/'
+ UUID.randomUUID().toString().split("-")[0] + '/';
private static final Logger LOGGER = Logger.getLogger(ExternalFSTest.class);
@BeforeClass
public void setUpClass() throws IOException {
HadoopUtil.recreateDir(clusterFS, baseTestDir);
Configuration conf = new Configuration();
conf.set("fs.defaultFS", WASB_END_POINT);
conf.set("fs.azure.account.key." + MerlinConstants.WASB_ACCOUNT,
MerlinConstants.WASB_SECRET);
conf.setBoolean("fs.hdfs.impl.disable.cache", false);
wasbFS = FileSystem.get(conf);
LOGGER.info("creating base wasb dir" + baseWasbDir);
}
@BeforeMethod(alwaysRun = true)
public void setUp() throws JAXBException, IOException {
Bundle bundle = BundleUtil.readFeedReplicationBundle();
bundles[0] = new Bundle(bundle, cluster);
externalBundle = new Bundle(bundle, cluster);
bundles[0].generateUniqueBundle(this);
externalBundle.generateUniqueBundle(this);
LOGGER.info("checking wasb credentials with location: " + testWasbTargetDir);
wasbFS.create(new Path(testWasbTargetDir));
wasbFS.delete(new Path(testWasbTargetDir), true);
}
@AfterMethod
public void tearDown() throws IOException {
removeTestClassEntities();
wasbFS.delete(new Path(testWasbTargetDir), true);
}
@AfterClass(alwaysRun = true)
public void tearDownClass() throws IOException {
wasbFS.delete(new Path(baseWasbDir), true);
}
@Test(dataProvider = "getInvalidTargets")
public void invalidCredentialsExtFS(String endpoint) throws Exception {
bundles[0].setClusterInterface(Interfacetype.READONLY, endpoint);
bundles[0].setClusterInterface(Interfacetype.WRITE, endpoint);
AssertUtil.assertFailed(prism.getClusterHelper()
.submitEntity(bundles[0].getClusterElement().toString()));
}
@Test(dataProvider = "getData")
public void replicateToExternalFS(final FileSystem externalFS,
final String separator, final boolean withData) throws Exception {
final String endpoint = externalFS.getUri().toString();
Bundle.submitCluster(bundles[0], externalBundle);
String startTime = TimeUtil.getTimeWrtSystemTime(0);
String endTime = TimeUtil.addMinsToTime(startTime, 5);
LOGGER.info("Time range between : " + startTime + " and " + endTime);
String datePattern = StringUtils .join(
new String[]{"${YEAR}", "${MONTH}", "${DAY}", "${HOUR}", "${MINUTE}"}, separator);
//configure feed
FeedMerlin feed = new FeedMerlin(bundles[0].getDataSets().get(0));
String targetDataLocation = endpoint + testWasbTargetDir + datePattern;
feed.setFilePath(sourcePath + '/' + datePattern);
//erase all clusters from feed definition
feed.clearFeedClusters();
//set local cluster as source
feed.addFeedCluster(
new FeedMerlin.FeedClusterBuilder(Util.readEntityName(bundles[0].getClusters().get(0)))
.withRetention("days(1000000)", ActionType.DELETE)
.withValidity(startTime, endTime)
.withClusterType(ClusterType.SOURCE)
.build());
//set externalFS cluster as target
feed.addFeedCluster(
new FeedMerlin.FeedClusterBuilder(Util.readEntityName(externalBundle.getClusters().get(0)))
.withRetention("days(1000000)", ActionType.DELETE)
.withValidity(startTime, endTime)
.withClusterType(ClusterType.TARGET)
.withDataLocation(targetDataLocation)
.build());
//submit and schedule feed
LOGGER.info("Feed : " + Util.prettyPrintXml(feed.toString()));
AssertUtil.assertSucceeded(prism.getFeedHelper().submitAndSchedule(feed.toString()));
datePattern = StringUtils.join(new String[]{"yyyy", "MM", "dd", "HH", "mm"}, separator);
//upload necessary data
DateTime date = new DateTime(startTime, DateTimeZone.UTC);
DateTimeFormatter fmt = DateTimeFormat.forPattern(datePattern);
String timePattern = fmt.print(date);
HadoopUtil.recreateDir(clusterFS, sourcePath + '/' + timePattern);
if (withData) {
HadoopUtil.copyDataToFolder(clusterFS, sourcePath + '/' + timePattern, OSUtil.SINGLE_FILE);
}
Path srcPath = new Path(sourcePath + '/' + timePattern);
Path dstPath = new Path(endpoint + testWasbTargetDir + '/' + timePattern);
//check if coordinator exists
TimeUtil.sleepSeconds(10);
InstanceUtil.waitTillInstancesAreCreated(clusterOC, feed.toString(), 0);
Assert.assertEquals(OozieUtil.checkIfFeedCoordExist(clusterOC, feed.getName(), "REPLICATION"), 1);
//replication should start, wait while it ends
InstanceUtil.waitTillInstanceReachState(clusterOC, Util.readEntityName(feed.toString()), 1,
CoordinatorAction.Status.SUCCEEDED, EntityType.FEED);
//check if data has been replicated correctly
List<Path> cluster1ReplicatedData =
HadoopUtil.getAllFilesRecursivelyHDFS(clusterFS, srcPath);
List<Path> cluster2ReplicatedData =
HadoopUtil.getAllFilesRecursivelyHDFS(externalFS, dstPath);
AssertUtil.checkForListSizes(cluster1ReplicatedData, cluster2ReplicatedData);
final ContentSummary srcSummary = clusterFS.getContentSummary(srcPath);
final ContentSummary dstSummary = externalFS.getContentSummary(dstPath);
Assert.assertEquals(dstSummary.getLength(), srcSummary.getLength());
}
@DataProvider
public Object[][] getData() {
//"-" for single directory, "/" - for dir with subdirs };
return MatrixUtil.crossProduct(new FileSystem[]{wasbFS},
new String[]{"/", "-"},
new Boolean[]{true, false});
}
@DataProvider
public Object[][] getInvalidTargets() {
return new Object[][]{{"wasb://invalid@invalid.blob.core.windows.net/"}};
}
}