/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.ignite.internal.processors.hadoop.impl.igfs; import java.io.OutputStream; import java.net.URI; import java.net.URISyntaxException; import java.util.Collection; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.ignite.cache.CacheWriteSynchronizationMode; import org.apache.ignite.configuration.CacheConfiguration; import org.apache.ignite.configuration.FileSystemConfiguration; import org.apache.ignite.configuration.IgniteConfiguration; import org.apache.ignite.configuration.NearCacheConfiguration; import org.apache.ignite.igfs.IgfsGroupDataBlocksKeyMapper; import org.apache.ignite.igfs.IgfsIpcEndpointConfiguration; import org.apache.ignite.igfs.IgfsIpcEndpointType; import org.apache.ignite.internal.util.ipc.shmem.IpcSharedMemoryServerEndpoint; import org.apache.ignite.internal.util.typedef.F; import org.apache.ignite.internal.util.typedef.G; import org.apache.ignite.internal.util.typedef.internal.U; import org.apache.ignite.lang.IgniteBiTuple; import org.apache.ignite.spi.discovery.tcp.TcpDiscoverySpi; import org.apache.ignite.spi.discovery.tcp.ipfinder.TcpDiscoveryIpFinder; import org.apache.ignite.spi.discovery.tcp.ipfinder.vm.TcpDiscoveryVmIpFinder; import org.apache.ignite.testframework.junits.common.GridCommonAbstractTest; import static org.apache.ignite.cache.CacheAtomicityMode.TRANSACTIONAL; import static org.apache.ignite.cache.CacheMode.PARTITIONED; import static org.apache.ignite.events.EventType.EVT_JOB_MAPPED; import static org.apache.ignite.events.EventType.EVT_TASK_FAILED; import static org.apache.ignite.events.EventType.EVT_TASK_FINISHED; /** * Test hadoop file system implementation. */ public class IgfsNearOnlyMultiNodeSelfTest extends GridCommonAbstractTest { /** Path to the default hadoop configuration. */ public static final String HADOOP_FS_CFG = "examples/config/filesystem/core-site.xml"; /** Group size. */ public static final int GRP_SIZE = 128; /** IP finder. */ private static final TcpDiscoveryIpFinder IP_FINDER = new TcpDiscoveryVmIpFinder(true); /** Node count. */ private int cnt; /** {@inheritDoc} */ @Override protected void beforeTestsStarted() throws Exception { startGrids(nodeCount()); grid(0).createNearCache("data", new NearCacheConfiguration()); grid(0).createNearCache("meta", new NearCacheConfiguration()); } /** {@inheritDoc} */ @Override protected void afterTestsStopped() throws Exception { G.stopAll(true); } /** {@inheritDoc} */ @Override protected IgniteConfiguration getConfiguration(String igniteInstanceName) throws Exception { IgniteConfiguration cfg = super.getConfiguration(igniteInstanceName); cfg.setDiscoverySpi(new TcpDiscoverySpi().setIpFinder(IP_FINDER).setForceServerMode(true)); FileSystemConfiguration igfsCfg = new FileSystemConfiguration(); igfsCfg.setName("igfs"); igfsCfg.setDataCacheConfiguration(cacheConfiguration(igniteInstanceName, "data")); igfsCfg.setMetaCacheConfiguration(cacheConfiguration(igniteInstanceName, "meta")); IgfsIpcEndpointConfiguration endpointCfg = new IgfsIpcEndpointConfiguration(); endpointCfg.setType(IgfsIpcEndpointType.SHMEM); endpointCfg.setPort(IpcSharedMemoryServerEndpoint.DFLT_IPC_PORT + cnt); igfsCfg.setIpcEndpointConfiguration(endpointCfg); igfsCfg.setBlockSize(512 * 1024); // Together with group blocks mapper will yield 64M per node groups. cfg.setFileSystemConfiguration(igfsCfg); cfg.setIncludeEventTypes(EVT_TASK_FAILED, EVT_TASK_FINISHED, EVT_JOB_MAPPED); if (cnt == 0) cfg.setClientMode(true); cnt++; return cfg; } /** @return Node count for test. */ protected int nodeCount() { return 4; } /** * Gets cache configuration. * * @param igniteInstanceName Ignite instance name. * @return Cache configuration. */ protected CacheConfiguration cacheConfiguration(String igniteInstanceName, String cacheName) { CacheConfiguration cacheCfg = defaultCacheConfiguration(); cacheCfg.setName(cacheName); cacheCfg.setCacheMode(PARTITIONED); cacheCfg.setWriteSynchronizationMode(CacheWriteSynchronizationMode.FULL_SYNC); cacheCfg.setAffinityMapper(new IgfsGroupDataBlocksKeyMapper(GRP_SIZE)); cacheCfg.setBackups(0); cacheCfg.setAtomicityMode(TRANSACTIONAL); return cacheCfg; } /** * Gets config of concrete File System. * * @return Config of concrete File System. */ protected Configuration getFileSystemConfig() { Configuration cfg = new Configuration(); cfg.addResource(U.resolveIgniteUrl(HADOOP_FS_CFG)); return cfg; } /** * Gets File System name. * * @param grid Grid index. * @return File System name. */ protected URI getFileSystemURI(int grid) { try { return new URI("igfs://127.0.0.1:" + (IpcSharedMemoryServerEndpoint.DFLT_IPC_PORT + grid)); } catch (URISyntaxException e) { throw new RuntimeException(e); } } /** @throws Exception If failed. */ public void testContentsConsistency() throws Exception { try (FileSystem fs = FileSystem.get(getFileSystemURI(0), getFileSystemConfig())) { Collection<IgniteBiTuple<String, Long>> files = F.asList( F.t("/dir1/dir2/file1", 1024L), F.t("/dir1/dir2/file2", 8 * 1024L), F.t("/dir1/file1", 1024 * 1024L), F.t("/dir1/file2", 5 * 1024 * 1024L), F.t("/file1", 64 * 1024L + 13), F.t("/file2", 13L), F.t("/file3", 123764L) ); for (IgniteBiTuple<String, Long> file : files) { info("Writing file: " + file.get1()); try (OutputStream os = fs.create(new Path(file.get1()), (short)3)) { byte[] data = new byte[file.get2().intValue()]; data[0] = 25; data[data.length - 1] = 26; os.write(data); } info("Finished writing file: " + file.get1()); } for (int i = 1; i < nodeCount(); i++) { try (FileSystem ignored = FileSystem.get(getFileSystemURI(i), getFileSystemConfig())) { for (IgniteBiTuple<String, Long> file : files) { Path path = new Path(file.get1()); FileStatus fileStatus = fs.getFileStatus(path); assertEquals(file.get2(), (Long)fileStatus.getLen()); byte[] read = new byte[file.get2().intValue()]; info("Reading file: " + path); try (FSDataInputStream in = fs.open(path)) { in.readFully(read); assert read[0] == 25; assert read[read.length - 1] == 26; } info("Finished reading file: " + path); } } } } } }