/** * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.camel.component.hdfs2.osgi; import org.apache.hadoop.util.ShutdownHookManager; import org.osgi.framework.BundleActivator; import org.osgi.framework.BundleContext; public class HdfsActivator implements BundleActivator { @Override public void start(BundleContext context) throws Exception { } @Override public void stop(BundleContext context) throws Exception { // There's problem inside OSGi when framwork is being shutdown // hadoop.fs code registers some JVM shutdown hooks throughout the code and this ordered // list of hooks is run in shutdown thread. // At that time bundle class loader / bundle wiring is no longer valid (bundle is stopped) // so ShutdownHookManager can't load additional classes. But there are some inner classes // loaded when iterating over registered hadoop shutdown hooks. // Let's explicitely load these inner classes when bundle is stopped, as there's last chance // to use valid bundle class loader. // This is based on the knowledge of what's contained in SMX bundle // org.apache.servicemix.bundles.hadoop-client-*.jar // the above is just a warning that hadopp may have some quirks when running inside OSGi ClassLoader hadoopCl = ShutdownHookManager.class.getClassLoader(); if (hadoopCl != null) { String shm = ShutdownHookManager.class.getName(); hadoopCl.loadClass(shm + "$1"); hadoopCl.loadClass(shm + "$2"); hadoopCl.loadClass(shm + "$HookEntry"); } } }