/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, * software distributed under the License is distributed on an * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. */ package org.apache.ranger.services.hive.client; import java.io.Closeable; import java.io.File; import java.net.MalformedURLException; import java.security.PrivilegedAction; import java.security.PrivilegedExceptionAction; import java.sql.Connection; import java.sql.Driver; import java.sql.DriverManager; import java.sql.ResultSet; import java.sql.SQLException; import java.sql.SQLTimeoutException; import java.sql.Statement; import java.util.ArrayList; import java.util.HashMap; import java.util.List; import java.util.Map; import java.util.Properties; import javax.security.auth.Subject; import org.apache.commons.collections.CollectionUtils; import org.apache.commons.io.FilenameUtils; import org.apache.commons.lang.StringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.HiveMetaStoreClient; import org.apache.hadoop.hive.metastore.api.FieldSchema; import org.apache.hadoop.hive.metastore.api.MetaException; import org.apache.ranger.plugin.client.BaseClient; import org.apache.ranger.plugin.client.HadoopException; import org.apache.thrift.TException; public class HiveClient extends BaseClient implements Closeable { private static final Log LOG = LogFactory.getLog(HiveClient.class); private static final String ERR_MSG = "You can still save the repository and start creating " + "policies, but you would not be able to use autocomplete for " + "resource names. Check ranger_admin.log for more info."; private Connection con; private HiveMetaStoreClient hiveClient; private String hiveSiteFilePath; private boolean isKerberosAuth; private boolean enableHiveMetastoreLookup; public HiveClient(String serviceName) throws Exception { super(serviceName, null); initHive(); } public HiveClient(String serviceName,Map<String,String> connectionProp) throws Exception{ super(serviceName,connectionProp); initHive(); } public void initHive() throws Exception { enableHiveMetastoreLookup = getConfigHolder().isEnableHiveMetastoreLookup(); hiveSiteFilePath = getConfigHolder().getHiveSiteFilePath(); isKerberosAuth = getConfigHolder().isKerberosAuthentication(); if (isKerberosAuth) { LOG.info("Secured Mode: JDBC Connection done with preAuthenticated Subject"); Subject.doAs(getLoginSubject(), new PrivilegedExceptionAction<Void>(){ public Void run() throws Exception { initConnection(); return null; }}); } else { LOG.info("Since Password is NOT provided, Trying to use UnSecure client with username and password"); final String userName = getConfigHolder().getUserName(); final String password = getConfigHolder().getPassword(); Subject.doAs(getLoginSubject(), new PrivilegedExceptionAction<Void>() { public Void run() throws Exception { initConnection(userName,password); return null; }}); } } public List<String> getDatabaseList(String databaseMatching, final List<String> databaseList) throws HadoopException{ final String dbMatching = databaseMatching; final List<String> dbList = databaseList; List<String> dblist = Subject.doAs(getLoginSubject(), new PrivilegedAction<List<String>>() { public List<String> run() { List<String> ret = null; try { if (enableHiveMetastoreLookup) { ret = getDBListFromHM(dbMatching,dbList); } else { ret = getDBList(dbMatching,dbList); } } catch (HadoopException he) { LOG.error("<== HiveClient getDatabaseList() :Unable to get the Database List", he); throw he; } return ret; } }); return dblist; } private List<String> getDBListFromHM(String databaseMatching, List<String>dbList) throws HadoopException { if (LOG.isDebugEnabled()) { LOG.debug("==> HiveClient getDBListFromHM databaseMatching : " + databaseMatching + " ExcludedbList : " + dbList); } List<String> ret = new ArrayList<String>(); try { List<String> hiveDBList = null; if (hiveClient != null) { if (databaseMatching.equalsIgnoreCase("*")) { hiveDBList = hiveClient.getAllDatabases(); } else { hiveDBList = hiveClient.getDatabases(databaseMatching); } } if (hiveDBList != null) { for (String dbName : hiveDBList) { if (dbList != null && dbList.contains(dbName)) { continue; } ret.add(dbName); } } } catch (MetaException e) { String msgDesc = "Unable to get Database"; HadoopException hdpException = new HadoopException(msgDesc, e); hdpException.generateResponseDataMap(false, getMessage(e), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug("<== HiveClient.getDBListFromHM() Error : " , e); } throw hdpException; } if (LOG.isDebugEnabled()) { LOG.debug("<== HiveClient.getDBListFromHM(): " + ret); } return ret; } private List<String> getDBList(String databaseMatching, List<String>dbList) throws HadoopException { if (LOG.isDebugEnabled()) { LOG.debug("==> HiveClient getDBList databaseMatching : " + databaseMatching + " ExcludedbList :" + dbList); } List<String> ret = new ArrayList<String>(); if (con != null) { Statement stat = null; ResultSet rs = null; String sql = "show databases"; if (databaseMatching != null && !databaseMatching.isEmpty()) { sql = sql + " like \"" + databaseMatching + "\""; } try { stat = con.createStatement() ; rs = stat.executeQuery(sql); while (rs.next()) { String dbName = rs.getString(1); if (dbList != null && dbList.contains(dbName)) { continue; } ret.add(rs.getString(1)); } } catch (SQLTimeoutException sqlt) { String msgDesc = "Time Out, Unable to execute SQL [" + sql + "]."; HadoopException hdpException = new HadoopException(msgDesc, sqlt); hdpException.generateResponseDataMap(false, getMessage(sqlt), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug("<== HiveClient.getDBList() Error : ", sqlt); } throw hdpException; } catch (SQLException sqle) { String msgDesc = "Unable to execute SQL [" + sql + "]."; HadoopException hdpException = new HadoopException(msgDesc, sqle); hdpException.generateResponseDataMap(false, getMessage(sqle), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug("<== HiveClient.getDBList() Error : " , sqle); } throw hdpException; } finally { close(rs); close(stat); } } if (LOG.isDebugEnabled()) { LOG.debug("<== HiveClient.getDBList(): " + ret); } return ret; } public List<String> getTableList(String tableNameMatching, List<String> databaseList, List<String> tblNameList) throws HadoopException { final String tblNameMatching = tableNameMatching; final List<String> dbList = databaseList; final List<String> tblList = tblNameList; List<String> tableList = Subject.doAs(getLoginSubject(), new PrivilegedAction<List<String>>() { public List<String> run() { List<String> ret = null; try { if (enableHiveMetastoreLookup) { ret = getTblListFromHM(tblNameMatching,dbList,tblList); } else { ret = getTblList(tblNameMatching,dbList,tblList); } } catch (HadoopException he) { LOG.error("<== HiveClient getTblList() :Unable to get the Table List", he); throw he; } return ret; } }); return tableList; } private List<String> getTblListFromHM(String tableNameMatching, List<String> dbList, List<String> tblList) throws HadoopException { if (LOG.isDebugEnabled()) { LOG.debug("==> HiveClient getTblListFromHM() tableNameMatching : " + tableNameMatching + " ExcludedbList :" + dbList + "ExcludeTableList :" + tblList); } List<String> ret = new ArrayList<String>(); if (hiveClient != null && dbList != null && !dbList.isEmpty()) { for (String dbName : dbList) { try { List<String> hiveTblList = hiveClient.getTables(dbName, tableNameMatching); for (String tblName : hiveTblList) { if (tblList != null && tblList.contains(tblName)) { continue; } ret.add(tblName); } } catch (MetaException e) { String msgDesc = "Unable to get Table."; HadoopException hdpException = new HadoopException(msgDesc,e); hdpException.generateResponseDataMap(false, getMessage(e), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug("<== HiveClient.getTblListFromHM() Error : " , e); } throw hdpException; } } } if (LOG.isDebugEnabled()) { LOG.debug("<== HiveClient getTblListFromHM() " + ret); } return ret; } private List<String> getTblList(String tableNameMatching, List<String> dbList, List<String> tblList) throws HadoopException { if (LOG.isDebugEnabled()) { LOG.debug("==> HiveClient getTblList() tableNameMatching : " + tableNameMatching + " ExcludedbList :" + dbList + "ExcludeTableList :" + tblList); } List<String> ret = new ArrayList<String>(); if (con != null) { Statement stat = null; ResultSet rs = null; String sql = null; try { if (dbList != null && !dbList.isEmpty()) { for (String db : dbList) { sql = "use " + db; try { stat = con.createStatement(); stat.execute(sql); } finally { close(stat); stat = null; } sql = "show tables "; if (tableNameMatching != null && !tableNameMatching.isEmpty()) { sql = sql + " like \"" + tableNameMatching + "\""; } try { stat = con.createStatement(); rs = stat.executeQuery(sql); while (rs.next()) { String tblName = rs.getString(1); if (tblList != null && tblList.contains(tblName)) { continue; } ret.add(tblName); } } finally { close(rs); close(stat); rs = null; stat = null; } } } } catch (SQLTimeoutException sqlt) { String msgDesc = "Time Out, Unable to execute SQL [" + sql + "]."; HadoopException hdpException = new HadoopException(msgDesc, sqlt); hdpException.generateResponseDataMap(false, getMessage(sqlt), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug("<== HiveClient.getTblList() Error : " , sqlt); } throw hdpException; } catch (SQLException sqle) { String msgDesc = "Unable to execute SQL [" + sql + "]."; HadoopException hdpException = new HadoopException(msgDesc, sqle); hdpException.generateResponseDataMap(false, getMessage(sqle), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug("<== HiveClient.getTblList() Error : " , sqle); } throw hdpException; } } if (LOG.isDebugEnabled()) { LOG.debug("<== HiveClient getTblList() " + ret); } return ret; } public List<String> getViewList(String database, String viewNameMatching) { List<String> ret = null; return ret; } public List<String> getUDFList(String database, String udfMatching) { List<String> ret = null; return ret; } public List<String> getColumnList(String columnNameMatching, List<String> dbList, List<String> tblList, List<String> colList) throws HadoopException { final String clmNameMatching = columnNameMatching; final List<String> databaseList = dbList; final List<String> tableList = tblList; final List<String> clmList = colList; List<String> columnList = Subject.doAs(getLoginSubject(), new PrivilegedAction<List<String>>() { public List<String> run() { List<String> ret = null; try { if (enableHiveMetastoreLookup) { ret = getClmListFromHM(clmNameMatching,databaseList,tableList,clmList); } else { ret = getClmList(clmNameMatching,databaseList,tableList,clmList); } } catch (HadoopException he) { LOG.error("<== HiveClient getColumnList() :Unable to get the Column List", he); throw he; } return ret; } }); return columnList; } private List<String> getClmListFromHM(String columnNameMatching,List<String> dbList, List<String> tblList, List<String> colList) throws HadoopException { if (LOG.isDebugEnabled()) { LOG.debug("==> HiveClient.getClmListFromHM() columnNameMatching: " + columnNameMatching + " dbList :" + dbList + " tblList: " + tblList + " colList: " + colList); } List<String> ret = new ArrayList<String>(); String columnNameMatchingRegEx = null; if (columnNameMatching != null && !columnNameMatching.isEmpty()) { columnNameMatchingRegEx = columnNameMatching; } if (hiveClient != null && dbList != null && !dbList.isEmpty() && tblList != null && !tblList.isEmpty()) { for (String db : dbList) { for (String tbl : tblList) { try { List<FieldSchema> hiveSch = hiveClient.getFields(db, tbl); if (hiveSch != null) { for (FieldSchema sch : hiveSch) { String columnName = sch.getName(); if (colList != null && colList.contains(columnName)) { continue; } if (columnNameMatchingRegEx == null) { ret.add(columnName); } else if (FilenameUtils.wildcardMatch(columnName, columnNameMatchingRegEx)) { ret.add(columnName); } } } } catch (TException e) { String msgDesc = "Unable to get Columns."; HadoopException hdpException = new HadoopException(msgDesc, e); hdpException.generateResponseDataMap(false, getMessage(e), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug("<== HiveClient.getClmListFromHM() Error : " ,e); } throw hdpException; } } } } if (LOG.isDebugEnabled()) { LOG.debug("<== HiveClient.getClmListFromHM() " + ret ); } return ret; } private List<String> getClmList(String columnNameMatching,List<String> dbList, List<String> tblList, List<String> colList) throws HadoopException { if (LOG.isDebugEnabled()) { LOG.debug("==> HiveClient.getClmList() columnNameMatching: " + columnNameMatching + " dbList :" + dbList + " tblList: " + tblList + " colList: " + colList); } List<String> ret = new ArrayList<String>(); if (con != null) { String columnNameMatchingRegEx = null; if (columnNameMatching != null && !columnNameMatching.isEmpty()) { columnNameMatchingRegEx = columnNameMatching; } Statement stat = null; ResultSet rs = null; String sql = null; if (dbList != null && !dbList.isEmpty() && tblList != null && !tblList.isEmpty()) { for (String db : dbList) { for (String tbl : tblList) { try { sql = "use " + db; try { stat = con.createStatement(); stat.execute(sql); } finally { close(stat); } sql = "describe " + tbl; stat = con.createStatement() ; rs = stat.executeQuery(sql); while (rs.next()) { String columnName = rs.getString(1); if (colList != null && colList.contains(columnName)) { continue; } if (columnNameMatchingRegEx == null) { ret.add(columnName); } else if (FilenameUtils.wildcardMatch(columnName,columnNameMatchingRegEx)) { ret.add(columnName); } } } catch (SQLTimeoutException sqlt) { String msgDesc = "Time Out, Unable to execute SQL [" + sql + "]."; HadoopException hdpException = new HadoopException(msgDesc, sqlt); hdpException.generateResponseDataMap(false, getMessage(sqlt), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug("<== HiveClient.getClmList() Error : " ,sqlt); } throw hdpException; } catch (SQLException sqle) { String msgDesc = "Unable to execute SQL [" + sql + "]."; HadoopException hdpException = new HadoopException(msgDesc, sqle); hdpException.generateResponseDataMap(false, getMessage(sqle), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug("<== HiveClient.getClmList() Error : " ,sqle); } throw hdpException; } finally { close(rs); close(stat); } } } } } if (LOG.isDebugEnabled()) { LOG.debug("<== HiveClient.getClmList() " + ret ); } return ret; } public void close() { Subject.doAs(getLoginSubject(), new PrivilegedAction<Void>(){ public Void run() { close(con); return null; } }); } private void close(Statement aStat) { try { if (aStat != null) { aStat.close(); } } catch (SQLException e) { LOG.error("Unable to close SQL statement", e); } } private void close(ResultSet aResultSet) { try { if (aResultSet != null) { aResultSet.close(); } } catch (SQLException e) { LOG.error("Unable to close ResultSet", e); } } private void close(Connection aCon) { try { if (aCon != null) { aCon.close(); } } catch (SQLException e) { LOG.error("Unable to close SQL Connection", e); } } private void initConnection() throws HadoopException{ try { initConnection(null,null); } catch (HadoopException he) { LOG.error("Unable to Connect to Hive", he); throw he; } } private void initConnection(String userName, String password) throws HadoopException { if (enableHiveMetastoreLookup) { try { HiveConf conf = new HiveConf(); if (!StringUtils.isEmpty(hiveSiteFilePath)) { File f = new File(hiveSiteFilePath); if (f.exists()) { conf.addResource(f.toURI().toURL()); } else { if(LOG.isDebugEnabled()) { LOG.debug("Hive site conf file path " + hiveSiteFilePath + " does not exists for Hive Metastore lookup"); } } } else { if (LOG.isDebugEnabled()) { LOG.debug("Hive site conf file path property not found for Hive Metastore lookup"); } } hiveClient = new HiveMetaStoreClient(conf); } catch (HadoopException he) { String msgDesc = "initConnection: Class or its nullary constructor might not accessible." + "So unable to initiate connection to hive thrift server instance."; HadoopException hdpException = new HadoopException(msgDesc, he); hdpException.generateResponseDataMap(false, getMessage(he), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug(msgDesc, hdpException); } throw hdpException; } catch (MalformedURLException e) { String msgDesc = "initConnection: URL might be malformed." + "So unable to initiate connection to hive thrift server instance."; HadoopException hdpException = new HadoopException(msgDesc, e); hdpException.generateResponseDataMap(false, getMessage(e), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug(msgDesc, hdpException); } throw hdpException; } catch (MetaException e) { String msgDesc = "initConnection: Meta info is not proper." + "So unable to initiate connection to hive thrift server instance."; HadoopException hdpException = new HadoopException(msgDesc, e); hdpException.generateResponseDataMap(false, getMessage(e), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug(msgDesc, hdpException); } throw hdpException; } catch ( Throwable t) { String msgDesc = "Unable to connect to Hive Thrift Server instance"; HadoopException hdpException = new HadoopException(msgDesc, t); hdpException.generateResponseDataMap(false, getMessage(t), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug(msgDesc, hdpException); } throw hdpException; } } else { Properties prop = getConfigHolder().getRangerSection(); String driverClassName = prop.getProperty("jdbc.driverClassName"); String url = prop.getProperty("jdbc.url"); if (driverClassName != null) { try { Driver driver = (Driver)Class.forName(driverClassName).newInstance(); DriverManager.registerDriver(driver); } catch (SQLException e) { String msgDesc = "initConnection: Caught SQLException while registering " + "Hive driver, so Unable to connect to Hive Thrift Server instance."; HadoopException hdpException = new HadoopException(msgDesc, e); hdpException.generateResponseDataMap(false, getMessage(e), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug(msgDesc, hdpException); } throw hdpException; } catch (IllegalAccessException ilae) { String msgDesc = "initConnection: Class or its nullary constructor might not accessible." + "So unable to initiate connection to hive thrift server instance."; HadoopException hdpException = new HadoopException(msgDesc, ilae); hdpException.generateResponseDataMap(false, getMessage(ilae), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug(msgDesc, hdpException); } throw hdpException; } catch (InstantiationException ie) { String msgDesc = "initConnection: Class may not have its nullary constructor or " + "may be the instantiation fails for some other reason." + "So unable to initiate connection to hive thrift server instance."; HadoopException hdpException = new HadoopException(msgDesc, ie); hdpException.generateResponseDataMap(false, getMessage(ie), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug(msgDesc, hdpException); } throw hdpException; } catch (ExceptionInInitializerError eie) { String msgDesc = "initConnection: Got ExceptionInInitializerError, " + "The initialization provoked by this method fails." + "So unable to initiate connection to hive thrift server instance."; HadoopException hdpException = new HadoopException(msgDesc, eie); hdpException.generateResponseDataMap(false, getMessage(eie), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug(msgDesc, hdpException); } throw hdpException; } catch (SecurityException se) { String msgDesc = "initConnection: unable to initiate connection to hive thrift server instance," + " The caller's class loader is not the same as or an ancestor " + "of the class loader for the current class and invocation of " + "s.checkPackageAccess() denies access to the package of this class."; HadoopException hdpException = new HadoopException(msgDesc, se); hdpException.generateResponseDataMap(false, getMessage(se), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug(msgDesc, hdpException); } throw hdpException; } catch (Throwable t) { String msgDesc = "initConnection: Unable to connect to Hive Thrift Server instance, " + "please provide valid value of field : {jdbc.driverClassName}."; HadoopException hdpException = new HadoopException(msgDesc, t); hdpException.generateResponseDataMap(false, getMessage(t), msgDesc + ERR_MSG, null, "jdbc.driverClassName"); if (LOG.isDebugEnabled()) { LOG.debug(msgDesc, hdpException); } throw hdpException; } } try { if (userName == null && password == null) { con = DriverManager.getConnection(url); } else { con = DriverManager.getConnection(url, userName, password); } } catch (SQLException e) { String msgDesc = "Unable to connect to Hive Thrift Server instance."; HadoopException hdpException = new HadoopException(msgDesc, e); hdpException.generateResponseDataMap(false, getMessage(e), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug(msgDesc, hdpException); } throw hdpException; } catch (SecurityException se) { String msgDesc = "Unable to connect to Hive Thrift Server instance."; HadoopException hdpException = new HadoopException(msgDesc, se); hdpException.generateResponseDataMap(false, getMessage(se), msgDesc + ERR_MSG, null, null); if (LOG.isDebugEnabled()) { LOG.debug(msgDesc, hdpException); } throw hdpException; } catch ( Throwable t) { String msgDesc = "Unable to connect to Hive Thrift Server instance"; HadoopException hdpException = new HadoopException(msgDesc, t); hdpException.generateResponseDataMap(false, getMessage(t), msgDesc + ERR_MSG, null, url); if (LOG.isDebugEnabled()) { LOG.debug(msgDesc, hdpException); } throw hdpException; } } } public static void main(String[] args) { HiveClient hc = null; if (args.length == 0) { System.err.println("USAGE: java " + HiveClient.class.getName() + " dataSourceName <databaseName> <tableName> <columnName>"); System.exit(1); } try { hc = new HiveClient(args[0]); if (args.length == 2) { List<String> dbList = null; try { dbList = hc.getDatabaseList(args[1],null); } catch (Exception e) { e.printStackTrace(); } if (CollectionUtils.isEmpty(dbList)) { System.out.println("No database found with db filter [" + args[1] + "]"); } else { if (CollectionUtils.isNotEmpty(dbList)) { for (String str : dbList) { System.out.println("database: " + str ); } } } } else if (args.length == 3) { List<String> tableList = hc.getTableList(args[2],null,null); if (tableList.size() == 0) { System.out.println("No tables found under database[" + args[1] + "] with table filter [" + args[2] + "]"); } else { for (String str : tableList) { System.out.println("Table: " + str); } } } else if (args.length == 4) { List<String> columnList = hc.getColumnList(args[3],null,null,null); if (columnList.size() == 0) { System.out.println("No columns found for db:" + args[1] + ", table: [" + args[2] + "], with column filter [" + args[3] + "]"); } else { for (String str : columnList) { System.out.println("Column: " + str); } } } } catch(Exception e) { e.printStackTrace(); } finally { if (hc != null) { hc.close(); } } } public static Map<String, Object> connectionTest(String serviceName, Map<String, String> connectionProperties) throws Exception { HiveClient connectionObj = null; Map<String, Object> responseData = new HashMap<String, Object>(); boolean connectivityStatus = false; List<String> testResult = null; try { connectionObj = new HiveClient(serviceName, connectionProperties); if (connectionObj != null) { testResult = connectionObj.getDatabaseList("*",null); if (testResult != null && testResult.size() != 0) { connectivityStatus = true; } if (connectivityStatus) { String successMsg = "ConnectionTest Successful"; generateResponseDataMap(connectivityStatus, successMsg, successMsg, null, null, responseData); } else { String failureMsg = "Unable to retrieve any databases using given parameters."; generateResponseDataMap(connectivityStatus, failureMsg, failureMsg + ERR_MSG, null, null, responseData); } } } catch (Exception e) { throw e; } finally { if (connectionObj != null) { connectionObj.close(); } } return responseData; } }