// Licensed to the Apache Software Foundation (ASF) under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use this file except in compliance // with the License. You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, // software distributed under the License is distributed on an // "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY // KIND, either express or implied. See the License for the // specific language governing permissions and limitations // under the License. package com.cloud.upgrade.dao; import java.io.File; import java.io.UnsupportedEncodingException; import java.sql.Connection; import java.sql.Date; import java.sql.PreparedStatement; import java.sql.ResultSet; import java.sql.SQLException; import java.sql.Types; import java.util.ArrayList; import java.util.HashMap; import java.util.Iterator; import java.util.List; import java.util.Map; import java.util.UUID; import org.apache.log4j.Logger; import org.apache.cloudstack.api.ApiConstants; import org.apache.cloudstack.engine.subsystem.api.storage.DataStoreProvider; import com.cloud.deploy.DeploymentPlanner; import com.cloud.hypervisor.Hypervisor.HypervisorType; import com.cloud.network.vpc.NetworkACL; import com.cloud.utils.Pair; import com.cloud.utils.crypt.DBEncryptionUtil; import com.cloud.utils.exception.CloudRuntimeException; import com.cloud.utils.script.Script; public class Upgrade410to420 implements DbUpgrade { final static Logger s_logger = Logger.getLogger(Upgrade410to420.class); @Override public String[] getUpgradableVersionRange() { return new String[] {"4.1.0", "4.2.0"}; } @Override public String getUpgradedVersion() { return "4.2.0"; } @Override public boolean supportsRollingUpgrade() { return false; } @Override public File[] getPrepareScripts() { String script = Script.findScript("", "db/schema-410to420.sql"); if (script == null) { throw new CloudRuntimeException("Unable to find db/schema-410to420.sql"); } return new File[] {new File(script)}; } @Override public void performDataMigration(Connection conn) { movePrivateZoneToDedicatedResource(conn); upgradeVmwareLabels(conn); persistLegacyZones(conn); persistVswitchConfiguration(conn); createPlaceHolderNics(conn); updateRemoteAccessVpn(conn); updateOverCommitRatioClusterDetails(conn); updatePrimaryStore(conn); addEgressFwRulesForSRXGuestNw(conn); upgradeEIPNetworkOfferings(conn); updateGlobalDeploymentPlanner(conn); upgradeDefaultVpcOffering(conn); upgradePhysicalNtwksWithInternalLbProvider(conn); updateNetworkACLs(conn); addHostDetailsIndex(conn); updateNetworksForPrivateGateways(conn); correctExternalNetworkDevicesSetup(conn); removeFirewallServiceFromSharedNetworkOfferingWithSGService(conn); fix22xKVMSnapshots(conn); setKVMSnapshotFlag(conn); addIndexForAlert(conn); fixBaremetalForeignKeys(conn); // storage refactor related migration // TODO: add clean-up scripts to delete the deprecated table. migrateSecondaryStorageToImageStore(conn); migrateVolumeHostRef(conn); migrateTemplateHostRef(conn); migrateSnapshotStoreRef(conn); migrateS3ToImageStore(conn); migrateSwiftToImageStore(conn); fixNiciraKeys(conn); fixRouterKeys(conn); encryptSite2SitePSK(conn); migrateDatafromIsoIdInVolumesTable(conn); setRAWformatForRBDVolumes(conn); migrateVolumeOnSecondaryStorage(conn); createFullCloneFlag(conn); upgradeVpcServiceMap(conn); upgradeResourceCount(conn); } private void createFullCloneFlag(Connection conn) { String update_sql; int numRows = 0; try (PreparedStatement delete = conn.prepareStatement("delete from `cloud`.`configuration` where name='vmware.create.full.clone';");) { delete.executeUpdate(); try(PreparedStatement query = conn.prepareStatement("select count(*) from `cloud`.`data_center`");) { try(ResultSet rs = query.executeQuery();) { if (rs.next()) { numRows = rs.getInt(1); } if (numRows > 0) { update_sql = "insert into `cloud`.`configuration` (`category`, `instance`, `component`, `name`, `value`, `description`) VALUES ('Advanced', 'DEFAULT', 'UserVmManager', 'vmware.create.full.clone' , 'false', 'If set to true, creates VMs as full clones on ESX hypervisor');"; } else { update_sql = "insert into `cloud`.`configuration` (`category`, `instance`, `component`, `name`, `value`, `description`) VALUES ('Advanced', 'DEFAULT', 'UserVmManager', 'vmware.create.full.clone' , 'true', 'If set to true, creates VMs as full clones on ESX hypervisor');"; } try(PreparedStatement update_pstmt = conn.prepareStatement(update_sql);) { update_pstmt.executeUpdate(); }catch (SQLException e) { throw new CloudRuntimeException("Failed to set global flag vmware.create.full.clone: ", e); } }catch (SQLException e) { throw new CloudRuntimeException("Failed to set global flag vmware.create.full.clone: ", e); } }catch (SQLException e) { throw new CloudRuntimeException("Failed to set global flag vmware.create.full.clone: ", e); } } catch (SQLException e) { throw new CloudRuntimeException("Failed to set global flag vmware.create.full.clone: ", e); } } private void migrateVolumeOnSecondaryStorage(Connection conn) { try (PreparedStatement sql = conn.prepareStatement("update `cloud`.`volumes` set state='Uploaded' where state='UploadOp'");){ sql.executeUpdate(); } catch (SQLException e) { throw new CloudRuntimeException("Failed to upgrade volume state: ", e); } } private void persistVswitchConfiguration(Connection conn) { Long clusterId; String clusterHypervisorType; final String NEXUS_GLOBAL_CONFIG_PARAM_NAME = "vmware.use.nexus.vswitch"; final String DVS_GLOBAL_CONFIG_PARAM_NAME = "vmware.use.dvswitch"; final String VSWITCH_GLOBAL_CONFIG_PARAM_CATEGORY = "Network"; final String VMWARE_STANDARD_VSWITCH = "vmwaresvs"; final String NEXUS_1000V_DVSWITCH = "nexusdvs"; String paramValStr; boolean readGlobalConfigParam = false; boolean nexusEnabled = false; String publicVswitchType = VMWARE_STANDARD_VSWITCH; String guestVswitchType = VMWARE_STANDARD_VSWITCH; Map<Long, List<Pair<String, String>>> detailsMap = new HashMap<Long, List<Pair<String, String>>>(); List<Pair<String, String>> detailsList; try (PreparedStatement clustersQuery = conn.prepareStatement("select id, hypervisor_type from `cloud`.`cluster` where removed is NULL");){ try(ResultSet clusters = clustersQuery.executeQuery();) { while (clusters.next()) { clusterHypervisorType = clusters.getString("hypervisor_type"); clusterId = clusters.getLong("id"); if (clusterHypervisorType.equalsIgnoreCase("VMware")) { if (!readGlobalConfigParam) { paramValStr = getConfigurationParameter(conn, VSWITCH_GLOBAL_CONFIG_PARAM_CATEGORY, NEXUS_GLOBAL_CONFIG_PARAM_NAME); if (paramValStr.equalsIgnoreCase("true")) { nexusEnabled = true; } } if (nexusEnabled) { publicVswitchType = NEXUS_1000V_DVSWITCH; guestVswitchType = NEXUS_1000V_DVSWITCH; } detailsList = new ArrayList<Pair<String, String>>(); detailsList.add(new Pair<String, String>(ApiConstants.VSWITCH_TYPE_GUEST_TRAFFIC, guestVswitchType)); detailsList.add(new Pair<String, String>(ApiConstants.VSWITCH_TYPE_PUBLIC_TRAFFIC, publicVswitchType)); detailsMap.put(clusterId, detailsList); updateClusterDetails(conn, detailsMap); s_logger.debug("Persist vSwitch Configuration: Successfully persisted vswitch configuration for cluster " + clusterId); } else { s_logger.debug("Persist vSwitch Configuration: Ignoring cluster " + clusterId + " with hypervisor type " + clusterHypervisorType); continue; } } // End cluster iteration }catch (SQLException e) { String msg = "Unable to persist vswitch configuration of VMware clusters." + e.getMessage(); s_logger.error(msg); throw new CloudRuntimeException(msg, e); } if (nexusEnabled) { // If Nexus global parameter is true, then set DVS configuration parameter to true. TODOS: Document that this mandates that MS need to be restarted. setConfigurationParameter(conn, VSWITCH_GLOBAL_CONFIG_PARAM_CATEGORY, DVS_GLOBAL_CONFIG_PARAM_NAME, "true"); } } catch (SQLException e) { String msg = "Unable to persist vswitch configuration of VMware clusters." + e.getMessage(); s_logger.error(msg); throw new CloudRuntimeException(msg, e); } } private void updateClusterDetails(Connection conn, Map<Long, List<Pair<String, String>>> detailsMap) { // Insert cluster details into cloud.cluster_details table for existing VMware clusters // Input parameter detailMap is a map of clusterId and list of key value pairs for that cluster Long clusterId; String key; String val; List<Pair<String, String>> keyValues; try { Iterator<Long> clusterIt = detailsMap.keySet().iterator(); while (clusterIt.hasNext()) { clusterId = clusterIt.next(); keyValues = detailsMap.get(clusterId); try( PreparedStatement clusterDetailsInsert = conn.prepareStatement("INSERT INTO `cloud`.`cluster_details` (cluster_id, name, value) VALUES (?, ?, ?)");) { for (Pair<String, String> keyValuePair : keyValues) { key = keyValuePair.first(); val = keyValuePair.second(); clusterDetailsInsert.setLong(1, clusterId); clusterDetailsInsert.setString(2, key); clusterDetailsInsert.setString(3, val); clusterDetailsInsert.executeUpdate(); } s_logger.debug("Inserted vswitch configuration details into cloud.cluster_details for cluster with id " + clusterId + "."); }catch (SQLException e) { throw new CloudRuntimeException("Unable insert cluster details into cloud.cluster_details table.", e); } } } catch (RuntimeException e) { throw new CloudRuntimeException("Unable insert cluster details into cloud.cluster_details table.", e); } } private String getConfigurationParameter(Connection conn, String category, String paramName) { try (PreparedStatement pstmt = conn.prepareStatement("select value from `cloud`.`configuration` where category=? and value is not NULL and name = ?;");) { pstmt.setString(1, category); pstmt.setString(2, paramName); try(ResultSet rs = pstmt.executeQuery();) { while (rs.next()) { return rs.getString("value"); } }catch (SQLException e) { throw new CloudRuntimeException("Unable read global configuration parameter " + paramName + ". ", e); } } catch (SQLException e) { throw new CloudRuntimeException("Unable read global configuration parameter " + paramName + ". ", e); } return "false"; } private void setConfigurationParameter(Connection conn, String category, String paramName, String paramVal) { try (PreparedStatement pstmt = conn.prepareStatement("UPDATE `cloud`.`configuration` SET value = ? WHERE name = ?;");) { pstmt.setString(1, paramVal); pstmt.setString(2, paramName); s_logger.debug("Updating global configuration parameter " + paramName + " with value " + paramVal + ". Update SQL statement is " + pstmt); pstmt.executeUpdate(); } catch (SQLException e) { throw new CloudRuntimeException("Unable to set global configuration parameter " + paramName + " to " + paramVal + ". ", e); } } private void movePrivateZoneToDedicatedResource(Connection conn) { String domainName = ""; try (PreparedStatement sel_dc_dom_id = conn.prepareStatement("SELECT distinct(`domain_id`) FROM `cloud`.`data_center` WHERE `domain_id` IS NOT NULL AND removed IS NULL");) { try (ResultSet rs3 = sel_dc_dom_id.executeQuery();) { while (rs3.next()) { long domainId = rs3.getLong(1); long affinityGroupId = 0; // create or find an affinity group for this domain of type // 'ExplicitDedication' try (PreparedStatement sel_aff_grp_pstmt = conn.prepareStatement("SELECT affinity_group.id FROM `cloud`.`affinity_group` INNER JOIN `cloud`.`affinity_group_domain_map` ON affinity_group.id=affinity_group_domain_map.affinity_group_id WHERE affinity_group.type = 'ExplicitDedication' AND affinity_group.acl_type = 'Domain' AND (affinity_group_domain_map.domain_id = ?)");) { sel_aff_grp_pstmt.setLong(1, domainId); try (ResultSet rs2 = sel_aff_grp_pstmt.executeQuery();) { if (rs2.next()) { // group exists, use it affinityGroupId = rs2.getLong(1); } else { // create new group try (PreparedStatement sel_dom_id_pstmt = conn.prepareStatement("SELECT name FROM `cloud`.`domain` where id = ?");) { sel_dom_id_pstmt.setLong(1, domainId); try (ResultSet sel_dom_id_res = sel_dom_id_pstmt.executeQuery();) { if (sel_dom_id_res.next()) { domainName = sel_dom_id_res.getString(1); } } } catch (SQLException e) { throw new CloudRuntimeException("Exception while Moving private zone information to dedicated resources", e); } // create new domain level group for this domain String type = "ExplicitDedication"; String uuid = UUID.randomUUID().toString(); String groupName = "DedicatedGrp-domain-" + domainName; s_logger.debug("Adding AffinityGroup of type " + type + " for domain id " + domainId); String sql = "INSERT INTO `cloud`.`affinity_group` (`name`, `type`, `uuid`, `description`, `domain_id`, `account_id`, `acl_type`) VALUES (?, ?, ?, ?, 1, 1, 'Domain')"; try (PreparedStatement insert_pstmt = conn.prepareStatement(sql);) { insert_pstmt.setString(1, groupName); insert_pstmt.setString(2, type); insert_pstmt.setString(3, uuid); insert_pstmt.setString(4, "dedicated resources group"); insert_pstmt.executeUpdate(); try (PreparedStatement sel_aff_pstmt = conn.prepareStatement("SELECT affinity_group.id FROM `cloud`.`affinity_group` where uuid = ?");) { sel_aff_pstmt.setString(1, uuid); try (ResultSet sel_aff_res = sel_aff_pstmt.executeQuery();) { if (sel_aff_res.next()) { affinityGroupId = sel_aff_res.getLong(1); } } catch (SQLException e) { throw new CloudRuntimeException("Exception while Moving private zone information to dedicated resources", e); } } catch (SQLException e) { throw new CloudRuntimeException("Exception while Moving private zone information to dedicated resources", e); } } catch (SQLException e) { throw new CloudRuntimeException("Exception while Moving private zone information to dedicated resources", e); } // add the domain map String sqlMap = "INSERT INTO `cloud`.`affinity_group_domain_map` (`domain_id`, `affinity_group_id`) VALUES (?, ?)"; try (PreparedStatement pstmtUpdate = conn.prepareStatement(sqlMap);) { pstmtUpdate.setLong(1, domainId); pstmtUpdate.setLong(2, affinityGroupId); pstmtUpdate.executeUpdate(); } catch (SQLException e) { throw new CloudRuntimeException("Exception while Moving private zone information to dedicated resources", e); } } } catch (SQLException e) { throw new CloudRuntimeException("Exception while Moving private zone information to dedicated resources", e); } } catch (SQLException e) { throw new CloudRuntimeException("Exception while Moving private zone information to dedicated resources", e); } try (PreparedStatement sel_pstmt = conn.prepareStatement("SELECT `id` FROM `cloud`.`data_center` WHERE `domain_id` = ? AND removed IS NULL");) { sel_pstmt.setLong(1, domainId); try (ResultSet sel_pstmt_rs = sel_pstmt.executeQuery();) { while (sel_pstmt_rs.next()) { long zoneId = sel_pstmt_rs.getLong(1); dedicateZone(conn, zoneId, domainId, affinityGroupId); } } catch (SQLException e) { throw new CloudRuntimeException("Exception while Moving private zone information to dedicated resources", e); } } catch (SQLException e) { throw new CloudRuntimeException("Exception while Moving private zone information to dedicated resources", e); } } } catch (SQLException e) { throw new CloudRuntimeException("Exception while Moving private zone information to dedicated resources", e); } }catch (SQLException e) { throw new CloudRuntimeException("Exception while Moving private zone information to dedicated resources", e); } } private void dedicateZone(Connection conn, long zoneId, long domainId, long affinityGroupId) { try( PreparedStatement pstmtUpdate2 = conn.prepareStatement("INSERT INTO `cloud`.`dedicated_resources` (`uuid`,`data_center_id`, `domain_id`, `affinity_group_id`) VALUES (?, ?, ?, ?)");) { // create the dedicated resources entry pstmtUpdate2.setString(1, UUID.randomUUID().toString()); pstmtUpdate2.setLong(2, zoneId); pstmtUpdate2.setLong(3, domainId); pstmtUpdate2.setLong(4, affinityGroupId); pstmtUpdate2.executeUpdate(); pstmtUpdate2.close(); } catch (SQLException e) { throw new CloudRuntimeException("Exception while saving zone to dedicated resources", e); } } private void fixBaremetalForeignKeys(Connection conn) { List<String> keys = new ArrayList<String>(); keys.add("fk_external_dhcp_devices_nsp_id"); keys.add("fk_external_dhcp_devices_host_id"); keys.add("fk_external_dhcp_devices_pod_id"); keys.add("fk_external_dhcp_devices_physical_network_id"); DbUpgradeUtils.dropKeysIfExist(conn, "baremetal_dhcp_devices", keys, true); keys.add("fk_external_pxe_devices_nsp_id"); keys.add("fk_external_pxe_devices_host_id"); keys.add("fk_external_pxe_devices_physical_network_id"); DbUpgradeUtils.dropKeysIfExist(conn, "baremetal_pxe_devices", keys, true); try (PreparedStatement alter_pstmt = conn.prepareStatement("ALTER TABLE `cloud`.`baremetal_dhcp_devices` ADD CONSTRAINT `fk_external_dhcp_devices_nsp_id` FOREIGN KEY (`nsp_id`) REFERENCES `physical_network_service_providers` (`id`) ON DELETE CASCADE");) { alter_pstmt.executeUpdate(); try(PreparedStatement alter_pstmt_id = conn.prepareStatement("ALTER TABLE `cloud`.`baremetal_dhcp_devices` ADD CONSTRAINT `fk_external_dhcp_devices_host_id` FOREIGN KEY (`host_id`) REFERENCES `host`(`id`) ON DELETE CASCADE"); ) { alter_pstmt_id.executeUpdate(); try(PreparedStatement alter_pstmt_phy_net = conn.prepareStatement("ALTER TABLE `cloud`.`baremetal_dhcp_devices` ADD CONSTRAINT `fk_external_dhcp_devices_physical_network_id` FOREIGN KEY (`physical_network_id`) REFERENCES `physical_network`(`id`) ON DELETE CASCADE");) { alter_pstmt_phy_net.executeUpdate(); }catch (SQLException e) { throw new CloudRuntimeException("Unable to add foreign keys to baremetal_dhcp_devices table", e); } }catch (SQLException e) { throw new CloudRuntimeException("Unable to add foreign keys to baremetal_dhcp_devices table", e); } s_logger.debug("Added foreign keys for table baremetal_dhcp_devices"); } catch (SQLException e) { throw new CloudRuntimeException("Unable to add foreign keys to baremetal_dhcp_devices table", e); } try (PreparedStatement alter_pxe_pstmt = conn.prepareStatement("ALTER TABLE `cloud`.`baremetal_pxe_devices` ADD CONSTRAINT `fk_external_pxe_devices_nsp_id` FOREIGN KEY (`nsp_id`) REFERENCES `physical_network_service_providers` (`id`) ON DELETE CASCADE");) { alter_pxe_pstmt.executeUpdate(); try(PreparedStatement alter_pxe_id_pstmt = conn.prepareStatement("ALTER TABLE `cloud`.`baremetal_pxe_devices` ADD CONSTRAINT `fk_external_pxe_devices_host_id` FOREIGN KEY (`host_id`) REFERENCES `host`(`id`) ON DELETE CASCADE");) { alter_pxe_id_pstmt.executeUpdate(); try(PreparedStatement alter_pxe_phy_net_pstmt = conn.prepareStatement("ALTER TABLE `cloud`.`baremetal_pxe_devices` ADD CONSTRAINT `fk_external_pxe_devices_physical_network_id` FOREIGN KEY (`physical_network_id`) REFERENCES `physical_network`(`id`) ON DELETE CASCADE");) { alter_pxe_phy_net_pstmt.executeUpdate(); }catch (SQLException e) { throw new CloudRuntimeException("Unable to add foreign keys to baremetal_pxe_devices table", e); } }catch (SQLException e) { throw new CloudRuntimeException("Unable to add foreign keys to baremetal_pxe_devices table", e); } s_logger.debug("Added foreign keys for table baremetal_pxe_devices"); } catch (SQLException e) { throw new CloudRuntimeException("Unable to add foreign keys to baremetal_pxe_devices table", e); } } private void addIndexForAlert(Connection conn) { //First drop if it exists. (Due to patches shipped to customers some will have the index and some wont.) List<String> indexList = new ArrayList<String>(); s_logger.debug("Dropping index i_alert__last_sent if it exists"); indexList.add("last_sent"); // in 4.1, we created this index that is not in convention. indexList.add("i_alert__last_sent"); DbUpgradeUtils.dropKeysIfExist(conn, "alert", indexList, false); //Now add index. try(PreparedStatement pstmt = conn.prepareStatement("ALTER TABLE `cloud`.`alert` ADD INDEX `i_alert__last_sent`(`last_sent`)");) { pstmt.executeUpdate(); s_logger.debug("Added index i_alert__last_sent for table alert"); } catch (SQLException e) { throw new CloudRuntimeException("Unable to add index i_alert__last_sent to alert table for the column last_sent", e); } } private void dropUploadTable(Connection conn) { try(PreparedStatement pstmt0 = conn.prepareStatement("SELECT url, created, type_id, host_id from upload where type=?");) { // Read upload table - Templates s_logger.debug("Populating template_store_ref table"); pstmt0.setString(1, "TEMPLATE"); try(ResultSet rs0 = pstmt0.executeQuery();) { try(PreparedStatement pstmt1 = conn.prepareStatement("UPDATE template_store_ref SET download_url=?, download_url_created=? where template_id=? and store_id=?");) { //Update template_store_ref while (rs0.next()) { pstmt1.setString(1, rs0.getString("url")); pstmt1.setDate(2, rs0.getDate("created")); pstmt1.setLong(3, rs0.getLong("type_id")); pstmt1.setLong(4, rs0.getLong("host_id")); pstmt1.executeUpdate(); } // Read upload table - Volumes s_logger.debug("Populating volume store ref table"); try(PreparedStatement pstmt2 = conn.prepareStatement("SELECT url, created, type_id, host_id, install_path from upload where type=?");) { pstmt2.setString(1, "VOLUME"); try(ResultSet rs2 = pstmt2.executeQuery();) { try(PreparedStatement pstmt3 = conn.prepareStatement("INSERT IGNORE INTO volume_store_ref (volume_id, store_id, zone_id, created, state, download_url, download_url_created, install_path) VALUES (?,?,?,?,?,?,?,?)");) { //insert into template_store_ref while (rs2.next()) { pstmt3.setLong(1, rs2.getLong("type_id")); pstmt3.setLong(2, rs2.getLong("host_id")); pstmt3.setLong(3, 1l);// ??? pstmt3.setDate(4, rs2.getDate("created")); pstmt3.setString(5, "Ready"); pstmt3.setString(6, rs2.getString("url")); pstmt3.setDate(7, rs2.getDate("created")); pstmt3.setString(8, rs2.getString("install_path")); pstmt3.executeUpdate(); } }catch (SQLException e) { throw new CloudRuntimeException("Unable add date into template/volume store ref from upload table.", e); } }catch (SQLException e) { throw new CloudRuntimeException("Unable add date into template/volume store ref from upload table.", e); } }catch (SQLException e) { throw new CloudRuntimeException("Unable add date into template/volume store ref from upload table.", e); } }catch (SQLException e) { throw new CloudRuntimeException("Unable add date into template/volume store ref from upload table.", e); } }catch (SQLException e) { throw new CloudRuntimeException("Unable add date into template/volume store ref from upload table.", e); } } catch (SQLException e) { throw new CloudRuntimeException("Unable add date into template/volume store ref from upload table.", e); } } //KVM snapshot flag: only turn on if Customers is using snapshot; private void setKVMSnapshotFlag(Connection conn) { s_logger.debug("Verify and set the KVM snapshot flag if snapshot was used. "); try(PreparedStatement pstmt = conn.prepareStatement("select count(*) from `cloud`.`snapshots` where hypervisor_type = 'KVM'");) { int numRows = 0; try(ResultSet rs = pstmt.executeQuery();) { if (rs.next()) { numRows = rs.getInt(1); } if (numRows > 0) { //Add the configuration flag try(PreparedStatement update_pstmt = conn.prepareStatement("UPDATE `cloud`.`configuration` SET value = ? WHERE name = 'kvm.snapshot.enabled'");) { update_pstmt.setString(1, "true"); update_pstmt.executeUpdate(); }catch (SQLException e) { throw new CloudRuntimeException("Failed to read the snapshot table for KVM upgrade. ", e); } } }catch (SQLException e) { throw new CloudRuntimeException("Failed to read the snapshot table for KVM upgrade. ", e); } } catch (SQLException e) { throw new CloudRuntimeException("Failed to read the snapshot table for KVM upgrade. ", e); } s_logger.debug("Done set KVM snapshot flag. "); } private void updatePrimaryStore(Connection conn) { try(PreparedStatement sql = conn.prepareStatement("update storage_pool set storage_provider_name = ? , scope = ? where pool_type = 'Filesystem' or pool_type = 'LVM'");) { sql.setString(1, DataStoreProvider.DEFAULT_PRIMARY); sql.setString(2, "HOST"); sql.executeUpdate(); try(PreparedStatement sql2 = conn.prepareStatement("update storage_pool set storage_provider_name = ? , scope = ? where pool_type != 'Filesystem' and pool_type != 'LVM'");) { sql2.setString(1, DataStoreProvider.DEFAULT_PRIMARY); sql2.setString(2, "CLUSTER"); sql2.executeUpdate(); }catch (SQLException e) { throw new CloudRuntimeException("Failed to upgrade vm template data store uuid: " + e.toString()); } } catch (SQLException e) { throw new CloudRuntimeException("Failed to upgrade vm template data store uuid: " + e.toString()); } } //update the cluster_details table with default overcommit ratios. private void updateOverCommitRatioClusterDetails(Connection conn) { try ( PreparedStatement pstmt = conn.prepareStatement("select id, hypervisor_type from `cloud`.`cluster` WHERE removed IS NULL"); PreparedStatement pstmt1 = conn.prepareStatement("INSERT INTO `cloud`.`cluster_details` (cluster_id, name, value) VALUES(?, 'cpuOvercommitRatio', ?)"); PreparedStatement pstmt2 = conn.prepareStatement("INSERT INTO `cloud`.`cluster_details` (cluster_id, name, value) VALUES(?, 'memoryOvercommitRatio', ?)"); PreparedStatement pstmt3 = conn.prepareStatement("select value from `cloud`.`configuration` where name=?");) { String global_cpu_overprovisioning_factor = "1"; String global_mem_overprovisioning_factor = "1"; pstmt3.setString(1, "cpu.overprovisioning.factor"); try (ResultSet rscpu_global = pstmt3.executeQuery();) { if (rscpu_global.next()) global_cpu_overprovisioning_factor = rscpu_global.getString(1); } pstmt3.setString(1, "mem.overprovisioning.factor"); try (ResultSet rsmem_global = pstmt3.executeQuery();) { if (rsmem_global.next()) global_mem_overprovisioning_factor = rsmem_global.getString(1); } try (ResultSet rs1 = pstmt.executeQuery();) { while (rs1.next()) { long id = rs1.getLong(1); String hypervisor_type = rs1.getString(2); if (HypervisorType.VMware.toString().equalsIgnoreCase(hypervisor_type)) { pstmt1.setLong(1, id); pstmt1.setString(2, global_cpu_overprovisioning_factor); pstmt1.execute(); pstmt2.setLong(1, id); pstmt2.setString(2, global_mem_overprovisioning_factor); pstmt2.execute(); } else { //update cluster_details table with the default overcommit ratios. pstmt1.setLong(1, id); pstmt1.setString(2, global_cpu_overprovisioning_factor); pstmt1.execute(); pstmt2.setLong(1, id); pstmt2.setString(2, "1"); pstmt2.execute(); } } } } catch (SQLException e) { throw new CloudRuntimeException("Unable to update cluster_details with default overcommit ratios.", e); } } @Override public File[] getCleanupScripts() { String script = Script.findScript("", "db/schema-410to420-cleanup.sql"); if (script == null) { throw new CloudRuntimeException("Unable to find db/schema-410to420-cleanup.sql"); } return new File[] {new File(script)}; } private String getNewLabel(ResultSet rs, String oldParamValue) { int separatorIndex; String oldGuestLabel; String newGuestLabel = oldParamValue; try { // No need to iterate because the global param setting applies to all physical networks irrespective of traffic type if ((rs != null) && (rs.next())) { oldGuestLabel = rs.getString("vmware_network_label"); // guestLabel is in format [[<VSWITCHNAME>],VLANID] separatorIndex = oldGuestLabel.indexOf(","); if (separatorIndex > -1) { newGuestLabel += oldGuestLabel.substring(separatorIndex); } } } catch (SQLException e) { s_logger.error(new CloudRuntimeException("Failed to read vmware_network_label : " + e)); } return newGuestLabel; } private void upgradeVmwareLabels(Connection conn) { String newLabel; String trafficType = null; String trafficTypeVswitchParam; String trafficTypeVswitchParamValue; try (PreparedStatement pstmt = conn.prepareStatement("select name,value from `cloud`.`configuration` where category='Hidden' and value is not NULL and name REGEXP 'vmware*.vswitch';");) { // update the existing vmware traffic labels try(ResultSet rsParams = pstmt.executeQuery();) { while (rsParams.next()) { trafficTypeVswitchParam = rsParams.getString("name"); trafficTypeVswitchParamValue = rsParams.getString("value"); // When upgraded from 4.0 to 4.1 update physical network traffic label with trafficTypeVswitchParam if (trafficTypeVswitchParam.equals("vmware.private.vswitch")) { trafficType = "Management"; //TODO(sateesh): Ignore storage traffic, as required physical network already implemented, anything else tobe done? } else if (trafficTypeVswitchParam.equals("vmware.public.vswitch")) { trafficType = "Public"; } else if (trafficTypeVswitchParam.equals("vmware.guest.vswitch")) { trafficType = "Guest"; } try(PreparedStatement sel_pstmt = conn.prepareStatement("select physical_network_id, traffic_type, vmware_network_label from physical_network_traffic_types where vmware_network_label is not NULL and traffic_type=?;");) { pstmt.setString(1, trafficType); try(ResultSet rsLabel = sel_pstmt.executeQuery();) { newLabel = getNewLabel(rsLabel, trafficTypeVswitchParamValue); try(PreparedStatement update_pstmt = conn.prepareStatement("update physical_network_traffic_types set vmware_network_label = ? where traffic_type = ? and vmware_network_label is not NULL;");) { s_logger.debug("Updating vmware label for " + trafficType + " traffic. Update SQL statement is " + pstmt); pstmt.setString(1, newLabel); pstmt.setString(2, trafficType); update_pstmt.executeUpdate(); }catch (SQLException e) { throw new CloudRuntimeException("Unable to set vmware traffic labels ", e); } }catch (SQLException e) { throw new CloudRuntimeException("Unable to set vmware traffic labels ", e); } }catch (SQLException e) { throw new CloudRuntimeException("Unable to set vmware traffic labels ", e); } } }catch (SQLException e) { throw new CloudRuntimeException("Unable to set vmware traffic labels ", e); } } catch (SQLException e) { throw new CloudRuntimeException("Unable to set vmware traffic labels ", e); } } private void persistLegacyZones(Connection conn) { List<Long> listOfLegacyZones = new ArrayList<Long>(); List<Long> listOfNonLegacyZones = new ArrayList<Long>(); Map<String, ArrayList<Long>> dcToZoneMap = new HashMap<String, ArrayList<Long>>(); ResultSet clusters = null; Long zoneId; Long clusterId; ArrayList<String> dcList = null; String clusterHypervisorType; boolean legacyZone; boolean ignoreZone; Long count; String dcOfPreviousCluster = null; String dcOfCurrentCluster = null; String[] tokens; String url; String vc = ""; String dcName = ""; try (PreparedStatement pstmt = conn.prepareStatement("select id from `cloud`.`data_center` where removed is NULL");) { try (ResultSet rs = pstmt.executeQuery();) { while (rs.next()) { zoneId = rs.getLong("id"); try (PreparedStatement clustersQuery = conn.prepareStatement("select id, hypervisor_type from `cloud`.`cluster` where removed is NULL AND data_center_id=?");) { clustersQuery.setLong(1, zoneId); legacyZone = false; ignoreZone = true; dcList = new ArrayList<String>(); count = 0L; // Legacy zone term is meant only for VMware // Legacy zone is a zone with atleast 2 clusters & with multiple DCs or VCs clusters = clustersQuery.executeQuery(); if (!clusters.next()) { continue; // Ignore the zone without any clusters } else { dcOfPreviousCluster = null; dcOfCurrentCluster = null; do { clusterHypervisorType = clusters.getString("hypervisor_type"); clusterId = clusters.getLong("id"); if (clusterHypervisorType.equalsIgnoreCase("VMware")) { ignoreZone = false; try (PreparedStatement clusterDetailsQuery = conn .prepareStatement("select value from `cloud`.`cluster_details` where name='url' and cluster_id=?");) { clusterDetailsQuery.setLong(1, clusterId); try (ResultSet clusterDetails = clusterDetailsQuery.executeQuery();) { clusterDetails.next(); url = clusterDetails.getString("value"); tokens = url.split("/"); // url format - http://vcenter/dc/cluster vc = tokens[2]; dcName = tokens[3]; dcOfPreviousCluster = dcOfCurrentCluster; dcOfCurrentCluster = dcName + "@" + vc; if (!dcList.contains(dcOfCurrentCluster)) { dcList.add(dcOfCurrentCluster); } if (count > 0) { if (!dcOfPreviousCluster.equalsIgnoreCase(dcOfCurrentCluster)) { legacyZone = true; s_logger.debug("Marking the zone " + zoneId + " as legacy zone."); } } } catch (SQLException e) { throw new CloudRuntimeException("Unable add zones to cloud.legacyzones table.", e); } } catch (SQLException e) { throw new CloudRuntimeException("Unable add zones to cloud.legacyzones table.", e); } } else { s_logger.debug("Ignoring zone " + zoneId + " with hypervisor type " + clusterHypervisorType); break; } count++; } while (clusters.next()); if (ignoreZone) { continue; // Ignore the zone with hypervisors other than VMware } } } catch (SQLException e) { throw new CloudRuntimeException("persistLegacyZones:Exception:" + e.getMessage(), e); } if (legacyZone) { listOfLegacyZones.add(zoneId); } else { listOfNonLegacyZones.add(zoneId); } for (String dc : dcList) { ArrayList<Long> dcZones = new ArrayList<Long>(); if (dcToZoneMap.get(dc) != null) { dcZones = dcToZoneMap.get(dc); } dcZones.add(zoneId); dcToZoneMap.put(dc, dcZones); } } // If a VMware datacenter in a vCenter maps to more than 1 CloudStack zone, mark all the zones it is mapped to as legacy for (Map.Entry<String, ArrayList<Long>> entry : dcToZoneMap.entrySet()) { if (entry.getValue().size() > 1) { for (Long newLegacyZone : entry.getValue()) { if (listOfNonLegacyZones.contains(newLegacyZone)) { listOfNonLegacyZones.remove(newLegacyZone); listOfLegacyZones.add(newLegacyZone); } } } } updateLegacyZones(conn, listOfLegacyZones); updateNonLegacyZones(conn, listOfNonLegacyZones); } catch (SQLException e) { s_logger.error("Unable to discover legacy zones." + e.getMessage(),e); throw new CloudRuntimeException("Unable to discover legacy zones." + e.getMessage(), e); } }catch (SQLException e) { s_logger.error("Unable to discover legacy zones." + e.getMessage(),e); throw new CloudRuntimeException("Unable to discover legacy zones." + e.getMessage(), e); } } private void updateLegacyZones(Connection conn, List<Long> zones) { //Insert legacy zones into table for legacy zones. try (PreparedStatement legacyZonesQuery = conn.prepareStatement("INSERT INTO `cloud`.`legacy_zones` (zone_id) VALUES (?)");){ for (Long zoneId : zones) { legacyZonesQuery.setLong(1, zoneId); legacyZonesQuery.executeUpdate(); s_logger.debug("Inserted zone " + zoneId + " into cloud.legacyzones table"); } } catch (SQLException e) { throw new CloudRuntimeException("Unable add zones to cloud.legacyzones table.", e); } } private void updateNonLegacyZones(Connection conn, List<Long> zones) { try { for (Long zoneId : zones) { s_logger.debug("Discovered non-legacy zone " + zoneId + ". Processing the zone to associate with VMware datacenter."); // All clusters in a non legacy zone will belong to the same VMware DC, hence pick the first cluster try (PreparedStatement clustersQuery = conn.prepareStatement("select id from `cloud`.`cluster` where removed is NULL AND data_center_id=?");) { clustersQuery.setLong(1, zoneId); try (ResultSet clusters = clustersQuery.executeQuery();) { clusters.next(); Long clusterId = clusters.getLong("id"); // Get VMware datacenter details from cluster_details table String user = null; String password = null; String url = null; try (PreparedStatement clusterDetailsQuery = conn.prepareStatement("select name, value from `cloud`.`cluster_details` where cluster_id=?");) { clusterDetailsQuery.setLong(1, clusterId); try (ResultSet clusterDetails = clusterDetailsQuery.executeQuery();) { while (clusterDetails.next()) { String key = clusterDetails.getString(1); String value = clusterDetails.getString(2); if (key.equalsIgnoreCase("username")) { user = value; } else if (key.equalsIgnoreCase("password")) { password = value; } else if (key.equalsIgnoreCase("url")) { url = value; } } String[] tokens = url.split("/"); // url format - http://vcenter/dc/cluster String vc = tokens[2]; String dcName = tokens[3]; String guid = dcName + "@" + vc; try (PreparedStatement insertVmWareDC = conn .prepareStatement("INSERT INTO `cloud`.`vmware_data_center` (uuid, name, guid, vcenter_host, username, password) values(?, ?, ?, ?, ?, ?)");) { insertVmWareDC.setString(1, UUID.randomUUID().toString()); insertVmWareDC.setString(2, dcName); insertVmWareDC.setString(3, guid); insertVmWareDC.setString(4, vc); insertVmWareDC.setString(5, user); insertVmWareDC.setString(6, password); insertVmWareDC.executeUpdate(); } try (PreparedStatement selectVmWareDC = conn.prepareStatement("SELECT id FROM `cloud`.`vmware_data_center` where guid=?");) { selectVmWareDC.setString(1, guid); try (ResultSet vmWareDcInfo = selectVmWareDC.executeQuery();) { Long vmwareDcId = -1L; if (vmWareDcInfo.next()) { vmwareDcId = vmWareDcInfo.getLong("id"); } try (PreparedStatement insertMapping = conn .prepareStatement("INSERT INTO `cloud`.`vmware_data_center_zone_map` (zone_id, vmware_data_center_id) values(?, ?)");) { insertMapping.setLong(1, zoneId); insertMapping.setLong(2, vmwareDcId); insertMapping.executeUpdate(); } } } } } } } } } catch (SQLException e) { String msg = "Unable to update non legacy zones." + e.getMessage(); s_logger.error(msg); throw new CloudRuntimeException(msg, e); } } private void createPlaceHolderNics(Connection conn) { try (PreparedStatement pstmt = conn.prepareStatement("SELECT network_id, gateway, ip4_address FROM `cloud`.`nics` WHERE reserver_name IN ('DirectNetworkGuru','DirectPodBasedNetworkGuru') and vm_type='DomainRouter' AND removed IS null");) { try(ResultSet rs = pstmt.executeQuery();) { while (rs.next()) { Long networkId = rs.getLong(1); String gateway = rs.getString(2); String ip = rs.getString(3); String uuid = UUID.randomUUID().toString(); //Insert placeholder nic for each Domain router nic in Shared network try(PreparedStatement insert_pstmt = conn.prepareStatement("INSERT INTO `cloud`.`nics` (uuid, ip4_address, gateway, network_id, state, strategy, vm_type, default_nic, created) VALUES (?, ?, ?, ?, 'Reserved', 'PlaceHolder', 'DomainRouter', 0, now())");) { insert_pstmt.setString(1, uuid); insert_pstmt.setString(2, ip); insert_pstmt.setString(3, gateway); insert_pstmt.setLong(4, networkId); insert_pstmt.executeUpdate(); }catch (SQLException e) { throw new CloudRuntimeException("Unable to create placeholder nics", e); } s_logger.debug("Created placeholder nic for the ipAddress " + ip + " and network " + networkId); } }catch (SQLException e) { throw new CloudRuntimeException("Unable to create placeholder nics", e); } } catch (SQLException e) { throw new CloudRuntimeException("Unable to create placeholder nics", e); } } private void updateRemoteAccessVpn(Connection conn) { try(PreparedStatement pstmt = conn.prepareStatement("SELECT vpn_server_addr_id FROM `cloud`.`remote_access_vpn`");) { try(ResultSet rs = pstmt.executeQuery();) { long id = 1; while (rs.next()) { String uuid = UUID.randomUUID().toString(); Long ipId = rs.getLong(1); try(PreparedStatement update_pstmt = conn.prepareStatement("UPDATE `cloud`.`remote_access_vpn` set uuid=?, id=? where vpn_server_addr_id=?");) { update_pstmt.setString(1, uuid); update_pstmt.setLong(2, id); update_pstmt.setLong(3, ipId); update_pstmt.executeUpdate(); id++; }catch (SQLException e) { throw new CloudRuntimeException("Unable to update id/uuid of remote_access_vpn table", e); } } }catch (SQLException e) { throw new CloudRuntimeException("Unable to update id/uuid of remote_access_vpn table", e); } } catch (SQLException e) { throw new CloudRuntimeException("Unable to update id/uuid of remote_access_vpn table", e); } } private void addEgressFwRulesForSRXGuestNw(Connection conn) { ResultSet rs = null; try(PreparedStatement pstmt = conn.prepareStatement("select network_id FROM `cloud`.`ntwk_service_map` where service='Firewall' and provider='JuniperSRX' ");) { rs = pstmt.executeQuery(); while (rs.next()) { long netId = rs.getLong(1); //checking for Isolated OR Virtual try(PreparedStatement sel_net_pstmt = conn.prepareStatement("select account_id, domain_id FROM `cloud`.`networks` where (guest_type='Isolated' OR guest_type='Virtual') and traffic_type='Guest' and vpc_id is NULL and (state='implemented' OR state='Shutdown') and id=? ");) { sel_net_pstmt.setLong(1, netId); s_logger.debug("Getting account_id, domain_id from networks table: "); try(ResultSet rsNw = pstmt.executeQuery();) { if (rsNw.next()) { long accountId = rsNw.getLong(1); long domainId = rsNw.getLong(2); //Add new rule for the existing networks s_logger.debug("Adding default egress firewall rule for network " + netId); try (PreparedStatement insert_pstmt = conn.prepareStatement("INSERT INTO firewall_rules (uuid, state, protocol, purpose, account_id, domain_id, network_id, xid, created, traffic_type) VALUES (?, 'Active', 'all', 'Firewall', ?, ?, ?, ?, now(), 'Egress')");) { insert_pstmt.setString(1, UUID.randomUUID().toString()); insert_pstmt.setLong(2, accountId); insert_pstmt.setLong(3, domainId); insert_pstmt.setLong(4, netId); insert_pstmt.setString(5, UUID.randomUUID().toString()); s_logger.debug("Inserting default egress firewall rule " + insert_pstmt); insert_pstmt.executeUpdate(); } catch (SQLException e) { throw new CloudRuntimeException("Unable to set egress firewall rules ", e); } try (PreparedStatement sel_firewall_pstmt = conn.prepareStatement("select id from firewall_rules where protocol='all' and network_id=?");) { sel_firewall_pstmt.setLong(1, netId); try (ResultSet rsId = sel_firewall_pstmt.executeQuery();) { long firewallRuleId; if (rsId.next()) { firewallRuleId = rsId.getLong(1); try (PreparedStatement insert_pstmt = conn.prepareStatement("insert into firewall_rules_cidrs (firewall_rule_id,source_cidr) values (?, '0.0.0.0/0')");) { insert_pstmt.setLong(1, firewallRuleId); s_logger.debug("Inserting rule for cidr 0.0.0.0/0 for the new Firewall rule id=" + firewallRuleId + " with statement " + insert_pstmt); insert_pstmt.executeUpdate(); } catch (SQLException e) { throw new CloudRuntimeException("Unable to set egress firewall rules ", e); } } } catch (SQLException e) { throw new CloudRuntimeException("Unable to set egress firewall rules ", e); } } catch (SQLException e) { throw new CloudRuntimeException("Unable to set egress firewall rules ", e); } } } } } } catch (SQLException e) { throw new CloudRuntimeException("Unable to set egress firewall rules ", e); } } private void upgradeEIPNetworkOfferings(Connection conn) { try (PreparedStatement pstmt = conn.prepareStatement("select id, elastic_ip_service from `cloud`.`network_offerings` where traffic_type='Guest'");) { try(ResultSet rs = pstmt.executeQuery();) { while (rs.next()) { long id = rs.getLong(1); // check if elastic IP service is enabled for network offering if (rs.getLong(2) != 0) { //update network offering with eip_associate_public_ip set to true try(PreparedStatement update_pstmt = conn.prepareStatement("UPDATE `cloud`.`network_offerings` set eip_associate_public_ip=? where id=?");) { update_pstmt.setBoolean(1, true); update_pstmt.setLong(2, id); update_pstmt.executeUpdate(); }catch (SQLException e) { throw new CloudRuntimeException("Unable to set elastic_ip_service for network offerings with EIP service enabled.", e); } } } }catch (SQLException e) { throw new CloudRuntimeException("Unable to set elastic_ip_service for network offerings with EIP service enabled.", e); } } catch (SQLException e) { throw new CloudRuntimeException("Unable to set elastic_ip_service for network offerings with EIP service enabled.", e); } } private void updateNetworkACLs(Connection conn) { //Fetch all VPC Tiers //For each tier create a network ACL and move all the acl_items to network_acl_item table // If there are no acl_items for a tier, associate it with default ACL s_logger.debug("Updating network ACLs"); //1,2 are default acl Ids, start acl Ids from 3 long nextAclId = 3; String sqlSelectNetworkIds = "SELECT id, vpc_id, uuid FROM `cloud`.`networks` where vpc_id is not null and removed is null"; String sqlSelectFirewallRules = "SELECT id, uuid, start_port, end_port, state, protocol, icmp_code, icmp_type, created, traffic_type FROM `cloud`.`firewall_rules` where network_id = ? and purpose = 'NetworkACL'"; String sqlInsertNetworkAcl = "INSERT INTO `cloud`.`network_acl` (id, uuid, vpc_id, description, name) values (?, UUID(), ? , ?, ?)"; String sqlSelectFirewallCidrs = "SELECT id, source_cidr FROM `cloud`.`firewall_rules_cidrs` where firewall_rule_id = ?"; String sqlDeleteFirewallCidr = "DELETE FROM `cloud`.`firewall_rules_cidrs` where id = ?"; String sqlInsertNetworkAclItem = "INSERT INTO `cloud`.`network_acl_item` (uuid, acl_id, start_port, end_port, state, protocol, icmp_code, icmp_type, created, traffic_type, cidr, number, action) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ? )"; String sqlDeleteFirewallRules = "DELETE FROM `cloud`.`firewall_rules` where id = ?"; String sqlUpdateNetworks = "UPDATE `cloud`.`networks` set network_acl_id=? where id=?"; try ( PreparedStatement pstmtSelectNetworkIds = conn.prepareStatement(sqlSelectNetworkIds); PreparedStatement pstmtUpdate = conn.prepareStatement(sqlUpdateNetworks); PreparedStatement pstmtInsertNetworkAclItem = conn.prepareStatement(sqlInsertNetworkAclItem); PreparedStatement pstmtSelectFirewallRules = conn.prepareStatement(sqlSelectFirewallRules); PreparedStatement pstmtInsertNetworkAcl = conn.prepareStatement(sqlInsertNetworkAcl); PreparedStatement pstmtSelectFirewallCidrs = conn.prepareStatement(sqlSelectFirewallCidrs); PreparedStatement pstmtDeleteFirewallCidr = conn.prepareStatement(sqlDeleteFirewallCidr); PreparedStatement pstmtDeleteFirewallRules = conn.prepareStatement(sqlDeleteFirewallRules); ResultSet rsNetworkIds = pstmtSelectNetworkIds.executeQuery();) { //Get all VPC tiers while (rsNetworkIds.next()) { Long networkId = rsNetworkIds.getLong(1); s_logger.debug("Updating network ACLs for network: " + networkId); Long vpcId = rsNetworkIds.getLong(2); String tierUuid = rsNetworkIds.getString(3); pstmtSelectFirewallRules.setLong(1, networkId); boolean hasAcls = false; Long aclId = null; int number = 1; try (ResultSet rsAcls = pstmtSelectFirewallRules.executeQuery();) { while (rsAcls.next()) { if (!hasAcls) { hasAcls = true; aclId = nextAclId++; //create ACL for the tier s_logger.debug("Creating network ACL for tier: " + tierUuid); pstmtInsertNetworkAcl.setLong(1, aclId); pstmtInsertNetworkAcl.setLong(2, vpcId); pstmtInsertNetworkAcl.setString(3, "ACL for tier " + tierUuid); pstmtInsertNetworkAcl.setString(4, "tier_" + tierUuid); pstmtInsertNetworkAcl.executeUpdate(); } Long fwRuleId = rsAcls.getLong(1); String cidr = null; //get cidr from firewall_rules_cidrs pstmtSelectFirewallCidrs.setLong(1, fwRuleId); try (ResultSet rsCidr = pstmtSelectFirewallCidrs.executeQuery();) { while (rsCidr.next()) { Long cidrId = rsCidr.getLong(1); String sourceCidr = rsCidr.getString(2); if (cidr == null) { cidr = sourceCidr; } else { cidr += "," + sourceCidr; } //Delete cidr entry pstmtDeleteFirewallCidr.setLong(1, cidrId); pstmtDeleteFirewallCidr.executeUpdate(); } } String aclItemUuid = rsAcls.getString(2); //Move acl to network_acl_item table s_logger.debug("Moving firewall rule: " + aclItemUuid); //uuid pstmtInsertNetworkAclItem.setString(1, aclItemUuid); //aclId pstmtInsertNetworkAclItem.setLong(2, aclId); //Start port Integer startPort = rsAcls.getInt(3); if (rsAcls.wasNull()) { pstmtInsertNetworkAclItem.setNull(3, Types.INTEGER); } else { pstmtInsertNetworkAclItem.setLong(3, startPort); } //End port Integer endPort = rsAcls.getInt(4); if (rsAcls.wasNull()) { pstmtInsertNetworkAclItem.setNull(4, Types.INTEGER); } else { pstmtInsertNetworkAclItem.setLong(4, endPort); } //State String state = rsAcls.getString(5); pstmtInsertNetworkAclItem.setString(5, state); //protocol String protocol = rsAcls.getString(6); pstmtInsertNetworkAclItem.setString(6, protocol); //icmp_code Integer icmpCode = rsAcls.getInt(7); if (rsAcls.wasNull()) { pstmtInsertNetworkAclItem.setNull(7, Types.INTEGER); } else { pstmtInsertNetworkAclItem.setLong(7, icmpCode); } //icmp_type Integer icmpType = rsAcls.getInt(8); if (rsAcls.wasNull()) { pstmtInsertNetworkAclItem.setNull(8, Types.INTEGER); } else { pstmtInsertNetworkAclItem.setLong(8, icmpType); } //created Date created = rsAcls.getDate(9); pstmtInsertNetworkAclItem.setDate(9, created); //traffic type String trafficType = rsAcls.getString(10); pstmtInsertNetworkAclItem.setString(10, trafficType); //cidr pstmtInsertNetworkAclItem.setString(11, cidr); //number pstmtInsertNetworkAclItem.setInt(12, number++); //action pstmtInsertNetworkAclItem.setString(13, "Allow"); pstmtInsertNetworkAclItem.executeUpdate(); //Delete firewall rule pstmtDeleteFirewallRules.setLong(1, fwRuleId); pstmtDeleteFirewallRules.executeUpdate(); } } if (!hasAcls) { //no network ACls for this network. // Assign default Deny ACL aclId = NetworkACL.DEFAULT_DENY; } //Assign acl to network pstmtUpdate.setLong(1, aclId); pstmtUpdate.setLong(2, networkId); pstmtUpdate.executeUpdate(); } s_logger.debug("Done updating network ACLs "); } catch (SQLException e) { throw new CloudRuntimeException("Unable to move network acls from firewall rules table to network_acl_item table", e); } } private void updateGlobalDeploymentPlanner(Connection conn) { try (PreparedStatement pstmt = conn.prepareStatement("select value from `cloud`.`configuration` where name = 'vm.allocation.algorithm'");){ try(ResultSet rs = pstmt.executeQuery();) { while (rs.next()) { String globalValue = rs.getString(1); String plannerName = "FirstFitPlanner"; if (globalValue != null) { if (globalValue.equals(DeploymentPlanner.AllocationAlgorithm.random.toString())) { plannerName = "FirstFitPlanner"; } else if (globalValue.equals(DeploymentPlanner.AllocationAlgorithm.firstfit.toString())) { plannerName = "FirstFitPlanner"; } else if (globalValue.equals(DeploymentPlanner.AllocationAlgorithm.userconcentratedpod_firstfit.toString())) { plannerName = "UserConcentratedPodPlanner"; } else if (globalValue.equals(DeploymentPlanner.AllocationAlgorithm.userconcentratedpod_random.toString())) { plannerName = "UserConcentratedPodPlanner"; } else if (globalValue.equals(DeploymentPlanner.AllocationAlgorithm.userdispersing.toString())) { plannerName = "UserDispersingPlanner"; } } // update vm.deployment.planner global config try (PreparedStatement update_pstmt = conn.prepareStatement("UPDATE `cloud`.`configuration` set value=? where name = 'vm.deployment.planner'");) { update_pstmt.setString(1, plannerName); update_pstmt.executeUpdate(); } catch (SQLException e) { throw new CloudRuntimeException("Unable to set vm.deployment.planner global config", e); } } }catch (SQLException e) { throw new CloudRuntimeException("Unable to set vm.deployment.planner global config", e); } } catch (SQLException e) { throw new CloudRuntimeException("Unable to set vm.deployment.planner global config", e); } } private void upgradeDefaultVpcOffering(Connection conn) { try(PreparedStatement pstmt = conn.prepareStatement("select distinct map.vpc_offering_id from `cloud`.`vpc_offering_service_map` map, `cloud`.`vpc_offerings` off where off.id=map.vpc_offering_id AND service='Lb'");) { try(ResultSet rs = pstmt.executeQuery();) { while (rs.next()) { long id = rs.getLong(1); //Add internal LB vm as a supported provider for the load balancer service try(PreparedStatement insert_pstmt = conn.prepareStatement("INSERT INTO `cloud`.`vpc_offering_service_map` (vpc_offering_id, service, provider) VALUES (?,?,?)");) { insert_pstmt.setLong(1, id); insert_pstmt.setString(2, "Lb"); insert_pstmt.setString(3, "InternalLbVm"); insert_pstmt.executeUpdate(); }catch (SQLException e) { throw new CloudRuntimeException("Unable update the default VPC offering with the internal lb service", e); } } }catch (SQLException e) { throw new CloudRuntimeException("Unable update the default VPC offering with the internal lb service", e); } } catch (SQLException e) { throw new CloudRuntimeException("Unable update the default VPC offering with the internal lb service", e); } } private void upgradePhysicalNtwksWithInternalLbProvider(Connection conn) { try (PreparedStatement pstmt = conn.prepareStatement("SELECT id FROM `cloud`.`physical_network` where removed is null");){ try(ResultSet rs = pstmt.executeQuery();) { while (rs.next()) { long pNtwkId = rs.getLong(1); String uuid = UUID.randomUUID().toString(); //Add internal LB VM to the list of physical network service providers try(PreparedStatement insert_pstmt = conn.prepareStatement("INSERT INTO `cloud`.`physical_network_service_providers` " + "(uuid, physical_network_id, provider_name, state, load_balance_service_provided, destination_physical_network_id)" + " VALUES (?, ?, 'InternalLbVm', 'Enabled', 1, 0)");) { insert_pstmt.setString(1, uuid); insert_pstmt.setLong(2, pNtwkId); insert_pstmt.executeUpdate(); //Add internal lb vm to the list of physical network elements try (PreparedStatement pstmt1 = conn.prepareStatement("SELECT id FROM `cloud`.`physical_network_service_providers`" + " WHERE physical_network_id=? AND provider_name='InternalLbVm'");) { pstmt1.setLong(1, pNtwkId); try (ResultSet rs1 = pstmt1.executeQuery();) { while (rs1.next()) { long providerId = rs1.getLong(1); uuid = UUID.randomUUID().toString(); try(PreparedStatement insert_cloud_pstmt = conn.prepareStatement("INSERT INTO `cloud`.`virtual_router_providers` (nsp_id, uuid, type, enabled) VALUES (?, ?, 'InternalLbVm', 1)");) { insert_cloud_pstmt.setLong(1, providerId); insert_cloud_pstmt.setString(2, uuid); insert_cloud_pstmt.executeUpdate(); }catch (SQLException e) { throw new CloudRuntimeException("Unable to update existing physical networks with internal lb provider", e); } } } catch (SQLException e) { throw new CloudRuntimeException("Unable to update existing physical networks with internal lb provider", e); } } catch (SQLException e) { throw new CloudRuntimeException("Unable to update existing physical networks with internal lb provider", e); } } } }catch (SQLException e) { throw new CloudRuntimeException("Unable to update existing physical networks with internal lb provider", e); } } catch (SQLException e) { throw new CloudRuntimeException("Unable to update existing physical networks with internal lb provider", e); } } private void addHostDetailsIndex(Connection conn) { s_logger.debug("Checking if host_details index exists, if not we will add it"); try(PreparedStatement pstmt = conn.prepareStatement("SHOW INDEX FROM `cloud`.`host_details` where KEY_NAME = 'fk_host_details__host_id'");) { try(ResultSet rs = pstmt.executeQuery();) { if (rs.next()) { s_logger.debug("Index already exists on host_details - not adding new one"); } else { // add the index try(PreparedStatement pstmtUpdate = conn.prepareStatement("ALTER TABLE `cloud`.`host_details` ADD INDEX `fk_host_details__host_id` (`host_id`)");) { pstmtUpdate.executeUpdate(); s_logger.debug("Index did not exist on host_details - added new one"); }catch (SQLException e) { throw new CloudRuntimeException("Failed to check/update the host_details index ", e); } } }catch (SQLException e) { throw new CloudRuntimeException("Failed to check/update the host_details index ", e); } } catch (SQLException e) { throw new CloudRuntimeException("Failed to check/update the host_details index ", e); } } private void updateNetworksForPrivateGateways(Connection conn) { try(PreparedStatement pstmt = conn.prepareStatement("SELECT network_id, vpc_id FROM `cloud`.`vpc_gateways` WHERE type='Private' AND removed IS null");) { //1) get all non removed gateways try(ResultSet rs = pstmt.executeQuery();) { while (rs.next()) { Long networkId = rs.getLong(1); Long vpcId = rs.getLong(2); //2) Update networks with vpc_id if its set to NULL try (PreparedStatement update_pstmt = conn.prepareStatement("UPDATE `cloud`.`networks` set vpc_id=? where id=? and vpc_id is NULL and removed is NULL");) { update_pstmt.setLong(1, vpcId); update_pstmt.setLong(2, networkId); update_pstmt.executeUpdate(); } catch (SQLException e) { throw new CloudRuntimeException("Failed to update private networks with VPC id.", e); } } }catch (SQLException e) { throw new CloudRuntimeException("Failed to update private networks with VPC id.", e); } } catch (SQLException e) { throw new CloudRuntimeException("Failed to update private networks with VPC id.", e); } } private void removeFirewallServiceFromSharedNetworkOfferingWithSGService(Connection conn) { try(PreparedStatement pstmt = conn.prepareStatement("select id from `cloud`.`network_offerings` where unique_name='DefaultSharedNetworkOfferingWithSGService'");) { try(ResultSet rs = pstmt.executeQuery();) { while (rs.next()) { long id = rs.getLong(1); // remove Firewall service for SG shared network offering try(PreparedStatement del_pstmt = conn.prepareStatement("DELETE from `cloud`.`ntwk_offering_service_map` where network_offering_id=? and service='Firewall'");) { del_pstmt.setLong(1, id); del_pstmt.executeUpdate(); }catch (SQLException e) { throw new CloudRuntimeException("Unable to remove Firewall service for SG shared network offering.", e); } } }catch (SQLException e) { throw new CloudRuntimeException("Unable to remove Firewall service for SG shared network offering.", e); } } catch (SQLException e) { throw new CloudRuntimeException("Unable to remove Firewall service for SG shared network offering.", e); } } private void fix22xKVMSnapshots(Connection conn) { s_logger.debug("Updating KVM snapshots"); try (PreparedStatement pstmt = conn.prepareStatement("select id, backup_snap_id from `cloud`.`snapshots` where hypervisor_type='KVM' and removed is null and backup_snap_id is not null");) { try(ResultSet rs = pstmt.executeQuery();) { while (rs.next()) { long id = rs.getLong(1); String backUpPath = rs.getString(2); // Update Backup Path. Remove anything before /snapshots/ // e.g 22x Path /mnt/0f14da63-7033-3ca5-bdbe-fa62f4e2f38a/snapshots/1/2/6/i-2-6-VM_ROOT-6_20121219072022 // Above path should change to /snapshots/1/2/6/i-2-6-VM_ROOT-6_20121219072022 int index = backUpPath.indexOf("snapshots" + File.separator); if (index > 1) { String correctedPath = backUpPath.substring(index); s_logger.debug("Updating Snapshot with id: " + id + " original backup path: " + backUpPath + " updated backup path: " + correctedPath); try(PreparedStatement update_pstmt = conn.prepareStatement("UPDATE `cloud`.`snapshots` set backup_snap_id=? where id = ?");) { update_pstmt.setString(1, correctedPath); update_pstmt.setLong(2, id); update_pstmt.executeUpdate(); }catch (SQLException e) { throw new CloudRuntimeException("Unable to update backup id for KVM snapshots", e); } } } s_logger.debug("Done updating KVM snapshots"); }catch (SQLException e) { throw new CloudRuntimeException("Unable to update backup id for KVM snapshots", e); } } catch (SQLException e) { throw new CloudRuntimeException("Unable to update backup id for KVM snapshots", e); } } // Corrects upgrade for deployment with F5 and SRX devices (pre 3.0) to network offering & // network service provider paradigm private void correctExternalNetworkDevicesSetup(Connection conn) { PreparedStatement pNetworkStmt = null, f5DevicesStmt = null, srxDevicesStmt = null; ResultSet pNetworksResults = null, f5DevicesResult = null, srxDevicesResult = null; try ( PreparedStatement zoneSearchStmt = conn.prepareStatement("SELECT id, networktype FROM `cloud`.`data_center`"); ResultSet zoneResults = zoneSearchStmt.executeQuery(); ){ while (zoneResults.next()) { long zoneId = zoneResults.getLong(1); String networkType = zoneResults.getString(2); if (!com.cloud.dc.DataCenter.NetworkType.Advanced.toString().equalsIgnoreCase(networkType)) { continue; } pNetworkStmt = conn.prepareStatement("SELECT id FROM `cloud`.`physical_network` where data_center_id=?"); pNetworkStmt.setLong(1, zoneId); pNetworksResults = pNetworkStmt.executeQuery(); while (pNetworksResults.next()) { long physicalNetworkId = pNetworksResults.getLong(1); PreparedStatement fetchF5NspStmt = conn.prepareStatement("SELECT id from `cloud`.`physical_network_service_providers` where physical_network_id=" + physicalNetworkId + " and provider_name = 'F5BigIp'"); ResultSet rsF5NSP = fetchF5NspStmt.executeQuery(); boolean hasF5Nsp = rsF5NSP.next(); fetchF5NspStmt.close(); // if there is no 'F5BigIP' physical network service provider added into physical network then // add 'F5BigIP' as network service provider and add the entry in 'external_load_balancer_devices' if (!hasF5Nsp) { f5DevicesStmt = conn.prepareStatement("SELECT id FROM host WHERE data_center_id=? AND type = 'ExternalLoadBalancer' AND removed IS NULL"); f5DevicesStmt.setLong(1, zoneId); f5DevicesResult = f5DevicesStmt.executeQuery(); // add F5BigIP provider and provider instance to physical network if there are any external load // balancers added in the zone while (f5DevicesResult.next()) { long f5HostId = f5DevicesResult.getLong(1); ; addF5ServiceProvider(conn, physicalNetworkId, zoneId); addF5LoadBalancer(conn, f5HostId, physicalNetworkId); } } boolean hasSrxNsp = false; try (PreparedStatement fetchSRXNspStmt = conn.prepareStatement("SELECT id from `cloud`.`physical_network_service_providers` where physical_network_id=" + physicalNetworkId + " and provider_name = 'JuniperSRX'"); ResultSet rsSRXNSP = fetchSRXNspStmt.executeQuery();) { hasSrxNsp = rsSRXNSP.next(); } // if there is no 'JuniperSRX' physical network service provider added into physical network then // add 'JuniperSRX' as network service provider and add the entry in 'external_firewall_devices' if (!hasSrxNsp) { srxDevicesStmt = conn.prepareStatement("SELECT id FROM host WHERE data_center_id=? AND type = 'ExternalFirewall' AND removed IS NULL"); srxDevicesStmt.setLong(1, zoneId); srxDevicesResult = srxDevicesStmt.executeQuery(); // add JuniperSRX provider and provider instance to physical network if there are any external // firewall instances added in to the zone while (srxDevicesResult.next()) { long srxHostId = srxDevicesResult.getLong(1); // add SRX provider and provider instance to physical network addSrxServiceProvider(conn, physicalNetworkId, zoneId); addSrxFirewall(conn, srxHostId, physicalNetworkId); } } } } // not the network service provider has been provisioned in to physical network, mark all guest network // to be using network offering 'Isolated with external providers' fixZoneUsingExternalDevices(conn); } catch (SQLException e) { throw new CloudRuntimeException("Exception while adding PhysicalNetworks", e); } } private void addF5LoadBalancer(Connection conn, long hostId, long physicalNetworkId) { String insertF5 = "INSERT INTO `cloud`.`external_load_balancer_devices` (physical_network_id, host_id, provider_name, " + "device_name, capacity, is_dedicated, device_state, allocation_state, is_managed, uuid) VALUES ( ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)"; try(PreparedStatement pstmtUpdate = conn.prepareStatement(insertF5);) { s_logger.debug("Adding F5 Big IP load balancer with host id " + hostId + " in to physical network" + physicalNetworkId); pstmtUpdate.setLong(1, physicalNetworkId); pstmtUpdate.setLong(2, hostId); pstmtUpdate.setString(3, "F5BigIp"); pstmtUpdate.setString(4, "F5BigIpLoadBalancer"); pstmtUpdate.setLong(5, 0); pstmtUpdate.setBoolean(6, false); pstmtUpdate.setString(7, "Enabled"); pstmtUpdate.setString(8, "Shared"); pstmtUpdate.setBoolean(9, false); pstmtUpdate.setString(10, UUID.randomUUID().toString()); pstmtUpdate.executeUpdate(); } catch (SQLException e) { throw new CloudRuntimeException("Exception while adding F5 load balancer device", e); } } private void addSrxFirewall(Connection conn, long hostId, long physicalNetworkId) { String insertSrx = "INSERT INTO `cloud`.`external_firewall_devices` (physical_network_id, host_id, provider_name, " + "device_name, capacity, is_dedicated, device_state, allocation_state, uuid) VALUES ( ?, ?, ?, ?, ?, ?, ?, ?, ?)"; try(PreparedStatement pstmtUpdate = conn.prepareStatement(insertSrx);) { s_logger.debug("Adding SRX firewall device with host id " + hostId + " in to physical network" + physicalNetworkId); pstmtUpdate.setLong(1, physicalNetworkId); pstmtUpdate.setLong(2, hostId); pstmtUpdate.setString(3, "JuniperSRX"); pstmtUpdate.setString(4, "JuniperSRXFirewall"); pstmtUpdate.setLong(5, 0); pstmtUpdate.setBoolean(6, false); pstmtUpdate.setString(7, "Enabled"); pstmtUpdate.setString(8, "Shared"); pstmtUpdate.setString(9, UUID.randomUUID().toString()); pstmtUpdate.executeUpdate(); } catch (SQLException e) { throw new CloudRuntimeException("Exception while adding SRX firewall device ", e); } } private void addF5ServiceProvider(Connection conn, long physicalNetworkId, long zoneId) { String insertPNSP = "INSERT INTO `cloud`.`physical_network_service_providers` (`uuid`, `physical_network_id` , `provider_name`, `state` ," + "`destination_physical_network_id`, `vpn_service_provided`, `dhcp_service_provided`, `dns_service_provided`, `gateway_service_provided`," + "`firewall_service_provided`, `source_nat_service_provided`, `load_balance_service_provided`, `static_nat_service_provided`," + "`port_forwarding_service_provided`, `user_data_service_provided`, `security_group_service_provided`) VALUES (?,?,?,?,0,0,0,0,0,0,0,1,0,0,0,0)"; try(PreparedStatement pstmtUpdate = conn.prepareStatement(insertPNSP);) { // add physical network service provider - F5BigIp s_logger.debug("Adding PhysicalNetworkServiceProvider F5BigIp" + " in to physical network" + physicalNetworkId); pstmtUpdate.setString(1, UUID.randomUUID().toString()); pstmtUpdate.setLong(2, physicalNetworkId); pstmtUpdate.setString(3, "F5BigIp"); pstmtUpdate.setString(4, "Enabled"); pstmtUpdate.executeUpdate(); } catch (SQLException e) { throw new CloudRuntimeException("Exception while adding PhysicalNetworkServiceProvider F5BigIp", e); } } private void addSrxServiceProvider(Connection conn, long physicalNetworkId, long zoneId) { String insertPNSP = "INSERT INTO `cloud`.`physical_network_service_providers` (`uuid`, `physical_network_id` , `provider_name`, `state` ," + "`destination_physical_network_id`, `vpn_service_provided`, `dhcp_service_provided`, `dns_service_provided`, `gateway_service_provided`," + "`firewall_service_provided`, `source_nat_service_provided`, `load_balance_service_provided`, `static_nat_service_provided`," + "`port_forwarding_service_provided`, `user_data_service_provided`, `security_group_service_provided`) VALUES (?,?,?,?,0,0,0,0,1,1,1,0,1,1,0,0)"; try( PreparedStatement pstmtUpdate = conn.prepareStatement(insertPNSP);) { // add physical network service provider - JuniperSRX s_logger.debug("Adding PhysicalNetworkServiceProvider JuniperSRX"); pstmtUpdate.setString(1, UUID.randomUUID().toString()); pstmtUpdate.setLong(2, physicalNetworkId); pstmtUpdate.setString(3, "JuniperSRX"); pstmtUpdate.setString(4, "Enabled"); pstmtUpdate.executeUpdate(); } catch (SQLException e) { throw new CloudRuntimeException("Exception while adding PhysicalNetworkServiceProvider JuniperSRX", e); } } // This method does two things // // 1) ensure that networks using external load balancer/firewall in deployments prior to release 3.0 // has entry in network_external_lb_device_map and network_external_firewall_device_map // // 2) Some keys of host details for F5 and SRX devices were stored in Camel Case in 2.x releases. From 3.0 // they are made in lowercase. On upgrade change the host details name to lower case private void fixZoneUsingExternalDevices(Connection conn) { //Get zones to upgrade List<Long> zoneIds = new ArrayList<Long>(); ResultSet rs = null; long networkOfferingId, networkId; long f5DeviceId, f5HostId; long srxDevivceId, srxHostId; try(PreparedStatement sel_id_pstmt = conn.prepareStatement("select id from `cloud`.`data_center` where lb_provider='F5BigIp' or firewall_provider='JuniperSRX' or gateway_provider='JuniperSRX'");) { try(ResultSet sel_id_rs = sel_id_pstmt.executeQuery();) { while (sel_id_rs.next()) { zoneIds.add(sel_id_rs.getLong(1)); } }catch (SQLException e) { throw new CloudRuntimeException("fixZoneUsingExternalDevices:Exception:"+e.getMessage(), e); } } catch (SQLException e) { throw new CloudRuntimeException("fixZoneUsingExternalDevices:Exception:"+e.getMessage(), e); } if (zoneIds.size() == 0) { return; // no zones using F5 and SRX devices so return } // find the default network offering created for external devices during upgrade from 2.2.14 try(PreparedStatement sel_id_off_pstmt = conn.prepareStatement("select id from `cloud`.`network_offerings` where unique_name='Isolated with external providers' ");) { try(ResultSet sel_id_off_rs = sel_id_off_pstmt.executeQuery();) { if (sel_id_off_rs.first()) { networkOfferingId = sel_id_off_rs.getLong(1); } else { throw new CloudRuntimeException("fixZoneUsingExternalDevices:Exception"); } }catch (SQLException e) { throw new CloudRuntimeException("fixZoneUsingExternalDevices:Exception:"+e.getMessage(), e); } } catch (SQLException e) { throw new CloudRuntimeException("fixZoneUsingExternalDevices:Exception:"+e.getMessage(), e); } for (Long zoneId : zoneIds) { try { // find the F5 device id in the zone try(PreparedStatement sel_id_host_pstmt = conn.prepareStatement("SELECT id FROM host WHERE data_center_id=? AND type = 'ExternalLoadBalancer' AND removed IS NULL");) { sel_id_host_pstmt.setLong(1, zoneId); try(ResultSet sel_id_host_pstmt_rs = sel_id_host_pstmt.executeQuery();) { if (sel_id_host_pstmt_rs.first()) { f5HostId = sel_id_host_pstmt_rs.getLong(1); } else { throw new CloudRuntimeException("Cannot upgrade as there is no F5 load balancer device found in data center " + zoneId); } }catch (SQLException e) { throw new CloudRuntimeException("fixZoneUsingExternalDevices:Exception:"+e.getMessage(), e); } }catch (SQLException e) { throw new CloudRuntimeException("fixZoneUsingExternalDevices:Exception:"+e.getMessage(), e); } try(PreparedStatement sel_id_ext_pstmt = conn.prepareStatement("SELECT id FROM external_load_balancer_devices WHERE host_id=?");) { sel_id_ext_pstmt.setLong(1, f5HostId); try(ResultSet sel_id_ext_rs = sel_id_ext_pstmt.executeQuery();) { if (sel_id_ext_rs.first()) { f5DeviceId = sel_id_ext_rs.getLong(1); } else { throw new CloudRuntimeException("Cannot upgrade as there is no F5 load balancer device with host ID " + f5HostId + " found in external_load_balancer_device"); } }catch (SQLException e) { throw new CloudRuntimeException("fixZoneUsingExternalDevices:Exception:"+e.getMessage(), e); } }catch (SQLException e) { throw new CloudRuntimeException("fixZoneUsingExternalDevices:Exception:"+e.getMessage(), e); } // find the SRX device id in the zone try(PreparedStatement sel_id_hostdc_pstmt = conn.prepareStatement("SELECT id FROM host WHERE data_center_id=? AND type = 'ExternalFirewall' AND removed IS NULL");) { sel_id_hostdc_pstmt.setLong(1, zoneId); try(ResultSet sel_id_hostdc_pstmt_rs = sel_id_hostdc_pstmt.executeQuery();) { if (sel_id_hostdc_pstmt_rs.first()) { srxHostId = sel_id_hostdc_pstmt_rs.getLong(1); } else { throw new CloudRuntimeException("Cannot upgrade as there is no SRX firewall device found in data center " + zoneId); } }catch (SQLException e) { throw new CloudRuntimeException("fixZoneUsingExternalDevices:Exception:"+e.getMessage(), e); } }catch (SQLException e) { throw new CloudRuntimeException("fixZoneUsingExternalDevices:Exception:"+e.getMessage(), e); } try(PreparedStatement sel_id_ext_frwl_pstmt = conn.prepareStatement("SELECT id FROM external_firewall_devices WHERE host_id=?");) { sel_id_ext_frwl_pstmt.setLong(1, srxHostId); try(ResultSet sel_id_ext_frwl_pstmt_rs = sel_id_ext_frwl_pstmt.executeQuery();) { if (sel_id_ext_frwl_pstmt_rs.first()) { srxDevivceId = sel_id_ext_frwl_pstmt_rs.getLong(1); } else { throw new CloudRuntimeException("Cannot upgrade as there is no SRX firewall device found with host ID " + srxHostId + " found in external_firewall_devices"); } }catch (SQLException e) { throw new CloudRuntimeException("fixZoneUsingExternalDevices:Exception:"+e.getMessage(), e); } }catch (SQLException e) { throw new CloudRuntimeException("fixZoneUsingExternalDevices:Exception:"+e.getMessage(), e); } // check if network any uses F5 or SRX devices in the zone try(PreparedStatement sel_id_cloud_pstmt = conn.prepareStatement("select id from `cloud`.`networks` where guest_type='Virtual' and data_center_id=? and network_offering_id=? and removed IS NULL");) { sel_id_cloud_pstmt.setLong(1, zoneId); sel_id_cloud_pstmt.setLong(2, networkOfferingId); try(ResultSet sel_id_cloud_pstmt_rs = sel_id_cloud_pstmt.executeQuery();) { while (sel_id_cloud_pstmt_rs.next()) { // get the network Id networkId = sel_id_cloud_pstmt_rs.getLong(1); // add mapping for the network in network_external_lb_device_map String insertLbMapping = "INSERT INTO `cloud`.`network_external_lb_device_map` (uuid, network_id, external_load_balancer_device_id, created) VALUES ( ?, ?, ?, now())"; try (PreparedStatement insert_lb_stmt = conn.prepareStatement(insertLbMapping);) { insert_lb_stmt.setString(1, UUID.randomUUID().toString()); insert_lb_stmt.setLong(2, networkId); insert_lb_stmt.setLong(3, f5DeviceId); insert_lb_stmt.executeUpdate(); } catch (SQLException e) { throw new CloudRuntimeException("Unable create a mapping for the networks in network_external_lb_device_map and network_external_firewall_device_map", e); } s_logger.debug("Successfully added entry in network_external_lb_device_map for network " + networkId + " and F5 device ID " + f5DeviceId); // add mapping for the network in network_external_firewall_device_map String insertFwMapping = "INSERT INTO `cloud`.`network_external_firewall_device_map` (uuid, network_id, external_firewall_device_id, created) VALUES ( ?, ?, ?, now())"; try (PreparedStatement insert_ext_firewall_stmt = conn.prepareStatement(insertFwMapping);) { insert_ext_firewall_stmt.setString(1, UUID.randomUUID().toString()); insert_ext_firewall_stmt.setLong(2, networkId); insert_ext_firewall_stmt.setLong(3, srxDevivceId); insert_ext_firewall_stmt.executeUpdate(); } catch (SQLException e) { throw new CloudRuntimeException("Unable create a mapping for the networks in network_external_lb_device_map and network_external_firewall_device_map", e); } s_logger.debug("Successfully added entry in network_external_firewall_device_map for network " + networkId + " and SRX device ID " + srxDevivceId); } }catch (SQLException e) { throw new CloudRuntimeException("Unable create a mapping for the networks in network_external_lb_device_map and network_external_firewall_device_map", e); } }catch (SQLException e) { throw new CloudRuntimeException("Unable create a mapping for the networks in network_external_lb_device_map and network_external_firewall_device_map", e); } // update host details for F5 and SRX devices s_logger.debug("Updating the host details for F5 and SRX devices"); try(PreparedStatement sel_pstmt = conn.prepareStatement("SELECT host_id, name FROM `cloud`.`host_details` WHERE host_id=? OR host_id=?");) { sel_pstmt.setLong(1, f5HostId); sel_pstmt.setLong(2, srxHostId); try(ResultSet sel_rs = sel_pstmt.executeQuery();) { while (sel_rs.next()) { long hostId = sel_rs.getLong(1); String camlCaseName = sel_rs.getString(2); if (!(camlCaseName.equalsIgnoreCase("numRetries") || camlCaseName.equalsIgnoreCase("publicZone") || camlCaseName.equalsIgnoreCase("privateZone") || camlCaseName.equalsIgnoreCase("publicInterface") || camlCaseName.equalsIgnoreCase("privateInterface") || camlCaseName.equalsIgnoreCase("usageInterface"))) { continue; } String lowerCaseName = camlCaseName.toLowerCase(); try (PreparedStatement update_pstmt = conn.prepareStatement("update `cloud`.`host_details` set name=? where host_id=? AND name=?");) { update_pstmt.setString(1, lowerCaseName); update_pstmt.setLong(2, hostId); update_pstmt.setString(3, camlCaseName); update_pstmt.executeUpdate(); } catch (SQLException e) { throw new CloudRuntimeException("Unable create a mapping for the networks in network_external_lb_device_map and network_external_firewall_device_map", e); } } }catch (SQLException e) { throw new CloudRuntimeException("Unable create a mapping for the networks in network_external_lb_device_map and network_external_firewall_device_map", e); } }catch (SQLException e) { throw new CloudRuntimeException("Unable create a mapping for the networks in network_external_lb_device_map and network_external_firewall_device_map", e); } s_logger.debug("Successfully updated host details for F5 and SRX devices"); } catch (RuntimeException e) { throw new CloudRuntimeException("Unable create a mapping for the networks in network_external_lb_device_map and network_external_firewall_device_map", e); } s_logger.info("Successfully upgraded network using F5 and SRX devices to have a entry in the network_external_lb_device_map and network_external_firewall_device_map"); } } // migrate secondary storages NFS from host tables to image_store table private void migrateSecondaryStorageToImageStore(Connection conn) { String sqlSelectS3Count = "select count(*) from `cloud`.`s3`"; String sqlSelectSwiftCount = "select count(*) from `cloud`.`swift`"; String sqlInsertStoreDetail = "INSERT INTO `cloud`.`image_store_details` (store_id, name, value) values(?, ?, ?)"; String sqlUpdateHostAsRemoved = "UPDATE `cloud`.`host` SET removed = now() WHERE type = 'SecondaryStorage' and removed is null"; s_logger.debug("Migrating secondary storage to image store"); boolean hasS3orSwift = false; try ( PreparedStatement pstmtSelectS3Count = conn.prepareStatement(sqlSelectS3Count); PreparedStatement pstmtSelectSwiftCount = conn.prepareStatement(sqlSelectSwiftCount); PreparedStatement storeDetailInsert = conn.prepareStatement(sqlInsertStoreDetail); PreparedStatement storeInsert = conn.prepareStatement("INSERT INTO `cloud`.`image_store` (id, uuid, name, image_provider_name, protocol, url, data_center_id, scope, role, parent, total_size, created, removed) values(?, ?, ?, 'NFS', 'nfs', ?, ?, 'ZONE', ?, ?, ?, ?, ?)"); PreparedStatement nfsQuery = conn.prepareStatement("select id, uuid, url, data_center_id, parent, total_size, created, removed from `cloud`.`host` where type = 'SecondaryStorage'"); PreparedStatement pstmtUpdateHostAsRemoved = conn.prepareStatement(sqlUpdateHostAsRemoved); ResultSet rsSelectS3Count = pstmtSelectS3Count.executeQuery(); ResultSet rsSelectSwiftCount = pstmtSelectSwiftCount.executeQuery(); ResultSet rsNfs = nfsQuery.executeQuery(); ) { s_logger.debug("Checking if we need to migrate NFS secondary storage to image store or staging store"); int numRows = 0; if (rsSelectS3Count.next()) { numRows = rsSelectS3Count.getInt(1); } // check if there is swift storage if (rsSelectSwiftCount.next()) { numRows += rsSelectSwiftCount.getInt(1); } if (numRows > 0) { hasS3orSwift = true; } String store_role = "Image"; if (hasS3orSwift) { store_role = "ImageCache"; } s_logger.debug("Migrating NFS secondary storage to " + store_role + " store"); // migrate NFS secondary storage, for nfs, keep previous host_id as the store_id while (rsNfs.next()) { Long nfs_id = rsNfs.getLong("id"); String nfs_uuid = rsNfs.getString("uuid"); String nfs_url = rsNfs.getString("url"); String nfs_parent = rsNfs.getString("parent"); int nfs_dcid = rsNfs.getInt("data_center_id"); Long nfs_totalsize = rsNfs.getObject("total_size") != null ? rsNfs.getLong("total_size") : null; Date nfs_created = rsNfs.getDate("created"); Date nfs_removed = rsNfs.getDate("removed"); // insert entry in image_store table and image_store_details // table and store host_id and store_id mapping storeInsert.setLong(1, nfs_id); storeInsert.setString(2, nfs_uuid); storeInsert.setString(3, nfs_uuid); storeInsert.setString(4, nfs_url); storeInsert.setInt(5, nfs_dcid); storeInsert.setString(6, store_role); storeInsert.setString(7, nfs_parent); if (nfs_totalsize != null) { storeInsert.setLong(8, nfs_totalsize); } else { storeInsert.setNull(8, Types.BIGINT); } storeInsert.setDate(9, nfs_created); storeInsert.setDate(10, nfs_removed); storeInsert.executeUpdate(); } s_logger.debug("Marking NFS secondary storage in host table as removed"); pstmtUpdateHostAsRemoved.executeUpdate(); } catch (SQLException e) { String msg = "Unable to migrate secondary storages." + e.getMessage(); s_logger.error(msg); throw new CloudRuntimeException(msg, e); } s_logger.debug("Completed migrating secondary storage to image store"); } // migrate volume_host_ref to volume_store_ref private void migrateVolumeHostRef(Connection conn) { s_logger.debug("Updating volume_store_ref table from volume_host_ref table"); try(PreparedStatement volStoreInsert = conn.prepareStatement("INSERT INTO `cloud`.`volume_store_ref` (store_id, volume_id, zone_id, created, last_updated, job_id, download_pct, size, physical_size, download_state, checksum, error_str, local_path, install_path, url, destroyed, update_count, ref_cnt, state) select host_id, volume_id, zone_id, created, last_updated, job_id, download_pct, size, physical_size, download_state, checksum, error_str, local_path, install_path, url, destroyed, 0, 0, 'Allocated' from `cloud`.`volume_host_ref`");) { int rowCount = volStoreInsert.executeUpdate(); s_logger.debug("Insert modified " + rowCount + " rows"); try(PreparedStatement volStoreUpdate = conn.prepareStatement("update `cloud`.`volume_store_ref` set state = 'Ready' where download_state = 'DOWNLOADED'");) { rowCount = volStoreUpdate.executeUpdate(); s_logger.debug("Update modified " + rowCount + " rows"); }catch (SQLException e) { s_logger.error("Unable to migrate volume_host_ref." + e.getMessage(),e); throw new CloudRuntimeException("Unable to migrate volume_host_ref." + e.getMessage(),e); } } catch (SQLException e) { s_logger.error("Unable to migrate volume_host_ref." + e.getMessage(),e); throw new CloudRuntimeException("Unable to migrate volume_host_ref." + e.getMessage(),e); } s_logger.debug("Completed updating volume_store_ref table from volume_host_ref table"); } // migrate template_host_ref to template_store_ref private void migrateTemplateHostRef(Connection conn) { s_logger.debug("Updating template_store_ref table from template_host_ref table"); try (PreparedStatement tmplStoreInsert = conn.prepareStatement("INSERT INTO `cloud`.`template_store_ref` (store_id, template_id, created, last_updated, job_id, download_pct, size, physical_size, download_state, error_str, local_path, install_path, url, destroyed, is_copy, update_count, ref_cnt, store_role, state) select host_id, template_id, created, last_updated, job_id, download_pct, size, physical_size, download_state, error_str, local_path, install_path, url, destroyed, is_copy, 0, 0, 'Image', 'Allocated' from `cloud`.`template_host_ref`");) { int rowCount = tmplStoreInsert.executeUpdate(); s_logger.debug("Insert modified " + rowCount + " rows"); try(PreparedStatement tmplStoreUpdate = conn.prepareStatement("update `cloud`.`template_store_ref` set state = 'Ready' where download_state = 'DOWNLOADED'");) { rowCount = tmplStoreUpdate.executeUpdate(); }catch (SQLException e) { s_logger.error("Unable to migrate template_host_ref." + e.getMessage(),e); throw new CloudRuntimeException("Unable to migrate template_host_ref." + e.getMessage(), e); } s_logger.debug("Update modified " + rowCount + " rows"); } catch (SQLException e) { s_logger.error("Unable to migrate template_host_ref." + e.getMessage(),e); throw new CloudRuntimeException("Unable to migrate template_host_ref." + e.getMessage(), e); } s_logger.debug("Completed updating template_store_ref table from template_host_ref table"); } // migrate some entry contents of snapshots to snapshot_store_ref private void migrateSnapshotStoreRef(Connection conn) { s_logger.debug("Updating snapshot_store_ref table from snapshots table"); try(PreparedStatement snapshotStoreInsert = conn.prepareStatement("INSERT INTO `cloud`.`snapshot_store_ref` (store_id, snapshot_id, created, size, parent_snapshot_id, install_path, volume_id, update_count, ref_cnt, store_role, state) select sechost_id, id, created, size, prev_snap_id, CONCAT('snapshots', '/', account_id, '/', volume_id, '/', backup_snap_id), volume_id, 0, 0, 'Image', 'Ready' from `cloud`.`snapshots` where status = 'BackedUp' and hypervisor_type <> 'KVM' and sechost_id is not null and removed is null"); ) { //Update all snapshots except KVM snapshots int rowCount = snapshotStoreInsert.executeUpdate(); s_logger.debug("Inserted " + rowCount + " snapshots into snapshot_store_ref"); //backsnap_id for KVM snapshots is complate path. CONCAT is not required try(PreparedStatement snapshotStoreInsert_2 = conn.prepareStatement("INSERT INTO `cloud`.`snapshot_store_ref` (store_id, snapshot_id, created, size, parent_snapshot_id, install_path, volume_id, update_count, ref_cnt, store_role, state) select sechost_id, id, created, size, prev_snap_id, backup_snap_id, volume_id, 0, 0, 'Image', 'Ready' from `cloud`.`snapshots` where status = 'BackedUp' and hypervisor_type = 'KVM' and sechost_id is not null and removed is null");) { rowCount = snapshotStoreInsert_2.executeUpdate(); s_logger.debug("Inserted " + rowCount + " KVM snapshots into snapshot_store_ref"); }catch (SQLException e) { s_logger.error("Unable to migrate snapshot_store_ref." + e.getMessage(),e); throw new CloudRuntimeException("Unable to migrate snapshot_store_ref." + e.getMessage(),e); } } catch (SQLException e) { s_logger.error("Unable to migrate snapshot_store_ref." + e.getMessage(),e); throw new CloudRuntimeException("Unable to migrate snapshot_store_ref." + e.getMessage(),e); } s_logger.debug("Completed updating snapshot_store_ref table from snapshots table"); } // migrate secondary storages S3 from s3 tables to image_store table private void migrateS3ToImageStore(Connection conn) { Long storeId = null; Map<Long, Long> s3_store_id_map = new HashMap<Long, Long>(); s_logger.debug("Migrating S3 to image store"); try ( PreparedStatement storeQuery = conn.prepareStatement("select id from `cloud`.`image_store` where uuid = ?"); PreparedStatement storeDetailInsert = conn.prepareStatement("INSERT INTO `cloud`.`image_store_details` (store_id, name, value) values(?, ?, ?)"); // migrate S3 to image_store PreparedStatement storeInsert = conn.prepareStatement("INSERT INTO `cloud`.`image_store` (uuid, name, image_provider_name, protocol, scope, role, created) " + "values(?, ?, 'S3', ?, 'REGION', 'Image', ?)"); PreparedStatement s3Query = conn.prepareStatement("select id, uuid, access_key, secret_key, end_point, bucket, https, connection_timeout, " + "max_error_retry, socket_timeout, created from `cloud`.`s3`"); ResultSet rs = s3Query.executeQuery(); ) { while (rs.next()) { Long s3_id = rs.getLong("id"); String s3_uuid = rs.getString("uuid"); String s3_accesskey = rs.getString("access_key"); String s3_secretkey = rs.getString("secret_key"); String s3_endpoint = rs.getString("end_point"); String s3_bucket = rs.getString("bucket"); boolean s3_https = rs.getObject("https") != null ? (rs.getInt("https") == 0 ? false : true) : false; Integer s3_connectiontimeout = rs.getObject("connection_timeout") != null ? rs.getInt("connection_timeout") : null; Integer s3_retry = rs.getObject("max_error_retry") != null ? rs.getInt("max_error_retry") : null; Integer s3_sockettimeout = rs.getObject("socket_timeout") != null ? rs.getInt("socket_timeout") : null; Date s3_created = rs.getDate("created"); // insert entry in image_store table and image_store_details // table and store s3_id and store_id mapping storeInsert.setString(1, s3_uuid); storeInsert.setString(2, s3_uuid); storeInsert.setString(3, s3_https ? "https" : "http"); storeInsert.setDate(4, s3_created); storeInsert.executeUpdate(); storeQuery.setString(1, s3_uuid); try (ResultSet storeInfo = storeQuery.executeQuery();) { if (storeInfo.next()) { storeId = storeInfo.getLong("id"); } } Map<String, String> detailMap = new HashMap<String, String>(); detailMap.put(ApiConstants.S3_ACCESS_KEY, s3_accesskey); detailMap.put(ApiConstants.S3_SECRET_KEY, s3_secretkey); detailMap.put(ApiConstants.S3_BUCKET_NAME, s3_bucket); detailMap.put(ApiConstants.S3_END_POINT, s3_endpoint); detailMap.put(ApiConstants.S3_HTTPS_FLAG, String.valueOf(s3_https)); if (s3_connectiontimeout != null) { detailMap.put(ApiConstants.S3_CONNECTION_TIMEOUT, String.valueOf(s3_connectiontimeout)); } if (s3_retry != null) { detailMap.put(ApiConstants.S3_MAX_ERROR_RETRY, String.valueOf(s3_retry)); } if (s3_sockettimeout != null) { detailMap.put(ApiConstants.S3_SOCKET_TIMEOUT, String.valueOf(s3_sockettimeout)); } Iterator<String> keyIt = detailMap.keySet().iterator(); while (keyIt.hasNext()) { String key = keyIt.next(); String val = detailMap.get(key); storeDetailInsert.setLong(1, storeId); storeDetailInsert.setString(2, key); storeDetailInsert.setString(3, val); storeDetailInsert.executeUpdate(); } s3_store_id_map.put(s3_id, storeId); } } catch (SQLException e) { String msg = "Unable to migrate S3 secondary storages." + e.getMessage(); s_logger.error(msg); throw new CloudRuntimeException(msg, e); } s_logger.debug("Migrating template_s3_ref to template_store_ref"); migrateTemplateS3Ref(conn, s3_store_id_map); s_logger.debug("Migrating s3 backedup snapshots to snapshot_store_ref"); migrateSnapshotS3Ref(conn, s3_store_id_map); s_logger.debug("Completed migrating S3 secondary storage to image store"); } // migrate template_s3_ref to template_store_ref private void migrateTemplateS3Ref(Connection conn, Map<Long, Long> s3StoreMap) { s_logger.debug("Updating template_store_ref table from template_s3_ref table"); try(PreparedStatement tmplStoreInsert = conn.prepareStatement("INSERT INTO `cloud`.`template_store_ref` (store_id, template_id, created, download_pct, size, physical_size, download_state, local_path, install_path, update_count, ref_cnt, store_role, state) values(?, ?, ?, 100, ?, ?, 'DOWNLOADED', '?', '?', 0, 0, 'Image', 'Ready')"); ) { try(PreparedStatement s3Query = conn.prepareStatement("select template_s3_ref.s3_id, template_s3_ref.template_id, template_s3_ref.created, template_s3_ref.size, template_s3_ref.physical_size, vm_template.account_id from `cloud`.`template_s3_ref`, `cloud`.`vm_template` where vm_template.id = template_s3_ref.template_id");) { try(ResultSet rs = s3Query.executeQuery();) { while (rs.next()) { Long s3_id = rs.getLong("s3_id"); Long s3_tmpl_id = rs.getLong("template_id"); Date s3_created = rs.getDate("created"); Long s3_size = rs.getObject("size") != null ? rs.getLong("size") : null; Long s3_psize = rs.getObject("physical_size") != null ? rs.getLong("physical_size") : null; Long account_id = rs.getLong("account_id"); tmplStoreInsert.setLong(1, s3StoreMap.get(s3_id)); tmplStoreInsert.setLong(2, s3_tmpl_id); tmplStoreInsert.setDate(3, s3_created); if (s3_size != null) { tmplStoreInsert.setLong(4, s3_size); } else { tmplStoreInsert.setNull(4, Types.BIGINT); } if (s3_psize != null) { tmplStoreInsert.setLong(5, s3_psize); } else { tmplStoreInsert.setNull(5, Types.BIGINT); } String path = "template/tmpl/" + account_id + "/" + s3_tmpl_id; tmplStoreInsert.setString(6, path); tmplStoreInsert.setString(7, path); tmplStoreInsert.executeUpdate(); } }catch (SQLException e) { s_logger.error("Unable to migrate template_s3_ref." + e.getMessage(),e); throw new CloudRuntimeException("Unable to migrate template_s3_ref." + e.getMessage(),e); } }catch (SQLException e) { s_logger.error("Unable to migrate template_s3_ref." + e.getMessage(),e); throw new CloudRuntimeException("Unable to migrate template_s3_ref." + e.getMessage(),e); } } catch (SQLException e) { s_logger.error("Unable to migrate template_s3_ref." + e.getMessage(),e); throw new CloudRuntimeException("Unable to migrate template_s3_ref." + e.getMessage(),e); } s_logger.debug("Completed migrating template_s3_ref table."); } // migrate some entry contents of snapshots to snapshot_store_ref private void migrateSnapshotS3Ref(Connection conn, Map<Long, Long> s3StoreMap) { s_logger.debug("Updating snapshot_store_ref table from snapshots table for s3"); try(PreparedStatement snapshotStoreInsert = conn.prepareStatement("INSERT INTO `cloud`.`snapshot_store_ref` (store_id, snapshot_id, created, size, parent_snapshot_id, install_path, volume_id, update_count, ref_cnt, store_role, state) values(?, ?, ?, ?, ?, ?, ?, 0, 0, 'Image', 'Ready')"); ) { try(PreparedStatement s3Query = conn.prepareStatement("select s3_id, id, created, size, prev_snap_id, CONCAT('snapshots', '/', account_id, '/', volume_id, '/', backup_snap_id), volume_id, 0, 0, 'Image', 'Ready' from `cloud`.`snapshots` where status = 'BackedUp' and hypervisor_type <> 'KVM' and s3_id is not null and removed is null");) { try(ResultSet rs = s3Query.executeQuery();) { while (rs.next()) { Long s3_id = rs.getLong("s3_id"); Long snapshot_id = rs.getLong("id"); Date s3_created = rs.getDate("created"); Long s3_size = rs.getObject("size") != null ? rs.getLong("size") : null; Long s3_prev_id = rs.getObject("prev_snap_id") != null ? rs.getLong("prev_snap_id") : null; String install_path = rs.getString(6); Long s3_vol_id = rs.getLong("volume_id"); snapshotStoreInsert.setLong(1, s3StoreMap.get(s3_id)); snapshotStoreInsert.setLong(2, snapshot_id); snapshotStoreInsert.setDate(3, s3_created); if (s3_size != null) { snapshotStoreInsert.setLong(4, s3_size); } else { snapshotStoreInsert.setNull(4, Types.BIGINT); } if (s3_prev_id != null) { snapshotStoreInsert.setLong(5, s3_prev_id); } else { snapshotStoreInsert.setNull(5, Types.BIGINT); } snapshotStoreInsert.setString(6, install_path); snapshotStoreInsert.setLong(7, s3_vol_id); snapshotStoreInsert.executeUpdate(); } }catch (SQLException e) { s_logger.error("migrateSnapshotS3Ref:Exception:"+e.getMessage(),e); throw new CloudRuntimeException("migrateSnapshotS3Ref:Exception:"+e.getMessage(),e); } }catch (SQLException e) { s_logger.error("migrateSnapshotS3Ref:Exception:"+e.getMessage(),e); throw new CloudRuntimeException("migrateSnapshotS3Ref:Exception:"+e.getMessage(),e); } } catch (SQLException e) { s_logger.error("Unable to migrate s3 backedup snapshots to snapshot_store_ref." + e.getMessage()); throw new CloudRuntimeException("Unable to migrate s3 backedup snapshots to snapshot_store_ref." + e.getMessage(), e); } s_logger.debug("Completed updating snapshot_store_ref table from s3 snapshots entries"); } // migrate secondary storages Swift from swift tables to image_store table private void migrateSwiftToImageStore(Connection conn) { Long storeId = null; Map<Long, Long> swift_store_id_map = new HashMap<Long, Long>(); s_logger.debug("Migrating Swift to image store"); try ( PreparedStatement storeQuery = conn.prepareStatement("select id from `cloud`.`image_store` where uuid = ?"); PreparedStatement storeDetailInsert = conn.prepareStatement("INSERT INTO `cloud`.`image_store_details` (store_id, name, value) values(?, ?, ?)"); // migrate SWIFT secondary storage PreparedStatement storeInsert = conn.prepareStatement("INSERT INTO `cloud`.`image_store` (uuid, name, image_provider_name, protocol, url, scope, role, created) values(?, ?, 'Swift', 'http', ?, 'REGION', 'Image', ?)"); PreparedStatement swiftQuery = conn.prepareStatement("select id, uuid, url, account, username, swift.key, created from `cloud`.`swift`"); ResultSet rs = swiftQuery.executeQuery(); ) { while (rs.next()) { Long swift_id = rs.getLong("id"); String swift_uuid = rs.getString("uuid"); String swift_url = rs.getString("url"); String swift_account = rs.getString("account"); String swift_username = rs.getString("username"); String swift_key = rs.getString("key"); Date swift_created = rs.getDate("created"); // insert entry in image_store table and image_store_details // table and store swift_id and store_id mapping storeInsert.setString(1, swift_uuid); storeInsert.setString(2, swift_uuid); storeInsert.setString(3, swift_url); storeInsert.setDate(4, swift_created); storeInsert.executeUpdate(); storeQuery.setString(1, swift_uuid); try (ResultSet storeInfo = storeQuery.executeQuery();) { if (storeInfo.next()) { storeId = storeInfo.getLong("id"); } } Map<String, String> detailMap = new HashMap<String, String>(); detailMap.put(ApiConstants.ACCOUNT, swift_account); detailMap.put(ApiConstants.USERNAME, swift_username); detailMap.put(ApiConstants.KEY, swift_key); Iterator<String> keyIt = detailMap.keySet().iterator(); while (keyIt.hasNext()) { String key = keyIt.next(); String val = detailMap.get(key); storeDetailInsert.setLong(1, storeId); storeDetailInsert.setString(2, key); storeDetailInsert.setString(3, val); storeDetailInsert.executeUpdate(); } swift_store_id_map.put(swift_id, storeId); } } catch (SQLException e) { String msg = "Unable to migrate swift secondary storages." + e.getMessage(); s_logger.error(msg); throw new CloudRuntimeException(msg, e); } s_logger.debug("Migrating template_swift_ref to template_store_ref"); migrateTemplateSwiftRef(conn, swift_store_id_map); s_logger.debug("Migrating swift backedup snapshots to snapshot_store_ref"); migrateSnapshotSwiftRef(conn, swift_store_id_map); s_logger.debug("Completed migrating Swift secondary storage to image store"); } // migrate template_s3_ref to template_store_ref private void migrateTemplateSwiftRef(Connection conn, Map<Long, Long> swiftStoreMap) { s_logger.debug("Updating template_store_ref table from template_swift_ref table"); try ( PreparedStatement tmplStoreInsert = conn.prepareStatement("INSERT INTO `cloud`.`template_store_ref` (store_id, template_id, created, download_pct, size, physical_size, download_state, local_path, install_path, update_count, ref_cnt, store_role, state) values(?, ?, ?, 100, ?, ?, 'DOWNLOADED', '?', '?', 0, 0, 'Image', 'Ready')"); PreparedStatement s3Query = conn.prepareStatement("select swift_id, template_id, created, path, size, physical_size from `cloud`.`template_swift_ref`"); ResultSet rs = s3Query.executeQuery(); ) { while (rs.next()) { Long swift_id = rs.getLong("swift_id"); Long tmpl_id = rs.getLong("template_id"); Date created = rs.getDate("created"); String path = rs.getString("path"); Long size = rs.getObject("size") != null ? rs.getLong("size") : null; Long psize = rs.getObject("physical_size") != null ? rs.getLong("physical_size") : null; tmplStoreInsert.setLong(1, swiftStoreMap.get(swift_id)); tmplStoreInsert.setLong(2, tmpl_id); tmplStoreInsert.setDate(3, created); if (size != null) { tmplStoreInsert.setLong(4, size); } else { tmplStoreInsert.setNull(4, Types.BIGINT); } if (psize != null) { tmplStoreInsert.setLong(5, psize); } else { tmplStoreInsert.setNull(5, Types.BIGINT); } tmplStoreInsert.setString(6, path); tmplStoreInsert.setString(7, path); tmplStoreInsert.executeUpdate(); } } catch (SQLException e) { String msg = "Unable to migrate template_swift_ref." + e.getMessage(); s_logger.error(msg); throw new CloudRuntimeException(msg, e); } s_logger.debug("Completed migrating template_swift_ref table."); } // migrate some entry contents of snapshots to snapshot_store_ref private void migrateSnapshotSwiftRef(Connection conn, Map<Long, Long> swiftStoreMap) { s_logger.debug("Updating snapshot_store_ref table from snapshots table for swift"); try (PreparedStatement snapshotStoreInsert = conn.prepareStatement("INSERT INTO `cloud`.`snapshot_store_ref` (store_id, snapshot_id, created, size, parent_snapshot_id, install_path, volume_id, update_count, ref_cnt, store_role, state) values(?, ?, ?, ?, ?, ?, ?, 0, 0, 'Image', 'Ready')"); ){ try(PreparedStatement s3Query = conn.prepareStatement("select swift_id, id, created, size, prev_snap_id, CONCAT('snapshots', '/', account_id, '/', volume_id, '/', backup_snap_id), volume_id, 0, 0, 'Image', 'Ready' from `cloud`.`snapshots` where status = 'BackedUp' and hypervisor_type <> 'KVM' and swift_id is not null and removed is null");) { try(ResultSet rs = s3Query.executeQuery();) { while (rs.next()) { Long swift_id = rs.getLong("swift_id"); Long snapshot_id = rs.getLong("id"); Date created = rs.getDate("created"); Long size = rs.getLong("size"); Long prev_id = rs.getLong("prev_snap_id"); String install_path = rs.getString(6); Long vol_id = rs.getLong("volume_id"); snapshotStoreInsert.setLong(1, swiftStoreMap.get(swift_id)); snapshotStoreInsert.setLong(2, snapshot_id); snapshotStoreInsert.setDate(3, created); snapshotStoreInsert.setLong(4, size); snapshotStoreInsert.setLong(5, prev_id); snapshotStoreInsert.setString(6, install_path); snapshotStoreInsert.setLong(7, vol_id); snapshotStoreInsert.executeUpdate(); } }catch (SQLException e) { s_logger.error("migrateSnapshotSwiftRef:Exception:"+e.getMessage(),e); throw new CloudRuntimeException("migrateSnapshotSwiftRef:Exception:"+e.getMessage(),e); } }catch (SQLException e) { s_logger.error("migrateSnapshotSwiftRef:Exception:"+e.getMessage(),e); throw new CloudRuntimeException("migrateSnapshotSwiftRef:Exception:"+e.getMessage(),e); } } catch (SQLException e) { s_logger.error("migrateSnapshotSwiftRef:Exception:"+e.getMessage(),e); throw new CloudRuntimeException("migrateSnapshotSwiftRef:Exception:"+e.getMessage(),e); } s_logger.debug("Completed updating snapshot_store_ref table from swift snapshots entries"); } private void fixNiciraKeys(Connection conn) { //First drop the key if it exists. List<String> keys = new ArrayList<String>(); s_logger.debug("Dropping foreign key fk_nicira_nvp_nic_map__nic from the table nicira_nvp_nic_map if it exists"); keys.add("fk_nicira_nvp_nic_map__nic"); DbUpgradeUtils.dropKeysIfExist(conn, "nicira_nvp_nic_map", keys, true); //Now add foreign key. try(PreparedStatement pstmt = conn.prepareStatement("ALTER TABLE `cloud`.`nicira_nvp_nic_map` ADD CONSTRAINT `fk_nicira_nvp_nic_map__nic` FOREIGN KEY (`nic`) REFERENCES `nics` (`uuid`) ON DELETE CASCADE");) { pstmt.executeUpdate(); s_logger.debug("Added foreign key fk_nicira_nvp_nic_map__nic to the table nicira_nvp_nic_map"); } catch (SQLException e) { throw new CloudRuntimeException("Unable to add foreign key fk_nicira_nvp_nic_map__nic to the table nicira_nvp_nic_map", e); } } private void fixRouterKeys(Connection conn) { //First drop the key if it exists. List<String> keys = new ArrayList<String>(); s_logger.debug("Dropping foreign key fk_router_network_ref__router_id from the table router_network_ref if it exists"); keys.add("fk_router_network_ref__router_id"); DbUpgradeUtils.dropKeysIfExist(conn, "router_network_ref", keys, true); //Now add foreign key. try (PreparedStatement pstmt = conn.prepareStatement("ALTER TABLE `cloud`.`router_network_ref` ADD CONSTRAINT `fk_router_network_ref__router_id` FOREIGN KEY (`router_id`) REFERENCES `domain_router` (`id`) ON DELETE CASCADE");) { pstmt.executeUpdate(); s_logger.debug("Added foreign key fk_router_network_ref__router_id to the table router_network_ref"); } catch (SQLException e) { throw new CloudRuntimeException("Unable to add foreign key fk_router_network_ref__router_id to the table router_network_ref", e); } } private void encryptSite2SitePSK(Connection conn) { s_logger.debug("Encrypting Site2Site Customer Gateway pre-shared key"); try (PreparedStatement select_pstmt = conn.prepareStatement("select id, ipsec_psk from `cloud`.`s2s_customer_gateway`");){ try(ResultSet rs = select_pstmt.executeQuery();) { while (rs.next()) { long id = rs.getLong(1); String value = rs.getString(2); if (value == null) { continue; } String encryptedValue = DBEncryptionUtil.encrypt(value); try(PreparedStatement update_pstmt = conn.prepareStatement("update `cloud`.`s2s_customer_gateway` set ipsec_psk=? where id=?");) { update_pstmt.setBytes(1, encryptedValue.getBytes("UTF-8")); update_pstmt.setLong(2, id); update_pstmt.executeUpdate(); }catch (SQLException e) { throw new CloudRuntimeException("encryptSite2SitePSK:Exception:"+e.getMessage(), e); } } }catch (SQLException e) { throw new CloudRuntimeException("encryptSite2SitePSK:Exception:"+e.getMessage(), e); } } catch (SQLException e) { throw new CloudRuntimeException("Unable to encrypt Site2Site Customer Gateway pre-shared key ", e); } catch (UnsupportedEncodingException e) { throw new CloudRuntimeException("Unable to encrypt Site2Site Customer Gateway pre-shared key ", e); } s_logger.debug("Done encrypting Site2Site Customer Gateway pre-shared key"); } protected void updateConcurrentConnectionsInNetworkOfferings(Connection conn) { try { try (PreparedStatement sel_pstmt = conn.prepareStatement("SELECT * FROM information_schema.COLUMNS WHERE TABLE_SCHEMA = 'cloud' AND TABLE_NAME = 'network_offerings' AND COLUMN_NAME = 'concurrent_connections'");) { try(ResultSet rs = sel_pstmt.executeQuery();) { if (!rs.next()) { try(PreparedStatement alter_pstmt = conn.prepareStatement("ALTER TABLE `cloud`.`network_offerings` ADD COLUMN `concurrent_connections` int(10) unsigned COMMENT 'Load Balancer(haproxy) maximum number of concurrent connections(global max)'");) { alter_pstmt.executeUpdate(); }catch (SQLException e) { throw new CloudRuntimeException("migration of concurrent connections from network_details failed"); } } }catch (SQLException e) { throw new CloudRuntimeException("migration of concurrent connections from network_details failed"); } } catch (SQLException e) { throw new CloudRuntimeException("migration of concurrent connections from network_details failed"); } try(PreparedStatement sel_net_pstmt = conn.prepareStatement("select network_id, value from `cloud`.`network_details` where name='maxconnections'");) { try(ResultSet rs = sel_net_pstmt.executeQuery();) { while (rs.next()) { long networkId = rs.getLong(1); int maxconnections = Integer.parseInt(rs.getString(2)); try(PreparedStatement sel_net_off_pstmt = conn.prepareStatement("select network_offering_id from `cloud`.`networks` where id= ?");) { sel_net_off_pstmt.setLong(1, networkId); try(ResultSet rs1 = sel_net_off_pstmt.executeQuery();) { if (rs1.next()) { long network_offering_id = rs1.getLong(1); try(PreparedStatement pstmt = conn.prepareStatement("select concurrent_connections from `cloud`.`network_offerings` where id= ?");) { pstmt.setLong(1, network_offering_id); try(ResultSet rs2 = pstmt.executeQuery();) { if ((!rs2.next()) || (rs2.getInt(1) < maxconnections)) { try(PreparedStatement update_net_pstmt = conn.prepareStatement("update network_offerings set concurrent_connections=? where id=?");) { update_net_pstmt.setInt(1, maxconnections); update_net_pstmt.setLong(2, network_offering_id); update_net_pstmt.executeUpdate(); }catch (SQLException e) { throw new CloudRuntimeException("migration of concurrent connections from network_details failed"); } } }catch (SQLException e) { throw new CloudRuntimeException("migration of concurrent connections from network_details failed"); } }catch (SQLException e) { throw new CloudRuntimeException("migration of concurrent connections from network_details failed"); } } }catch (SQLException e) { throw new CloudRuntimeException("migration of concurrent connections from network_details failed"); } }catch (SQLException e) { throw new CloudRuntimeException("migration of concurrent connections from network_details failed"); } } }catch (SQLException e) { throw new CloudRuntimeException("migration of concurrent connections from network_details failed"); } }catch (SQLException e) { throw new CloudRuntimeException("migration of concurrent connections from network_details failed"); } } catch (RuntimeException e) { throw new CloudRuntimeException("migration of concurrent connections from network_details failed",e); } } private void migrateDatafromIsoIdInVolumesTable(Connection conn) { try(PreparedStatement pstmt = conn.prepareStatement("SELECT iso_id1 From `cloud`.`volumes`");) { try(ResultSet rs = pstmt.executeQuery();) { if (rs.next()) { try(PreparedStatement alter_pstmt = conn.prepareStatement("ALTER TABLE `cloud`.`volumes` DROP COLUMN `iso_id`");) { alter_pstmt.executeUpdate(); try(PreparedStatement alter_iso_pstmt = conn.prepareStatement("ALTER TABLE `cloud`.`volumes` CHANGE COLUMN `iso_id1` `iso_id` bigint(20) unsigned COMMENT 'The id of the iso from which the volume was created'");) { alter_iso_pstmt.executeUpdate(); }catch (SQLException e) { s_logger.error("migrateDatafromIsoIdInVolumesTable:Exception:"+e.getMessage(),e); //implies iso_id1 is not present, so do nothing. } }catch (SQLException e) { s_logger.error("migrateDatafromIsoIdInVolumesTable:Exception:"+e.getMessage(),e); //implies iso_id1 is not present, so do nothing. } } }catch (SQLException e) { s_logger.error("migrateDatafromIsoIdInVolumesTable:Exception:"+e.getMessage(),e); //implies iso_id1 is not present, so do nothing. } } catch (SQLException e) { //implies iso_id1 is not present, so do nothing. } } protected void setRAWformatForRBDVolumes(Connection conn) { try(PreparedStatement pstmt = conn.prepareStatement("UPDATE volumes SET format = 'RAW' WHERE pool_id IN(SELECT id FROM storage_pool WHERE pool_type = 'RBD')");) { s_logger.debug("Setting format to RAW for all volumes on RBD primary storage pools"); pstmt.executeUpdate(); } catch (SQLException e) { throw new CloudRuntimeException("Failed to update volume format to RAW for volumes on RBD pools due to exception ", e); } } private void upgradeVpcServiceMap(Connection conn) { s_logger.debug("Upgrading VPC service Map"); try(PreparedStatement listVpc = conn.prepareStatement("SELECT id, vpc_offering_id FROM `cloud`.`vpc` where removed is NULL");) { //Get all vpc Ids along with vpc offering Id try(ResultSet rs = listVpc.executeQuery();) { while (rs.next()) { long vpc_id = rs.getLong(1); long offering_id = rs.getLong(2); //list all services and providers in offering try(PreparedStatement listServiceProviders = conn.prepareStatement("SELECT service, provider FROM `cloud`.`vpc_offering_service_map` where vpc_offering_id = ?");) { listServiceProviders.setLong(1, offering_id); try(ResultSet rs1 = listServiceProviders.executeQuery();) { //Insert entries in vpc_service_map while (rs1.next()) { String service = rs1.getString(1); String provider = rs1.getString(2); try (PreparedStatement insertProviders = conn.prepareStatement("INSERT INTO `cloud`.`vpc_service_map` (`vpc_id`, `service`, `provider`, `created`) VALUES (?, ?, ?, now());");) { insertProviders.setLong(1, vpc_id); insertProviders.setString(2, service); insertProviders.setString(3, provider); insertProviders.executeUpdate(); } catch (SQLException e) { throw new CloudRuntimeException("Error during VPC service map upgrade", e); } } }catch (SQLException e) { throw new CloudRuntimeException("Error during VPC service map upgrade", e); } }catch (SQLException e) { throw new CloudRuntimeException("Error during VPC service map upgrade", e); } s_logger.debug("Upgraded service map for VPC: " + vpc_id); } } } catch (SQLException e) { throw new CloudRuntimeException("Error during VPC service map upgrade", e); } } private void upgradeResourceCount(Connection conn) { s_logger.debug("upgradeResourceCount start"); try( PreparedStatement sel_dom_pstmt = conn.prepareStatement("select id, domain_id FROM `cloud`.`account` where removed is NULL "); ResultSet rsAccount = sel_dom_pstmt.executeQuery(); ) { while (rsAccount.next()) { long account_id = rsAccount.getLong(1); long domain_id = rsAccount.getLong(2); // 1. update cpu,memory for all accounts try(PreparedStatement sel_sum_pstmt = conn.prepareStatement("SELECT SUM(service_offering.cpu), SUM(service_offering.ram_size)" + " FROM `cloud`.`vm_instance`, `cloud`.`service_offering`" + " WHERE vm_instance.service_offering_id = service_offering.id AND vm_instance.account_id = ?" + " AND vm_instance.removed is NULL" + " AND vm_instance.vm_type='User' AND state not in ('Destroyed', 'Error', 'Expunging')");) { sel_sum_pstmt.setLong(1, account_id); try(ResultSet sel_sum_pstmt_res = sel_sum_pstmt.executeQuery();) { if (sel_sum_pstmt_res.next()) { upgradeResourceCountforAccount(conn, account_id, domain_id, "cpu", sel_sum_pstmt_res.getLong(1)); upgradeResourceCountforAccount(conn, account_id, domain_id, "memory", sel_sum_pstmt_res.getLong(2)); } else { upgradeResourceCountforAccount(conn, account_id, domain_id, "cpu", 0L); upgradeResourceCountforAccount(conn, account_id, domain_id, "memory", 0L); } // 2. update primary_storage for all accounts try(PreparedStatement sel_cloud_vol_pstmt = conn.prepareStatement("SELECT sum(size) FROM `cloud`.`volumes` WHERE account_id= ?" + " AND (path is not NULL OR state in ('Allocated')) AND removed is NULL" + " AND instance_id IN (SELECT id FROM `cloud`.`vm_instance` WHERE vm_type='User')");) { sel_cloud_vol_pstmt.setLong(1, account_id); try(ResultSet sel_cloud_vol_count = sel_cloud_vol_pstmt.executeQuery();) { if (sel_cloud_vol_count.next()) { upgradeResourceCountforAccount(conn, account_id, domain_id, "primary_storage", sel_cloud_vol_count.getLong(1)); } else { upgradeResourceCountforAccount(conn, account_id, domain_id, "primary_storage", 0L); } }catch (SQLException e) { throw new CloudRuntimeException("upgradeResourceCount:Exception:"+e.getMessage(),e); } }catch (SQLException e) { throw new CloudRuntimeException("upgradeResourceCount:Exception:"+e.getMessage(),e); } // 3. update secondary_storage for all accounts long totalVolumesSize = 0; long totalSnapshotsSize = 0; long totalTemplatesSize = 0; try(PreparedStatement sel_cloud_vol_alloc_pstmt = conn.prepareStatement("SELECT sum(size) FROM `cloud`.`volumes` WHERE account_id= ?" + " AND path is NULL AND state not in ('Allocated') AND removed is NULL");) { sel_cloud_vol_alloc_pstmt.setLong(1, account_id); try(ResultSet sel_cloud_vol_res = sel_cloud_vol_alloc_pstmt.executeQuery();) { if (sel_cloud_vol_res.next()) { totalVolumesSize = sel_cloud_vol_res.getLong(1); } try(PreparedStatement sel_cloud_snapshot_pstmt = conn.prepareStatement("SELECT sum(size) FROM `cloud`.`snapshots` WHERE account_id= ? AND removed is NULL");) { sel_cloud_snapshot_pstmt.setLong(1, account_id); try(ResultSet sel_cloud_snapshot_res = sel_cloud_snapshot_pstmt.executeQuery();) { if (sel_cloud_snapshot_res.next()) { totalSnapshotsSize = sel_cloud_snapshot_res.getLong(1); } try (PreparedStatement sel_templ_store_pstmt = conn.prepareStatement("SELECT sum(template_store_ref.size) FROM `cloud`.`template_store_ref`,`cloud`.`vm_template` WHERE account_id = ?" + " AND template_store_ref.template_id = vm_template.id AND download_state = 'DOWNLOADED' AND destroyed = false AND removed is NULL");) { sel_templ_store_pstmt.setLong(1, account_id); try (ResultSet templ_store_count = sel_templ_store_pstmt.executeQuery();) { if (templ_store_count.next()) { totalTemplatesSize = templ_store_count.getLong(1); } } catch (SQLException e) { throw new CloudRuntimeException("upgradeResourceCount:Exception:" + e.getMessage(), e); } } catch (SQLException e) { throw new CloudRuntimeException("upgradeResourceCount:Exception:" + e.getMessage(), e); } upgradeResourceCountforAccount(conn, account_id, domain_id, "secondary_storage", totalVolumesSize + totalSnapshotsSize + totalTemplatesSize); }catch (SQLException e) { throw new CloudRuntimeException("upgradeResourceCount:Exception:" + e.getMessage(), e); } }catch (SQLException e) { throw new CloudRuntimeException("upgradeResourceCount:Exception:" + e.getMessage(), e); } }catch (SQLException e) { throw new CloudRuntimeException("upgradeResourceCount:Exception:" + e.getMessage(), e); } }catch (SQLException e) { throw new CloudRuntimeException("upgradeResourceCount:Exception:"+e.getMessage(),e); } }catch (SQLException e) { throw new CloudRuntimeException("upgradeResourceCount:Exception:"+e.getMessage(),e); } }catch (SQLException e) { throw new CloudRuntimeException("upgradeResourceCount:Exception:"+e.getMessage(),e); } } // 4. upgrade cpu,memory,primary_storage,secondary_storage for domains String resource_types[] = {"cpu", "memory", "primary_storage", "secondary_storage"}; try(PreparedStatement sel_id_pstmt = conn.prepareStatement("select id FROM `cloud`.`domain`");) { try(ResultSet sel_id_res = sel_id_pstmt.executeQuery();) { while (sel_id_res.next()) { long domain_id = sel_id_res.getLong(1); for (int count = 0; count < resource_types.length; count++) { String resource_type = resource_types[count]; upgradeResourceCountforDomain(conn, domain_id, resource_type, 0L); // reset value to 0 before statistics } } }catch (SQLException e) { throw new CloudRuntimeException("Unable to upgrade resource count (cpu,memory,primary_storage,secondary_storage) ", e); } }catch (SQLException e) { throw new CloudRuntimeException("Unable to upgrade resource count (cpu,memory,primary_storage,secondary_storage) ", e); } for (int count = 0; count < resource_types.length; count++) { String resource_type = resource_types[count]; try(PreparedStatement sel_dom_id_pstmt = conn.prepareStatement("select account.domain_id,sum(resource_count.count) from `cloud`.`account` left join `cloud`.`resource_count` on account.id=resource_count.account_id " + "where resource_count.type=? group by account.domain_id;");) { sel_dom_id_pstmt.setString(1, resource_type); try(ResultSet sel_dom_res = sel_dom_id_pstmt.executeQuery();) { while (sel_dom_res.next()) { long domain_id = sel_dom_res.getLong(1); long resource_count = sel_dom_res.getLong(2); upgradeResourceCountforDomain(conn, domain_id, resource_type, resource_count); } }catch (SQLException e) { throw new CloudRuntimeException("Unable to upgrade resource count (cpu,memory,primary_storage,secondary_storage) ", e); } }catch (SQLException e) { throw new CloudRuntimeException("Unable to upgrade resource count (cpu,memory,primary_storage,secondary_storage) ", e); } } s_logger.debug("upgradeResourceCount finish"); } catch (SQLException e) { throw new CloudRuntimeException("Unable to upgrade resource count (cpu,memory,primary_storage,secondary_storage) ", e); } } private static void upgradeResourceCountforAccount(Connection conn, Long accountId, Long domainId, String type, Long resourceCount) throws SQLException { //update or insert into resource_count table. try(PreparedStatement pstmt = conn.prepareStatement("INSERT INTO `cloud`.`resource_count` (account_id, type, count) VALUES (?,?,?) ON DUPLICATE KEY UPDATE id=LAST_INSERT_ID(id), count=?");) { pstmt.setLong(1, accountId); pstmt.setString(2, type); pstmt.setLong(3, resourceCount); pstmt.setLong(4, resourceCount); pstmt.executeUpdate(); }catch (SQLException e) { throw new CloudRuntimeException("upgradeResourceCountforAccount:Exception:"+e.getMessage(),e); } } private static void upgradeResourceCountforDomain(Connection conn, Long domainId, String type, Long resourceCount) throws SQLException { //update or insert into resource_count table. try(PreparedStatement pstmt = conn.prepareStatement("INSERT INTO `cloud`.`resource_count` (domain_id, type, count) VALUES (?,?,?) ON DUPLICATE KEY UPDATE id=LAST_INSERT_ID(id), count=?");) { pstmt.setLong(1, domainId); pstmt.setString(2, type); pstmt.setLong(3, resourceCount); pstmt.setLong(4, resourceCount); pstmt.executeUpdate(); }catch (SQLException e) { throw new CloudRuntimeException("upgradeResourceCountforDomain:Exception:"+e.getMessage(),e); } } }