/* * Copyright (c) 2010-2013, 2015, 2016 Eike Stepper (Berlin, Germany) and others. * All rights reserved. This program and the accompanying materials * are made available under the terms of the Eclipse Public License v1.0 * which accompanies this distribution, and is available at * http://www.eclipse.org/legal/epl-v10.html * * Contributors: * Stefan Winkler - initial API and implementation taken from AuditFeatureMapTableMappingWithRanges * Stefan Winkler - Bug 329025: [DB] Support branching for range-based mapping strategy */ package org.eclipse.emf.cdo.server.internal.db.mapping.horizontal; import org.eclipse.emf.cdo.common.branch.CDOBranch; import org.eclipse.emf.cdo.common.branch.CDOBranchManager; import org.eclipse.emf.cdo.common.branch.CDOBranchPoint; import org.eclipse.emf.cdo.common.id.CDOID; import org.eclipse.emf.cdo.common.id.CDOIDUtil; import org.eclipse.emf.cdo.common.revision.CDOList; import org.eclipse.emf.cdo.common.revision.CDORevision; import org.eclipse.emf.cdo.common.revision.CDORevisionUtil; import org.eclipse.emf.cdo.common.revision.delta.CDOAddFeatureDelta; import org.eclipse.emf.cdo.common.revision.delta.CDOClearFeatureDelta; import org.eclipse.emf.cdo.common.revision.delta.CDOContainerFeatureDelta; import org.eclipse.emf.cdo.common.revision.delta.CDOFeatureDelta; import org.eclipse.emf.cdo.common.revision.delta.CDOFeatureDeltaVisitor; import org.eclipse.emf.cdo.common.revision.delta.CDOListFeatureDelta; import org.eclipse.emf.cdo.common.revision.delta.CDOMoveFeatureDelta; import org.eclipse.emf.cdo.common.revision.delta.CDORemoveFeatureDelta; import org.eclipse.emf.cdo.common.revision.delta.CDOSetFeatureDelta; import org.eclipse.emf.cdo.common.revision.delta.CDOUnsetFeatureDelta; import org.eclipse.emf.cdo.server.IStoreAccessor.QueryXRefsContext; import org.eclipse.emf.cdo.server.IStoreChunkReader; import org.eclipse.emf.cdo.server.IStoreChunkReader.Chunk; import org.eclipse.emf.cdo.server.db.IDBStore; import org.eclipse.emf.cdo.server.db.IDBStoreAccessor; import org.eclipse.emf.cdo.server.db.IDBStoreChunkReader; import org.eclipse.emf.cdo.server.db.IIDHandler; import org.eclipse.emf.cdo.server.db.mapping.IListMappingDeltaSupport; import org.eclipse.emf.cdo.server.db.mapping.IMappingStrategy; import org.eclipse.emf.cdo.server.db.mapping.ITypeMapping; import org.eclipse.emf.cdo.server.internal.db.bundle.OM; import org.eclipse.emf.cdo.spi.common.revision.InternalCDORevision; import org.eclipse.emf.cdo.spi.common.revision.InternalCDORevisionManager; import org.eclipse.emf.cdo.spi.server.InternalRepository; import org.eclipse.net4j.db.DBException; import org.eclipse.net4j.db.DBType; import org.eclipse.net4j.db.DBUtil; import org.eclipse.net4j.db.IDBDatabase; import org.eclipse.net4j.db.IDBPreparedStatement; import org.eclipse.net4j.db.IDBPreparedStatement.ReuseProbability; import org.eclipse.net4j.db.ddl.IDBIndex.Type; import org.eclipse.net4j.db.ddl.IDBTable; import org.eclipse.net4j.util.ImplementationError; import org.eclipse.net4j.util.collection.MoveableList; import org.eclipse.net4j.util.om.trace.ContextTracer; import org.eclipse.emf.ecore.EClass; import org.eclipse.emf.ecore.EStructuralFeature; import org.eclipse.emf.ecore.util.FeatureMap; import java.sql.ResultSet; import java.sql.SQLException; import java.util.ArrayList; import java.util.Collection; import java.util.Collections; import java.util.Iterator; import java.util.List; import java.util.Map; /** * This is a featuremap-table mapping for audit mode. It is optimized for frequent insert operations at the list's end, * which causes just 1 DB row to be changed. This is achieved by introducing a version range (columns * {@link IMappingConstants#LIST_REVISION_VERSION_ADDED cdo_version_added} and * {@link IMappingConstants#LIST_REVISION_VERSION_REMOVED cdo_version_removed}) which records for which revisions a particular * entry existed. Also, this mapping is mainly optimized for potentially very large lists: the need for having the * complete list stored in memory to do in-the-middle-moved and inserts is traded in for a few more DB access * operations. * * @author Eike Stepper * @author Stefan Winkler * @author Lothar Werzinger * @since 3.0 */ public class BranchingFeatureMapTableMappingWithRanges extends AbstractBasicListTableMapping implements IListMappingDeltaSupport { private static final ContextTracer TRACER = new ContextTracer(OM.DEBUG, BranchingFeatureMapTableMappingWithRanges.class); /** * Used to clean up lists for detached objects. */ private static final int FINAL_VERSION = Integer.MAX_VALUE; /** * The table of this mapping. */ private IDBTable table; /** * The tags mapped to column names */ private Map<CDOID, String> tagMap; /** * Column name Set */ private List<String> columnNames; /** * The type mappings for the value fields. */ private Map<CDOID, ITypeMapping> typeMappings; private List<DBType> dbTypes; // --------- SQL strings - see initSQLStrings() ----------------- private String sqlSelectChunksPrefix; private String sqlOrderByIndex; private String sqlInsert; private String sqlRemoveEntry; private String sqlDeleteEntry; private String sqlUpdateIndex; private String sqlGetValue; private String sqlClearList; public BranchingFeatureMapTableMappingWithRanges(IMappingStrategy mappingStrategy, EClass eClass, EStructuralFeature feature) { super(mappingStrategy, eClass, feature); initDBTypes(); initTable(); initSQLStrings(); } private void initDBTypes() { // TODO add annotation processing here ... ITypeMapping.Registry registry = ITypeMapping.Registry.INSTANCE; dbTypes = new ArrayList<DBType>(registry.getDefaultFeatureMapDBTypes()); } private void initTable() { String tableName = getMappingStrategy().getTableName(getContainingClass(), getFeature()); IDBStore store = getMappingStrategy().getStore(); DBType idType = store.getIDHandler().getDBType(); int idLength = store.getIDColumnLength(); IDBDatabase database = getMappingStrategy().getStore().getDatabase(); table = database.getSchema().getTable(tableName); if (table == null) { table = database.getSchemaTransaction().getWorkingCopy().addTable(tableName); table.addField(FEATUREMAP_REVISION_ID, idType, idLength); table.addField(LIST_REVISION_BRANCH, DBType.INTEGER); table.addField(FEATUREMAP_VERSION_ADDED, DBType.INTEGER); table.addField(FEATUREMAP_VERSION_REMOVED, DBType.INTEGER); table.addField(FEATUREMAP_IDX, DBType.INTEGER); table.addField(FEATUREMAP_TAG, idType, idLength); tagMap = CDOIDUtil.createMap(); typeMappings = CDOIDUtil.createMap(); columnNames = new ArrayList<String>(); initTypeColumns(true); table.addIndex(Type.NON_UNIQUE, FEATUREMAP_REVISION_ID); table.addIndex(Type.NON_UNIQUE, LIST_REVISION_BRANCH); table.addIndex(Type.NON_UNIQUE, FEATUREMAP_VERSION_ADDED); table.addIndex(Type.NON_UNIQUE, FEATUREMAP_VERSION_REMOVED); table.addIndex(Type.NON_UNIQUE, FEATUREMAP_IDX); table.addIndex(Type.NON_UNIQUE, FEATUREMAP_TAG); } else { initTypeColumns(false); } } private void initTypeColumns(boolean create) { for (DBType type : getDBTypes()) { String column = FEATUREMAP_VALUE + "_" + type.name(); if (create) { table.addField(column, type); } columnNames.add(column); } } private void initSQLStrings() { String tableName = getTable().getName(); // ---------------- SELECT to read chunks ---------------------------- StringBuilder builder = new StringBuilder(); builder.append("SELECT "); //$NON-NLS-1$ builder.append(FEATUREMAP_IDX); builder.append(", "); //$NON-NLS-1$ builder.append(FEATUREMAP_TAG); builder.append(", "); //$NON-NLS-1$ Iterator<String> iter = columnNames.iterator(); while (iter.hasNext()) { builder.append(iter.next()); if (iter.hasNext()) { builder.append(", "); //$NON-NLS-1$ } } builder.append(" FROM "); //$NON-NLS-1$ builder.append(tableName); builder.append(" WHERE "); //$NON-NLS-1$ builder.append(FEATUREMAP_REVISION_ID); builder.append("=? AND "); //$NON-NLS-1$ builder.append(FEATUREMAP_BRANCH); builder.append("=? AND "); //$NON-NLS-1$ builder.append(FEATUREMAP_VERSION_ADDED); builder.append("<=? AND ("); //$NON-NLS-1$ builder.append(FEATUREMAP_VERSION_REMOVED); builder.append(" IS NULL OR "); //$NON-NLS-1$ builder.append(FEATUREMAP_VERSION_REMOVED); builder.append(">?)"); //$NON-NLS-1$ sqlSelectChunksPrefix = builder.toString(); sqlOrderByIndex = " ORDER BY " + FEATUREMAP_IDX; //$NON-NLS-1$ // ----------------- INSERT - prefix ----------------- builder = new StringBuilder("INSERT INTO "); //$NON-NLS-1$ builder.append(tableName); builder.append("("); //$NON-NLS-1$ builder.append(FEATUREMAP_REVISION_ID); builder.append(", "); //$NON-NLS-1$ builder.append(FEATUREMAP_BRANCH); builder.append(", "); //$NON-NLS-1$ builder.append(FEATUREMAP_VERSION_ADDED); builder.append(", "); //$NON-NLS-1$ builder.append(FEATUREMAP_VERSION_REMOVED); builder.append(", "); //$NON-NLS-1$ builder.append(FEATUREMAP_IDX); builder.append(", "); //$NON-NLS-1$ builder.append(FEATUREMAP_TAG); for (int i = 0; i < columnNames.size(); i++) { builder.append(", "); //$NON-NLS-1$ builder.append(columnNames.get(i)); } builder.append(") VALUES (?, ?, ?, ?, ?, ?"); //$NON-NLS-1$ for (int i = 0; i < columnNames.size(); i++) { builder.append(", ?"); //$NON-NLS-1$ } builder.append(")"); //$NON-NLS-1$ sqlInsert = builder.toString(); // ----------------- remove current entry ----------------- builder = new StringBuilder("UPDATE "); //$NON-NLS-1$ builder.append(tableName); builder.append(" SET "); //$NON-NLS-1$ builder.append(FEATUREMAP_VERSION_REMOVED); builder.append("=? "); //$NON-NLS-1$ builder.append(" WHERE "); //$NON-NLS-1$ builder.append(FEATUREMAP_REVISION_ID); builder.append("=? AND "); //$NON-NLS-1$ builder.append(FEATUREMAP_BRANCH); builder.append("=? AND "); //$NON-NLS-1$ builder.append(FEATUREMAP_IDX); builder.append("=? AND "); //$NON-NLS-1$ builder.append(FEATUREMAP_VERSION_REMOVED); builder.append(" IS NULL"); //$NON-NLS-1$ sqlRemoveEntry = builder.toString(); // ----------------- delete temporary entry ----------------- builder = new StringBuilder("DELETE FROM "); //$NON-NLS-1$ builder.append(tableName); builder.append(" WHERE "); //$NON-NLS-1$ builder.append(FEATUREMAP_REVISION_ID); builder.append("=? AND "); //$NON-NLS-1$ builder.append(FEATUREMAP_BRANCH); builder.append("=? AND "); //$NON-NLS-1$ builder.append(FEATUREMAP_IDX); builder.append("=? AND "); //$NON-NLS-1$ builder.append(FEATUREMAP_VERSION_ADDED); builder.append("=?"); //$NON-NLS-1$ sqlDeleteEntry = builder.toString(); // ----------------- update index ----------------- builder = new StringBuilder("UPDATE "); //$NON-NLS-1$ builder.append(tableName); builder.append(" SET "); //$NON-NLS-1$ builder.append(FEATUREMAP_IDX); builder.append("=? WHERE "); //$NON-NLS-1$ builder.append(FEATUREMAP_REVISION_ID); builder.append("=? AND "); //$NON-NLS-1$ builder.append(FEATUREMAP_BRANCH); builder.append("=? AND "); //$NON-NLS-1$ builder.append(FEATUREMAP_VERSION_ADDED); builder.append("=? AND "); //$NON-NLS-1$ builder.append(FEATUREMAP_IDX); builder.append("=?"); //$NON-NLS-1$ sqlUpdateIndex = builder.toString(); // ----------------- get current value ----------------- builder = new StringBuilder("SELECT "); //$NON-NLS-1$ builder.append(FEATUREMAP_TAG); builder.append(", "); //$NON-NLS-1$ iter = columnNames.iterator(); while (iter.hasNext()) { builder.append(iter.next()); if (iter.hasNext()) { builder.append(", "); //$NON-NLS-1$ } } builder.append(" FROM "); //$NON-NLS-1$ builder.append(tableName); builder.append(" WHERE "); //$NON-NLS-1$ builder.append(FEATUREMAP_REVISION_ID); builder.append("=? AND "); //$NON-NLS-1$ builder.append(FEATUREMAP_BRANCH); builder.append("=? AND "); //$NON-NLS-1$ builder.append(FEATUREMAP_IDX); builder.append("=? AND "); //$NON-NLS-1$ builder.append(FEATUREMAP_VERSION_REMOVED); builder.append(" IS NULL"); //$NON-NLS-1$ sqlGetValue = builder.toString(); // ----------- clear list items ------------------------- builder = new StringBuilder("UPDATE "); //$NON-NLS-1$ builder.append(tableName); builder.append(" SET "); //$NON-NLS-1$ builder.append(FEATUREMAP_VERSION_REMOVED); builder.append("=? "); //$NON-NLS-1$ builder.append(" WHERE "); //$NON-NLS-1$ builder.append(FEATUREMAP_REVISION_ID); builder.append("=? AND "); //$NON-NLS-1$ builder.append(FEATUREMAP_BRANCH); builder.append("=? AND "); //$NON-NLS-1$ builder.append(FEATUREMAP_VERSION_REMOVED); builder.append(" IS NULL"); //$NON-NLS-1$ sqlClearList = builder.toString(); } public Collection<IDBTable> getDBTables() { return Collections.singleton(table); } protected List<DBType> getDBTypes() { return dbTypes; } protected final IDBTable getTable() { return table; } protected final List<String> getColumnNames() { return columnNames; } protected final Map<CDOID, ITypeMapping> getTypeMappings() { return typeMappings; } protected final Map<CDOID, String> getTagMap() { return tagMap; } public void readValues(IDBStoreAccessor accessor, InternalCDORevision revision, int listChunk) { MoveableList<Object> list = revision.getList(getFeature()); int valuesToRead = list.size(); if (listChunk != CDORevision.UNCHUNKED && listChunk < valuesToRead) { valuesToRead = listChunk; } if (valuesToRead == 0) { // nothing to read take shortcut return; } if (TRACER.isEnabled()) { TRACER.format("Reading list values for feature {0}.{1} of {2}", getContainingClass().getName(), getFeature() //$NON-NLS-1$ .getName(), revision); } String sql = sqlSelectChunksPrefix + sqlOrderByIndex; IIDHandler idHandler = getMappingStrategy().getStore().getIDHandler(); IDBPreparedStatement stmt = accessor.getDBConnection().prepareStatement(sql, ReuseProbability.HIGH); ResultSet resultSet = null; IStoreChunkReader baseReader = null; try { CDOID id = revision.getID(); int branchID = revision.getBranch().getID(); idHandler.setCDOID(stmt, 1, id); stmt.setInt(2, branchID); stmt.setInt(3, revision.getVersion()); stmt.setInt(4, revision.getVersion()); stmt.setMaxRows(valuesToRead); // optimization - don't read unneeded rows. resultSet = stmt.executeQuery(); int currentIndex = 0; while (valuesToRead > 0 && resultSet.next()) { int index = resultSet.getInt(1); if (index > currentIndex) { if (baseReader == null) { baseReader = createBaseChunkReader(accessor, id, branchID); } baseReader.addRangedChunk(currentIndex, index); if (TRACER.isEnabled()) { TRACER.format("Scheduling range {0}-{1} to be read from base revision", currentIndex, index); //$NON-NLS-1$ } valuesToRead -= index - currentIndex; currentIndex = index; } CDOID tag = idHandler.getCDOID(resultSet, 2); Object value = getTypeMapping(tag).readValue(resultSet); if (TRACER.isEnabled()) { TRACER.format("Read value for index {0} from result set: {1}", currentIndex, value); //$NON-NLS-1$ } list.set(currentIndex++, CDORevisionUtil.createFeatureMapEntry(getFeatureByTag(tag), value)); valuesToRead--; } if (valuesToRead > 0) { if (baseReader == null) { baseReader = createBaseChunkReader(accessor, id, branchID); } baseReader.addRangedChunk(currentIndex, currentIndex + valuesToRead); } } catch (SQLException ex) { throw new DBException(ex); } finally { DBUtil.close(resultSet); DBUtil.close(stmt); } if (baseReader != null) { if (TRACER.isEnabled()) { TRACER.format("Reading base revision chunks for featureMap {0}.{1} of {2} from base revision {3}", //$NON-NLS-1$ getContainingClass().getName(), getFeature().getName(), revision, baseReader.getRevision()); } List<Chunk> baseChunks = baseReader.executeRead(); for (Chunk chunk : baseChunks) { int startIndex = chunk.getStartIndex(); for (int i = 0; i < chunk.size(); i++) { list.set(startIndex + i, chunk.get(i)); } } } if (TRACER.isEnabled()) { TRACER.format("Reading list values done for feature {0}.{1} of {2}", getContainingClass().getName(), //$NON-NLS-1$ getFeature().getName(), revision); } } public final void readChunks(IDBStoreChunkReader chunkReader, List<Chunk> chunks, String where) { CDORevision revision = chunkReader.getRevision(); if (TRACER.isEnabled()) { TRACER.format("Reading list chunk values for feature {0}.{1} of {2}", getContainingClass().getName(), //$NON-NLS-1$ getFeature().getName(), revision); } StringBuilder builder = new StringBuilder(sqlSelectChunksPrefix); if (where != null) { builder.append(" AND "); //$NON-NLS-1$ builder.append(where); } builder.append(sqlOrderByIndex); String sql = builder.toString(); IIDHandler idHandler = getMappingStrategy().getStore().getIDHandler(); IDBPreparedStatement stmt = chunkReader.getAccessor().getDBConnection().prepareStatement(sql, ReuseProbability.LOW); ResultSet resultSet = null; IStoreChunkReader baseReader = null; try { idHandler.setCDOID(stmt, 1, revision.getID()); stmt.setInt(2, revision.getBranch().getID()); stmt.setInt(3, revision.getVersion()); stmt.setInt(4, revision.getVersion()); resultSet = stmt.executeQuery(); int nextDBIndex = Integer.MAX_VALUE; // next available DB index if (resultSet.next()) { nextDBIndex = resultSet.getInt(1); } for (Chunk chunk : chunks) { int startIndex = chunk.getStartIndex(); int missingValueStartIndex = -1; for (int i = 0; i < chunk.size(); i++) { int nextListIndex = startIndex + i; // next expected list index if (nextDBIndex == nextListIndex) { // DB value is available. check first if missing indexes were present before. if (missingValueStartIndex != -1) { // read missing indexes from missingValueStartIndex to currentIndex if (baseReader == null) { baseReader = createBaseChunkReader(chunkReader.getAccessor(), chunkReader.getRevision().getID(), chunkReader.getRevision().getBranch().getID()); } if (TRACER.isEnabled()) { TRACER.format("Scheduling range {0}-{1} to be read from base revision", missingValueStartIndex, //$NON-NLS-1$ nextListIndex); } baseReader.addRangedChunk(missingValueStartIndex, nextListIndex); // reset missingValueStartIndex missingValueStartIndex = -1; } // now read value and set to chunk CDOID tag = idHandler.getCDOID(resultSet, 2); Object value = getTypeMapping(tag).readValue(resultSet); if (TRACER.isEnabled()) { TRACER.format("ChunkReader read value for index {0} from result set: {1}", nextDBIndex, value); //$NON-NLS-1$ } chunk.add(i, CDORevisionUtil.createFeatureMapEntry(getFeatureByTag(tag), value)); // advance DB cursor and read next available index if (resultSet.next()) { nextDBIndex = resultSet.getInt(1); } else { // no more DB indexes available, but we have to continue checking for gaps, therefore set to MAX_VALUE nextDBIndex = Integer.MAX_VALUE; } } else { // gap between next DB index and next list index detected. // skip until end of chunk or until DB value becomes available if (missingValueStartIndex == -1) { missingValueStartIndex = nextListIndex; } } } // chunk complete. check for missing values at the end of the chunk. if (missingValueStartIndex != -1) { // read missing indexes from missingValueStartIndex to last chunk index if (baseReader == null) { baseReader = createBaseChunkReader(chunkReader.getAccessor(), chunkReader.getRevision().getID(), chunkReader.getRevision().getBranch().getID()); } baseReader.addRangedChunk(missingValueStartIndex, chunk.getStartIndex() + chunk.size()); } } } catch (SQLException ex) { throw new DBException(ex); } finally { DBUtil.close(resultSet); DBUtil.close(stmt); } // now read missing values from base revision. if (baseReader != null) { List<Chunk> baseChunks = baseReader.executeRead(); Iterator<Chunk> thisIterator = chunks.iterator(); Chunk thisChunk = thisIterator.next(); for (Chunk baseChunk : baseChunks) { int baseStartIndex = baseChunk.getStartIndex(); while (baseStartIndex > thisChunk.getStartIndex() + thisChunk.size()) { // advance thisChunk, because it does not match baseChunk thisChunk = thisIterator.next(); } // baseChunk now corresponds to this chunk, but startIndex of baseChunk may be higher. // therefore calculate offset int offset = thisChunk.getStartIndex() - baseStartIndex; // and copy values. for (int i = 0; i < baseChunk.size(); i++) { thisChunk.add(i + offset, baseChunk.get(i)); } } // finally, continue with the next baseChunk } if (TRACER.isEnabled()) { TRACER.format("Reading list chunk values done for feature {0}.{1} of {2}", getContainingClass().getName(), //$NON-NLS-1$ getFeature(), revision); } } public void writeValues(IDBStoreAccessor accessor, InternalCDORevision revision) { CDOList values = revision.getList(getFeature()); int idx = 0; for (Object element : values) { writeValue(accessor, revision, idx++, element); } if (TRACER.isEnabled()) { TRACER.format("Writing done"); //$NON-NLS-1$ } } protected final void writeValue(IDBStoreAccessor accessor, CDORevision revision, int idx, Object value) { if (TRACER.isEnabled()) { TRACER.format("Writing value for feature {0}.{1} index {2} of {3} : {4}", getContainingClass().getName(), //$NON-NLS-1$ getFeature(), idx, revision, value); } addEntry(accessor, revision.getID(), revision.getBranch().getID(), revision.getVersion(), idx, value, revision.getTimeStamp()); } /** * Get column name (lazy). * * @param tag * The feature's MetaID in CDO * @return the column name where the values are stored */ protected String getColumnName(CDOID tag) { String column = tagMap.get(tag); if (column == null) { addFeature(tag); column = tagMap.get(tag); } return column; } /** * Get type mapping (lazy). * * @param tag * The feature's MetaID in CDO * @return the corresponding type mapping */ protected ITypeMapping getTypeMapping(CDOID tag) { ITypeMapping typeMapping = typeMappings.get(tag); if (typeMapping == null) { addFeature(tag); typeMapping = typeMappings.get(tag); } return typeMapping; } private void addFeature(CDOID tag) { EStructuralFeature modelFeature = getFeatureByTag(tag); ITypeMapping typeMapping = getMappingStrategy().createValueMapping(modelFeature); String column = FEATUREMAP_VALUE + "_" + typeMapping.getDBType(); //$NON-NLS-1$ tagMap.put(tag, column); typeMapping.setDBField(table, column); typeMappings.put(tag, typeMapping); } /** * @param metaID * @return the column name where the values are stored */ private EStructuralFeature getFeatureByTag(CDOID tag) { return (EStructuralFeature)getMappingStrategy().getStore().getMetaDataManager().getMetaInstance(tag); } /** * @param feature * The EStructuralFeature * @return The feature's MetaID in CDO */ protected CDOID getTagByFeature(EStructuralFeature feature, long created) { return getMappingStrategy().getStore().getMetaDataManager().getMetaID(feature, created); } /** * Clear a list of a given revision. * * @param accessor * the accessor to use * @param id * the id of the revision from which to remove all items */ public void clearList(IDBStoreAccessor accessor, CDOID id, int branchId, int oldVersion, int newVersion, int lastIndex, long timestamp) { // check for each index if the value exists in the current branch for (int i = 0; i <= lastIndex; i++) { if (getValue(accessor, id, branchId, i, false) == null) { // if not, add a historic entry for missing ones. addHistoricEntry(accessor, id, branchId, 0, newVersion, i, getValueFromBase(accessor, id, branchId, i), timestamp); } } IIDHandler idHandler = getMappingStrategy().getStore().getIDHandler(); IDBPreparedStatement stmt = accessor.getDBConnection().prepareStatement(sqlClearList, ReuseProbability.HIGH); try { // clear rest of the list stmt.setInt(1, newVersion); idHandler.setCDOID(stmt, 2, id); stmt.setInt(3, branchId); int result = DBUtil.update(stmt, false); if (TRACER.isEnabled()) { TRACER.format("ClearList result: {0}", result); //$NON-NLS-1$ } } catch (SQLException e) { throw new DBException(e); } finally { DBUtil.close(stmt); } } public void objectDetached(IDBStoreAccessor accessor, CDOID id, long revised) { InternalCDORevision revision = (InternalCDORevision)accessor.getTransaction().getRevision(id); int branchId = accessor.getTransaction().getBranch().getID(); if (TRACER.isEnabled()) { TRACER.format("objectDetached {1}", revision); //$NON-NLS-1$ } clearList(accessor, id, branchId, revision.getVersion(), FINAL_VERSION, revision.getList(getFeature()).size() - 1, revised); } @Override public void rawDeleted(IDBStoreAccessor accessor, CDOID id, CDOBranch branch, int version) { throw new UnsupportedOperationException("Raw deletion does not work in range-based mappings"); } public void processDelta(final IDBStoreAccessor accessor, final CDOID id, final int branchId, int oldVersion, final int newVersion, long created, CDOListFeatureDelta delta) { List<CDOFeatureDelta> listChanges = delta.getListChanges(); if (listChanges.size() == 0) { // nothing to do. return; } InternalCDORevision originalRevision = (InternalCDORevision)accessor.getTransaction().getRevision(id); int oldListSize = originalRevision.getList(getFeature()).size(); if (TRACER.isEnabled()) { TRACER.format("ListTableMapping.processDelta for revision {0} - previous list size: {1}", originalRevision, //$NON-NLS-1$ oldListSize); } // let the visitor collect the changes ListDeltaVisitor visitor = new ListDeltaVisitor(accessor, originalRevision, branchId, oldVersion, newVersion, created); if (TRACER.isEnabled()) { TRACER.format("Processing deltas..."); //$NON-NLS-1$ } // optimization: it's only necessary to process deltas // starting with the last feature delta which clears the list // (any operation before the clear is cascaded by it anyway) int index = listChanges.size() - 1; while (index > 0) { CDOFeatureDelta listDelta = listChanges.get(index); if (listDelta instanceof CDOClearFeatureDelta || listDelta instanceof CDOUnsetFeatureDelta) { break; } index--; } while (index < listChanges.size()) { listChanges.get(index++).accept(visitor); } } private class ListDeltaVisitor implements CDOFeatureDeltaVisitor { private IDBStoreAccessor accessor; private InternalCDORevision originalRevision; private CDOID id; private int branchID; private int oldVersion; private int newVersion; private int lastIndex; private long timestamp; public ListDeltaVisitor(IDBStoreAccessor accessor, InternalCDORevision originalRevision, int targetBranchID, int oldVersion, int newVersion, long timestamp) { this.accessor = accessor; this.originalRevision = originalRevision; id = this.originalRevision.getID(); branchID = targetBranchID; this.oldVersion = oldVersion; this.newVersion = newVersion; lastIndex = originalRevision.getList(getFeature()).size() - 1; this.timestamp = timestamp; } public void visit(CDOMoveFeatureDelta delta) { int fromIdx = delta.getOldPosition(); int toIdx = delta.getNewPosition(); if (TRACER.isEnabled()) { TRACER.format("Delta Moving: {0} to {1}", fromIdx, toIdx); //$NON-NLS-1$ } Object value = getValue(accessor, id, branchID, fromIdx, true); // remove the item removeEntry(accessor, id, branchID, oldVersion, newVersion, fromIdx, timestamp); // adjust indexes and shift either up or down if (fromIdx < toIdx) { moveOneUp(accessor, id, branchID, oldVersion, newVersion, fromIdx + 1, toIdx); } else { // fromIdx > toIdx here moveOneDown(accessor, id, branchID, oldVersion, newVersion, toIdx, fromIdx - 1); } // create the item addEntry(accessor, id, branchID, newVersion, toIdx, value, timestamp); } public void visit(CDOAddFeatureDelta delta) { int startIndex = delta.getIndex(); int endIndex = lastIndex; if (TRACER.isEnabled()) { TRACER.format("Delta Adding at: {0}", startIndex); //$NON-NLS-1$ } if (startIndex <= endIndex) { // make room for the new item moveOneDown(accessor, id, branchID, oldVersion, newVersion, startIndex, endIndex); } // create the item addEntry(accessor, id, branchID, newVersion, startIndex, delta.getValue(), timestamp); ++lastIndex; } public void visit(CDORemoveFeatureDelta delta) { int startIndex = delta.getIndex(); int endIndex = lastIndex; if (TRACER.isEnabled()) { TRACER.format("Delta Removing at: {0}", startIndex); //$NON-NLS-1$ } // remove the item removeEntry(accessor, id, branchID, oldVersion, newVersion, startIndex, timestamp); // make room for the new item moveOneUp(accessor, id, branchID, oldVersion, newVersion, startIndex + 1, endIndex); --lastIndex; } public void visit(CDOSetFeatureDelta delta) { int index = delta.getIndex(); if (TRACER.isEnabled()) { TRACER.format("Delta Setting at: {0}", index); //$NON-NLS-1$ } // remove the item removeEntry(accessor, id, branchID, oldVersion, newVersion, index, timestamp); // create the item addEntry(accessor, id, branchID, newVersion, index, delta.getValue(), timestamp); } public void visit(CDOUnsetFeatureDelta delta) { if (delta.getFeature().isUnsettable()) { throw new ImplementationError("Should not be called"); //$NON-NLS-1$ } if (TRACER.isEnabled()) { TRACER.format("Delta Unsetting"); //$NON-NLS-1$ } clearList(accessor, id, branchID, oldVersion, newVersion, lastIndex, timestamp); lastIndex = -1; } public void visit(CDOListFeatureDelta delta) { throw new ImplementationError("Should not be called"); //$NON-NLS-1$ } public void visit(CDOClearFeatureDelta delta) { if (TRACER.isEnabled()) { TRACER.format("Delta Clearing"); //$NON-NLS-1$ } clearList(accessor, id, branchID, oldVersion, newVersion, lastIndex, timestamp); lastIndex = -1; } public void visit(CDOContainerFeatureDelta delta) { throw new ImplementationError("Should not be called"); //$NON-NLS-1$ } private void moveOneUp(IDBStoreAccessor accessor, CDOID id, int branchId, int oldVersion, int newVersion, int startIndex, int endIndex) { IIDHandler idHandler = getMappingStrategy().getStore().getIDHandler(); IDBPreparedStatement stmt = accessor.getDBConnection().prepareStatement(sqlUpdateIndex, ReuseProbability.HIGH); try { for (int index = startIndex; index <= endIndex; ++index) { if (TRACER.isEnabled()) { TRACER.format("moveOneUp moving: {0} -> {1}", index, index - 1); //$NON-NLS-1$ } int column = 1; stmt.setInt(column++, index - 1); idHandler.setCDOID(stmt, startIndex++, id); stmt.setInt(column++, branchId); stmt.setInt(column++, newVersion); stmt.setInt(column++, index); int result = DBUtil.update(stmt, false); switch (result) { case 1: // entry for current revision was already present. // index update succeeded. if (TRACER.isEnabled()) { TRACER.format("moveOneUp updated: {0} -> {1}", index, index - 1); //$NON-NLS-1$ } break; case 0: Object value = getValue(accessor, id, branchId, index, false); if (value != null) { if (TRACER.isEnabled()) { TRACER.format("moveOneUp remove: {0}", index); //$NON-NLS-1$ } removeEntry(accessor, id, branchId, oldVersion, newVersion, index, timestamp); } else { value = getValueFromBase(accessor, id, branchId, index); { TRACER.format("moveOneUp add historic entry at: {0}", index); //$NON-NLS-1$ } addHistoricEntry(accessor, id, branchId, 0, newVersion, index, value, timestamp); } if (TRACER.isEnabled()) { TRACER.format("moveOneUp add: {0}", index - 1); //$NON-NLS-1$ } addEntry(accessor, id, branchId, newVersion, index - 1, value, timestamp); break; default: if (TRACER.isEnabled()) { TRACER.format("moveOneUp Too many results: {0} -> {1}: {2}", index, index + 1, result); //$NON-NLS-1$ } throw new DBException("Too many results"); //$NON-NLS-1$ } } } catch (SQLException e) { throw new DBException(e); } finally { DBUtil.close(stmt); } } private void moveOneDown(IDBStoreAccessor accessor, CDOID id, int branchId, int oldVersion, int newVersion, int startIndex, int endIndex) { IIDHandler idHandler = getMappingStrategy().getStore().getIDHandler(); IDBPreparedStatement stmt = accessor.getDBConnection().prepareStatement(sqlUpdateIndex, ReuseProbability.HIGH); try { for (int index = endIndex; index >= startIndex; --index) { if (TRACER.isEnabled()) { TRACER.format("moveOneDown moving: {0} -> {1}", index, index + 1); //$NON-NLS-1$ } int column = 1; stmt.setInt(column++, index + 1); idHandler.setCDOID(stmt, column++, id); stmt.setInt(column++, branchId); stmt.setInt(column++, newVersion); stmt.setInt(column++, index); int result = DBUtil.update(stmt, false); switch (result) { case 1: // entry for current revision was already present. // index update succeeded. if (TRACER.isEnabled()) { TRACER.format("moveOneDown updated: {0} -> {1}", index, index + 1); //$NON-NLS-1$ } break; case 0: Object value = getValue(accessor, id, branchId, index, false); if (value != null) { if (TRACER.isEnabled()) { TRACER.format("moveOneDown remove: {0}", index); //$NON-NLS-1$ } removeEntry(accessor, id, branchId, oldVersion, newVersion, index, timestamp); } else { value = getValueFromBase(accessor, id, branchId, index); { TRACER.format("moveOneDown add historic entry at: {0}", index); //$NON-NLS-1$ } addHistoricEntry(accessor, id, branchId, 0, newVersion, index, value, timestamp); } addEntry(accessor, id, branchId, newVersion, index + 1, value, timestamp); break; default: if (TRACER.isEnabled()) { TRACER.format("moveOneDown Too many results: {0} -> {1}: {2}", index, index + 1, result); //$NON-NLS-1$ } throw new DBException("Too many results"); //$NON-NLS-1$ } } } catch (SQLException e) { throw new DBException(e); } finally { DBUtil.close(stmt); } } } private void addEntry(IDBStoreAccessor accessor, CDOID id, int branchId, int version, int index, Object value, long timestamp) { if (TRACER.isEnabled()) { TRACER.format("Adding value for feature() {0}.{1} index {2} of {3}v{4} : {5}", //$NON-NLS-1$ getContainingClass().getName(), getFeature().getName(), index, id, version, value); } FeatureMap.Entry entry = (FeatureMap.Entry)value; EStructuralFeature entryFeature = entry.getEStructuralFeature(); CDOID tag = getTagByFeature(entryFeature, timestamp); String columnName = getColumnName(tag); IIDHandler idHandler = getMappingStrategy().getStore().getIDHandler(); IDBPreparedStatement stmt = accessor.getDBConnection().prepareStatement(sqlInsert, ReuseProbability.HIGH); try { int column = 1; idHandler.setCDOID(stmt, column++, id); stmt.setInt(column++, branchId); stmt.setInt(column++, version); stmt.setNull(column++, DBType.INTEGER.getCode()); // versionRemoved stmt.setInt(column++, index); idHandler.setCDOID(stmt, column++, tag); for (int i = 0; i < columnNames.size(); i++) { if (columnNames.get(i).equals(columnName)) { getTypeMapping(tag).setValue(stmt, column++, entry.getValue()); } else { stmt.setNull(column++, getDBTypes().get(i).getCode()); } } DBUtil.update(stmt, true); } catch (SQLException e) { throw new DBException(e); } catch (IllegalStateException e) { throw new DBException(e); } finally { DBUtil.close(stmt); } } private void addHistoricEntry(IDBStoreAccessor accessor, CDOID id, int branchId, int versionAdded, int versionRemoved, int index, Object value, long timestamp) { if (TRACER.isEnabled()) { TRACER.format("Adding historic value for feature {0}.{1} index {2} of {3}:{4}v{5}-v{6} : {7}", //$NON-NLS-1$ getContainingClass().getName(), getFeature().getName(), index, id, branchId, versionAdded, versionRemoved, value); } FeatureMap.Entry entry = (FeatureMap.Entry)value; EStructuralFeature entryFeature = entry.getEStructuralFeature(); CDOID tag = getTagByFeature(entryFeature, timestamp); ITypeMapping typeMapping = getTypeMapping(tag); String columnName = getColumnName(tag); IIDHandler idHandler = getMappingStrategy().getStore().getIDHandler(); IDBPreparedStatement stmt = accessor.getDBConnection().prepareStatement(sqlInsert, ReuseProbability.HIGH); try { int column = 1; idHandler.setCDOID(stmt, column++, id); stmt.setInt(column++, branchId); stmt.setInt(column++, versionAdded); stmt.setNull(column++, versionRemoved); stmt.setInt(column++, index); idHandler.setCDOID(stmt, column++, tag); for (int i = 0; i < columnNames.size(); i++) { if (columnNames.get(i).equals(columnName)) { typeMapping.setValue(stmt, column++, entry.getValue()); } else { stmt.setNull(column++, getDBTypes().get(i).getCode()); } } DBUtil.update(stmt, true); } catch (SQLException e) { throw new DBException(e); } catch (IllegalStateException e) { throw new DBException(e); } finally { DBUtil.close(stmt); } } private void removeEntry(IDBStoreAccessor accessor, CDOID id, int branchId, int oldVersion, int newVersion, int index, long timestamp) { if (TRACER.isEnabled()) { TRACER.format("Removing value for feature() {0}.{1} index {2} of {3}v{4}", //$NON-NLS-1$ getContainingClass().getName(), getFeature().getName(), index, id, newVersion); } IIDHandler idHandler = getMappingStrategy().getStore().getIDHandler(); IDBPreparedStatement stmt = accessor.getDBConnection().prepareStatement(sqlDeleteEntry, ReuseProbability.HIGH); try { // try to delete a temporary entry first int column = 1; idHandler.setCDOID(stmt, column++, id); stmt.setInt(column++, branchId); stmt.setInt(column++, index); stmt.setInt(column++, newVersion); int result = DBUtil.update(stmt, false); if (result == 1) { if (TRACER.isEnabled()) { TRACER.format("removeEntry deleted: {0}", index); //$NON-NLS-1$ } } else if (result > 1) { if (TRACER.isEnabled()) { TRACER.format("removeEntry Too many results: {0}: {1}", index, result); //$NON-NLS-1$ } throw new DBException("Too many results"); //$NON-NLS-1$ } else { // no temporary entry found, so mark the entry as removed DBUtil.close(stmt); stmt = accessor.getDBConnection().prepareStatement(sqlRemoveEntry, ReuseProbability.HIGH); column = 1; stmt.setInt(column++, newVersion); idHandler.setCDOID(stmt, column++, id); stmt.setInt(column++, branchId); stmt.setInt(column++, index); result = DBUtil.update(stmt, false); if (result == 0) { // no entry removed -> this means that we are in a branch and // the entry has not been modified since the branch fork. // therefore, we have to copy the base value and mark it as removed Object value = getValueFromBase(accessor, id, branchId, index); addHistoricEntry(accessor, id, branchId, 0, newVersion, index, value, timestamp); } } } catch (SQLException e) { if (TRACER.isEnabled()) { TRACER.format("Removing value for feature() {0}.{1} index {2} of {3}v{4} FAILED {5}", //$NON-NLS-1$ getContainingClass().getName(), getFeature().getName(), index, id, newVersion, e.getMessage()); } throw new DBException(e); } catch (IllegalStateException e) { if (TRACER.isEnabled()) { TRACER.format("Removing value for feature() {0}.{1} index {2} of {3}v{4} FAILED {5}", //$NON-NLS-1$ getContainingClass().getName(), getFeature().getName(), index, id, newVersion, e.getMessage()); } throw new DBException(e); } finally { DBUtil.close(stmt); } } private FeatureMap.Entry getValue(IDBStoreAccessor accessor, CDOID id, int branchId, int index, boolean getFromBase) { IIDHandler idHandler = getMappingStrategy().getStore().getIDHandler(); IDBPreparedStatement stmt = accessor.getDBConnection().prepareStatement(sqlGetValue, ReuseProbability.HIGH); FeatureMap.Entry result = null; try { int column = 1; idHandler.setCDOID(stmt, column++, id); stmt.setInt(column++, branchId); stmt.setInt(column++, index); ResultSet resultSet = stmt.executeQuery(); if (resultSet.next()) { CDOID tag = idHandler.getCDOID(resultSet, 1); Object value = getTypeMapping(tag).readValue(resultSet); result = CDORevisionUtil.createFeatureMapEntry(getFeatureByTag(tag), value); } else { // value is not in this branch. // -> read from base revision if (getFromBase) { result = getValueFromBase(accessor, id, branchId, index); } // else: result remains null } if (TRACER.isEnabled()) { TRACER.format("Read value (index {0}) from result set: {1}", index, result); //$NON-NLS-1$ } } catch (SQLException e) { throw new DBException(e); } finally { DBUtil.close(stmt); } return result; } /** * Read a single value (at a given index) from the base revision * * @param accessor * the DBStoreAccessor * @param id * the ID of the revision * @param branchID * the ID of the current (child) branch * @param index * the index to read the value from * @return the value which is at index <code>index</code> in revision with ID <code>id</code> in the parent branch at * the base of this branch (indicated by <code>branchID</code>). */ private FeatureMap.Entry getValueFromBase(IDBStoreAccessor accessor, CDOID id, int branchID, int index) { IStoreChunkReader chunkReader = createBaseChunkReader(accessor, id, branchID); chunkReader.addSimpleChunk(index); List<Chunk> chunks = chunkReader.executeRead(); return (FeatureMap.Entry)chunks.get(0).get(0); } private IStoreChunkReader createBaseChunkReader(IDBStoreAccessor accessor, CDOID id, int branchID) { InternalRepository repository = (InternalRepository)accessor.getStore().getRepository(); CDOBranchManager branchManager = repository.getBranchManager(); CDOBranch branch = branchManager.getBranch(branchID); CDOBranchPoint base = branch.getBase(); if (base.getBranch() == null) { throw new IllegalArgumentException("Base branch is null: " + branch); } InternalCDORevisionManager revisionManager = repository.getRevisionManager(); InternalCDORevision baseRevision = revisionManager.getRevision(id, base, 0, CDORevision.DEPTH_NONE, true); return accessor.createChunkReader(baseRevision, getFeature()); } public final boolean queryXRefs(IDBStoreAccessor accessor, String mainTableName, String mainTableWhere, QueryXRefsContext context, String idString) { // must never be called (a feature map is not associated with an EReference feature, so XRefs are nor supported // here) throw new ImplementationError("Should never be called!"); } }