/*
* CDDL HEADER START
*
* The contents of this file are subject to the terms of the
* Common Development and Distribution License, Version 1.0 only
* (the "License"). You may not use this file except in compliance
* with the License.
*
* You can obtain a copy of the license at
* trunk/opends/resource/legal-notices/OpenDS.LICENSE
* or https://OpenDS.dev.java.net/OpenDS.LICENSE.
* See the License for the specific language governing permissions
* and limitations under the License.
*
* When distributing Covered Code, include this CDDL HEADER in each
* file and include the License file at
* trunk/opends/resource/legal-notices/OpenDS.LICENSE. If applicable,
* add the following below this CDDL HEADER, with the fields enclosed
* by brackets "[]" replaced with your own identifying information:
* Portions Copyright [yyyy] [name of copyright owner]
*
* CDDL HEADER END
*
*
* Copyright 2006-2008 Sun Microsystems, Inc.
*/
package org.opends.server.backends;
import java.util.HashMap;
import java.util.HashSet;
import java.util.LinkedHashMap;
import java.util.LinkedList;
import java.util.Set;
import org.opends.messages.Message;
import org.opends.server.admin.Configuration;
import org.opends.server.admin.std.server.MemoryBackendCfg;
import org.opends.server.api.Backend;
import org.opends.server.config.ConfigException;
import org.opends.server.controls.SubtreeDeleteControl;
import org.opends.server.core.AddOperation;
import org.opends.server.core.DeleteOperation;
import org.opends.server.core.DirectoryServer;
import org.opends.server.core.ModifyOperation;
import org.opends.server.core.ModifyDNOperation;
import org.opends.server.core.SearchOperation;
import org.opends.server.loggers.debug.DebugTracer;
import org.opends.server.types.AttributeType;
import org.opends.server.types.BackupConfig;
import org.opends.server.types.BackupDirectory;
import org.opends.server.types.ConditionResult;
import org.opends.server.types.Control;
import org.opends.server.types.DebugLogLevel;
import org.opends.server.types.DirectoryException;
import org.opends.server.types.DN;
import org.opends.server.types.Entry;
import org.opends.server.types.IndexType;
import org.opends.server.types.InitializationException;
import org.opends.server.types.LDIFExportConfig;
import org.opends.server.types.LDIFImportConfig;
import org.opends.server.types.LDIFImportResult;
import org.opends.server.types.RestoreConfig;
import org.opends.server.types.ResultCode;
import org.opends.server.types.SearchFilter;
import org.opends.server.types.SearchScope;
import org.opends.server.util.LDIFException;
import org.opends.server.util.LDIFReader;
import org.opends.server.util.LDIFWriter;
import org.opends.server.util.Validator;
import static org.opends.messages.BackendMessages.*;
import static org.opends.server.loggers.debug.DebugLogger.*;
import static org.opends.server.util.ServerConstants.*;
import static org.opends.server.util.StaticUtils.*;
/**
* This class defines a very simple backend that stores its information in
* memory. This is primarily intended for testing purposes with small data
* sets, as it does not have any indexing mechanism such as would be required to
* achieve high performance with large data sets. It is also heavily
* synchronized for simplicity at the expense of performance, rather than
* providing a more fine-grained locking mechanism.
* <BR><BR>
* Entries stored in this backend are held in a
* <CODE>LinkedHashMap<DN,Entry></CODE> object, which ensures that the
* order in which you iterate over the entries is the same as the order in which
* they were inserted. By combining this with the constraint that no entry can
* be added before its parent, you can ensure that iterating through the entries
* will always process the parent entries before their children, which is
* important for both search result processing and LDIF exports.
* <BR><BR>
* As mentioned above, no data indexing is performed, so all non-baseObject
* searches require iteration through the entire data set. If this is to become
* a more general-purpose backend, then additional
* <CODE>HashMap<ByteString,Set<DN>></CODE> objects could be used
* to provide that capability.
* <BR><BR>
* There is actually one index that does get maintained within this backend,
* which is a mapping between the DN of an entry and the DNs of any immediate
* children of that entry. This is needed to efficiently determine whether an
* entry has any children (which must not be the case for delete operations).
*/
public class MemoryBackend
extends Backend
{
/**
* The tracer object for the debug logger.
*/
private static final DebugTracer TRACER = getTracer();
// The base DNs for this backend.
private DN[] baseDNs;
// The mapping between parent DNs and their immediate children.
private HashMap<DN,HashSet<DN>> childDNs;
// The base DNs for this backend, in a hash set.
private HashSet<DN> baseDNSet;
// The set of supported controls for this backend.
private HashSet<String> supportedControls;
// The set of supported features for this backend.
private HashSet<String> supportedFeatures;
// The mapping between entry DNs and the corresponding entries.
private LinkedHashMap<DN,Entry> entryMap;
/**
* Creates a new backend with the provided information. All backend
* implementations must implement a default constructor that use
* <CODE>super()</CODE> to invoke this constructor.
*/
public MemoryBackend()
{
super();
// Perform all initialization in initializeBackend.
}
/**
* Set the base DNs for this backend. This is used by the unit tests
* to set the base DNs without having to provide a configuration
* object when initializing the backend.
* @param baseDNs The set of base DNs to be served by this memory backend.
*/
public void setBaseDNs(DN[] baseDNs)
{
this.baseDNs = baseDNs;
}
/**
* {@inheritDoc}
*/
@Override()
public void configureBackend(Configuration config)
throws ConfigException
{
if (config != null)
{
Validator.ensureTrue(config instanceof MemoryBackendCfg);
MemoryBackendCfg cfg = (MemoryBackendCfg)config;
DN[] baseDNs = new DN[cfg.getBaseDN().size()];
cfg.getBaseDN().toArray(baseDNs);
setBaseDNs(baseDNs);
}
}
/**
* {@inheritDoc}
*/
@Override()
public synchronized void initializeBackend()
throws ConfigException, InitializationException
{
// We won't support anything other than exactly one base DN in this
// implementation. If we were to add such support in the future, we would
// likely want to separate the data for each base DN into a separate entry
// map.
if ((baseDNs == null) || (baseDNs.length != 1))
{
Message message = ERR_MEMORYBACKEND_REQUIRE_EXACTLY_ONE_BASE.get();
throw new ConfigException(message);
}
baseDNSet = new HashSet<DN>();
for (DN dn : baseDNs)
{
baseDNSet.add(dn);
}
entryMap = new LinkedHashMap<DN,Entry>();
childDNs = new HashMap<DN,HashSet<DN>>();
supportedControls = new HashSet<String>();
supportedControls.add(OID_SUBTREE_DELETE_CONTROL);
supportedFeatures = new HashSet<String>();
for (DN dn : baseDNs)
{
try
{
DirectoryServer.registerBaseDN(dn, this, false);
}
catch (Exception e)
{
if (debugEnabled())
{
TRACER.debugCaught(DebugLogLevel.ERROR, e);
}
Message message = ERR_BACKEND_CANNOT_REGISTER_BASEDN.get(
dn.toString(), getExceptionMessage(e));
throw new InitializationException(message, e);
}
}
}
/**
* Removes any data that may have been stored in this backend.
*/
public synchronized void clearMemoryBackend()
{
entryMap.clear();
childDNs.clear();
}
/**
* {@inheritDoc}
*/
@Override()
public synchronized void finalizeBackend()
{
clearMemoryBackend();
for (DN dn : baseDNs)
{
try
{
DirectoryServer.deregisterBaseDN(dn);
}
catch (Exception e)
{
if (debugEnabled())
{
TRACER.debugCaught(DebugLogLevel.ERROR, e);
}
}
}
}
/**
* {@inheritDoc}
*/
@Override()
public DN[] getBaseDNs()
{
return baseDNs;
}
/**
* {@inheritDoc}
*/
@Override()
public synchronized long getEntryCount()
{
if (entryMap != null)
{
return entryMap.size();
}
return -1;
}
/**
* {@inheritDoc}
*/
@Override()
public boolean isLocal()
{
// For the purposes of this method, this is a local backend.
return true;
}
/**
* {@inheritDoc}
*/
@Override()
public boolean isIndexed(AttributeType attributeType, IndexType indexType)
{
// All searches in this backend will always be considered indexed.
return true;
}
/**
* {@inheritDoc}
*/
@Override
public synchronized ConditionResult hasSubordinates(DN entryDN)
throws DirectoryException
{
long ret = numSubordinates(entryDN, false);
if(ret < 0)
{
return ConditionResult.UNDEFINED;
}
else if(ret == 0)
{
return ConditionResult.FALSE;
}
else
{
return ConditionResult.TRUE;
}
}
/**
* {@inheritDoc}
*/
@Override()
public synchronized long numSubordinates(DN entryDN, boolean subtree)
throws DirectoryException
{
// Try to look up the immediate children for the DN
Set<DN> children = childDNs.get(entryDN);
if (children == null)
{
if(entryMap.get(entryDN) != null)
{
// The entry does exist but just no children.
return 0;
}
return -1;
}
if(!subtree)
{
return children.size();
}
else
{
long count = 0;
for(DN child : children)
{
count += numSubordinates(child, true);
count++;
}
return count;
}
}
/**
* {@inheritDoc}
*/
@Override()
public synchronized Entry getEntry(DN entryDN)
{
Entry entry = entryMap.get(entryDN);
if (entry != null)
{
entry = entry.duplicate(true);
}
return entry;
}
/**
* {@inheritDoc}
*/
@Override()
public synchronized boolean entryExists(DN entryDN)
{
return entryMap.containsKey(entryDN);
}
/**
* {@inheritDoc}
*/
@Override()
public synchronized void addEntry(Entry entry, AddOperation addOperation)
throws DirectoryException
{
Entry e = entry.duplicate(false);
// See if the target entry already exists. If so, then fail.
DN entryDN = e.getDN();
if (entryMap.containsKey(entryDN))
{
Message message =
ERR_MEMORYBACKEND_ENTRY_ALREADY_EXISTS.get(String.valueOf(entryDN));
throw new DirectoryException(ResultCode.ENTRY_ALREADY_EXISTS, message);
}
// If the entry is one of the base DNs, then add it.
if (baseDNSet.contains(entryDN))
{
entryMap.put(entryDN, e);
return;
}
// Get the parent DN and ensure that it exists in the backend.
DN parentDN = entryDN.getParentDNInSuffix();
if (parentDN == null)
{
Message message =
ERR_MEMORYBACKEND_ENTRY_DOESNT_BELONG.get(String.valueOf(entryDN));
throw new DirectoryException(ResultCode.NO_SUCH_OBJECT, message);
}
else if (! entryMap.containsKey(parentDN))
{
Message message = ERR_MEMORYBACKEND_PARENT_DOESNT_EXIST.get(
String.valueOf(entryDN), String.valueOf(parentDN));
throw new DirectoryException(ResultCode.NO_SUCH_OBJECT, message);
}
entryMap.put(entryDN, e);
HashSet<DN> children = childDNs.get(parentDN);
if (children == null)
{
children = new HashSet<DN>();
childDNs.put(parentDN, children);
}
children.add(entryDN);
}
/**
* {@inheritDoc}
*/
@Override()
public synchronized void deleteEntry(DN entryDN,
DeleteOperation deleteOperation)
throws DirectoryException
{
// Make sure the entry exists. If not, then throw an exception.
if (! entryMap.containsKey(entryDN))
{
Message message =
ERR_MEMORYBACKEND_ENTRY_DOESNT_EXIST.get(String.valueOf(entryDN));
throw new DirectoryException(ResultCode.NO_SUCH_OBJECT, message);
}
// Check to see if the entry contains a subtree delete control.
boolean subtreeDelete = false;
if (deleteOperation != null
&& deleteOperation
.getRequestControl(SubtreeDeleteControl.DECODER) != null)
{
subtreeDelete = true;
}
HashSet<DN> children = childDNs.get(entryDN);
if (subtreeDelete)
{
if (children != null)
{
HashSet<DN> childrenCopy = new HashSet<DN>(children);
for (DN childDN : childrenCopy)
{
try
{
deleteEntry(childDN, null);
}
catch (Exception e)
{
// This shouldn't happen, but we want the delete to continue anyway
// so just ignore it if it does for some reason.
if (debugEnabled())
{
TRACER.debugCaught(DebugLogLevel.ERROR, e);
}
}
}
}
}
else
{
// Make sure the entry doesn't have any children. If it does, then throw
// an exception.
if ((children != null) && (! children.isEmpty()))
{
Message message = ERR_MEMORYBACKEND_CANNOT_DELETE_ENTRY_WITH_CHILDREN.
get(String.valueOf(entryDN));
throw new DirectoryException(
ResultCode.NOT_ALLOWED_ON_NONLEAF, message);
}
}
// Remove the entry from the backend. Also remove the reference to it from
// its parent, if applicable.
childDNs.remove(entryDN);
entryMap.remove(entryDN);
DN parentDN = entryDN.getParentDNInSuffix();
if (parentDN != null)
{
HashSet<DN> parentsChildren = childDNs.get(parentDN);
if (parentsChildren != null)
{
parentsChildren.remove(entryDN);
if (parentsChildren.isEmpty())
{
childDNs.remove(parentDN);
}
}
}
}
/**
* {@inheritDoc}
*/
@Override()
public synchronized void replaceEntry(Entry oldEntry, Entry newEntry,
ModifyOperation modifyOperation) throws DirectoryException
{
Entry e = newEntry.duplicate(false);
// Make sure the entry exists. If not, then throw an exception.
DN entryDN = e.getDN();
if (! entryMap.containsKey(entryDN))
{
Message message =
ERR_MEMORYBACKEND_ENTRY_DOESNT_EXIST.get(String.valueOf(entryDN));
throw new DirectoryException(ResultCode.NO_SUCH_OBJECT, message);
}
// Replace the old entry with the new one.
entryMap.put(entryDN, e);
}
/**
* {@inheritDoc}
*/
@Override()
public synchronized void renameEntry(DN currentDN, Entry entry,
ModifyDNOperation modifyDNOperation)
throws DirectoryException
{
Entry e = entry.duplicate(false);
// Make sure that the target entry exists.
if (! entryMap.containsKey(currentDN))
{
Message message =
ERR_MEMORYBACKEND_ENTRY_DOESNT_EXIST.get(String.valueOf(currentDN));
throw new DirectoryException(ResultCode.NO_SUCH_OBJECT, message);
}
// Make sure that the target entry doesn't have any children.
HashSet<DN> children = childDNs.get(currentDN);
if (children != null)
{
if (children.isEmpty())
{
childDNs.remove(currentDN);
}
else
{
Message message = ERR_MEMORYBACKEND_CANNOT_RENAME_ENRY_WITH_CHILDREN.
get(String.valueOf(currentDN));
throw new DirectoryException(
ResultCode.NOT_ALLOWED_ON_NONLEAF, message);
}
}
// Make sure that no entry exists with the new DN.
if (entryMap.containsKey(e.getDN()))
{
Message message =
ERR_MEMORYBACKEND_ENTRY_ALREADY_EXISTS.get(String.valueOf(e.getDN()));
throw new DirectoryException(ResultCode.ENTRY_ALREADY_EXISTS, message);
}
// Make sure that the new DN is in this backend.
boolean matchFound = false;
for (DN dn : baseDNs)
{
if (dn.isAncestorOf(e.getDN()))
{
matchFound = true;
break;
}
}
if (! matchFound)
{
Message message = ERR_MEMORYBACKEND_CANNOT_RENAME_TO_ANOTHER_BACKEND.get(
String.valueOf(currentDN));
throw new DirectoryException(ResultCode.UNWILLING_TO_PERFORM, message);
}
// Make sure that the parent of the new entry exists.
DN parentDN = e.getDN().getParentDNInSuffix();
if ((parentDN == null) || (! entryMap.containsKey(parentDN)))
{
Message message = ERR_MEMORYBACKEND_RENAME_PARENT_DOESNT_EXIST.get(
String.valueOf(currentDN), String.valueOf(parentDN));
throw new DirectoryException(ResultCode.NO_SUCH_OBJECT, message);
}
// Delete the current entry and add the new one.
deleteEntry(currentDN, null);
addEntry(e, null);
}
/**
* {@inheritDoc}
*/
@Override()
public synchronized void search(SearchOperation searchOperation)
throws DirectoryException
{
// Get the base DN, scope, and filter for the search.
DN baseDN = searchOperation.getBaseDN();
SearchScope scope = searchOperation.getScope();
SearchFilter filter = searchOperation.getFilter();
// Make sure the base entry exists if it's supposed to be in this backend.
Entry baseEntry = entryMap.get(baseDN);
if ((baseEntry == null) && handlesEntry(baseDN))
{
DN matchedDN = baseDN.getParentDNInSuffix();
while (matchedDN != null)
{
if (entryMap.containsKey(matchedDN))
{
break;
}
matchedDN = matchedDN.getParentDNInSuffix();
}
Message message =
ERR_MEMORYBACKEND_ENTRY_DOESNT_EXIST.get(String.valueOf(baseDN));
throw new DirectoryException(
ResultCode.NO_SUCH_OBJECT, message, matchedDN, null);
}
if (baseEntry != null)
{
baseEntry = baseEntry.duplicate(true);
}
// If it's a base-level search, then just get that entry and return it if it
// matches the filter.
if (scope == SearchScope.BASE_OBJECT)
{
if (filter.matchesEntry(baseEntry))
{
searchOperation.returnEntry(baseEntry, new LinkedList<Control>());
}
}
else
{
// Walk through all entries and send the ones that match.
for (Entry e : entryMap.values())
{
e = e.duplicate(true);
if (e.matchesBaseAndScope(baseDN, scope) && filter.matchesEntry(e))
{
searchOperation.returnEntry(e, new LinkedList<Control>());
}
}
}
}
/**
* {@inheritDoc}
*/
@Override()
public HashSet<String> getSupportedControls()
{
return supportedControls;
}
/**
* {@inheritDoc}
*/
@Override()
public HashSet<String> getSupportedFeatures()
{
return supportedFeatures;
}
/**
* {@inheritDoc}
*/
@Override()
public boolean supportsLDIFExport()
{
return true;
}
/**
* {@inheritDoc}
*/
@Override()
public synchronized void exportLDIF(LDIFExportConfig exportConfig)
throws DirectoryException
{
// Create the LDIF writer.
LDIFWriter ldifWriter;
try
{
ldifWriter = new LDIFWriter(exportConfig);
}
catch (Exception e)
{
if (debugEnabled())
{
TRACER.debugCaught(DebugLogLevel.ERROR, e);
}
Message message =
ERR_MEMORYBACKEND_CANNOT_CREATE_LDIF_WRITER.get(String.valueOf(e));
throw new DirectoryException(DirectoryServer.getServerErrorResultCode(),
message, e);
}
// Walk through all the entries and write them to LDIF.
DN entryDN = null;
try
{
for (Entry entry : entryMap.values())
{
entryDN = entry.getDN();
ldifWriter.writeEntry(entry);
}
}
catch (Exception e)
{
Message message = ERR_MEMORYBACKEND_CANNOT_WRITE_ENTRY_TO_LDIF.get(
String.valueOf(entryDN), String.valueOf(e));
throw new DirectoryException(DirectoryServer.getServerErrorResultCode(),
message, e);
}
finally
{
try
{
ldifWriter.close();
}
catch (Exception e)
{
if (debugEnabled())
{
TRACER.debugCaught(DebugLogLevel.ERROR, e);
}
}
}
}
/**
* {@inheritDoc}
*/
@Override()
public boolean supportsLDIFImport()
{
return true;
}
/**
* {@inheritDoc}
*/
@Override()
public synchronized LDIFImportResult importLDIF(LDIFImportConfig importConfig)
throws DirectoryException
{
clearMemoryBackend();
LDIFReader reader;
try
{
reader = new LDIFReader(importConfig);
}
catch (Exception e)
{
Message message =
ERR_MEMORYBACKEND_CANNOT_CREATE_LDIF_READER.get(String.valueOf(e));
throw new DirectoryException(DirectoryServer.getServerErrorResultCode(),
message, e);
}
try
{
while (true)
{
Entry e = null;
try
{
e = reader.readEntry();
if (e == null)
{
break;
}
}
catch (LDIFException le)
{
if (! le.canContinueReading())
{
Message message =
ERR_MEMORYBACKEND_ERROR_READING_LDIF.get(String.valueOf(e));
throw new DirectoryException(
DirectoryServer.getServerErrorResultCode(),
message, le);
}
else
{
continue;
}
}
try
{
addEntry(e, null);
}
catch (DirectoryException de)
{
reader.rejectLastEntry(de.getMessageObject());
}
}
return new LDIFImportResult(reader.getEntriesRead(),
reader.getEntriesRejected(),
reader.getEntriesIgnored());
}
catch (DirectoryException de)
{
throw de;
}
catch (Exception e)
{
Message message =
ERR_MEMORYBACKEND_ERROR_DURING_IMPORT.get(String.valueOf(e));
throw new DirectoryException(DirectoryServer.getServerErrorResultCode(),
message, e);
}
finally
{
reader.close();
}
}
/**
* {@inheritDoc}
*/
@Override()
public boolean supportsBackup()
{
// This backend does not provide a backup/restore mechanism.
return false;
}
/**
* {@inheritDoc}
*/
@Override()
public boolean supportsBackup(BackupConfig backupConfig,
StringBuilder unsupportedReason)
{
// This backend does not provide a backup/restore mechanism.
return false;
}
/**
* {@inheritDoc}
*/
@Override()
public void createBackup(BackupConfig backupConfig)
throws DirectoryException
{
Message message = ERR_MEMORYBACKEND_BACKUP_RESTORE_NOT_SUPPORTED.get();
throw new DirectoryException(ResultCode.UNWILLING_TO_PERFORM, message);
}
/**
* {@inheritDoc}
*/
@Override()
public void removeBackup(BackupDirectory backupDirectory,
String backupID)
throws DirectoryException
{
Message message = ERR_MEMORYBACKEND_BACKUP_RESTORE_NOT_SUPPORTED.get();
throw new DirectoryException(ResultCode.UNWILLING_TO_PERFORM, message);
}
/**
* {@inheritDoc}
*/
@Override()
public boolean supportsRestore()
{
// This backend does not provide a backup/restore mechanism.
return false;
}
/**
* {@inheritDoc}
*/
@Override()
public void restoreBackup(RestoreConfig restoreConfig)
throws DirectoryException
{
Message message = ERR_MEMORYBACKEND_BACKUP_RESTORE_NOT_SUPPORTED.get();
throw new DirectoryException(ResultCode.UNWILLING_TO_PERFORM, message);
}
/**
* {@inheritDoc}
*/
@Override
public void preloadEntryCache() throws UnsupportedOperationException {
throw new UnsupportedOperationException("Operation not supported.");
}
}