IndexBuffer.java revision 20fdcbef0d17440c367d2943f9c5799bddfe661f
/*
* CDDL HEADER START
*
* The contents of this file are subject to the terms of the
* Common Development and Distribution License, Version 1.0 only
* (the "License"). You may not use this file except in compliance
* with the License.
*
* You can obtain a copy of the license at legal-notices/CDDLv1_0.txt
* or http://forgerock.org/license/CDDLv1.0.html.
* See the License for the specific language governing permissions
* and limitations under the License.
*
* When distributing Covered Code, include this CDDL HEADER in each
* file and include the License file at legal-notices/CDDLv1_0.txt.
* If applicable, add the following below this CDDL HEADER, with the
* fields enclosed by brackets "[]" replaced with your own identifying
* information:
* Portions Copyright [yyyy] [name of copyright owner]
*
* CDDL HEADER END
*
*
* Copyright 2006-2008 Sun Microsystems, Inc.
* Portions Copyright 2014-2015 ForgeRock AS
*/
package org.opends.server.backends.pluggable;
import static org.opends.server.backends.pluggable.EntryIDSet.newDefinedSet;
import java.util.Iterator;
import java.util.LinkedHashMap;
import java.util.Map;
import java.util.TreeMap;
import java.util.TreeSet;
import org.forgerock.opendj.ldap.ByteString;
import org.opends.server.backends.pluggable.spi.StorageRuntimeException;
import org.opends.server.backends.pluggable.spi.WriteableTransaction;
import org.opends.server.types.DirectoryException;
/**
* A buffered index is used to buffer multiple reads or writes to the
* same index key into a single read or write.
* It can only be used to buffer multiple reads and writes under
* the same transaction. The transaction may be null if it is known
* that there are no other concurrent updates to the index.
*/
class IndexBuffer
{
private final EntryContainer entryContainer;
/**
* The buffered records stored as a map from the record key to the
* buffered value for that key for each index.
*/
private final LinkedHashMap<Index, TreeMap<ByteString, BufferedIndexValues>> bufferedIndexes =
new LinkedHashMap<Index, TreeMap<ByteString, BufferedIndexValues>>();
/** The buffered records stored as a set of buffered VLV values for each index. */
private final LinkedHashMap<VLVIndex, BufferedVLVIndexValues> bufferedVLVIndexes =
new LinkedHashMap<VLVIndex, BufferedVLVIndexValues>();
/**
* A simple class representing a pair of added and deleted indexed IDs. Initially both addedIDs
* and deletedIDs are {@code null} indicating that that the whole record should be deleted. This
* state is only ever used when updating the id2children and id2subtree indexes when deleting an
* entry.
*/
private static class BufferedIndexValues
{
private EntryIDSet addedEntryIDs;
private EntryIDSet deletedEntryIDs;
void addEntryID(EntryID entryID)
{
if (!remove(deletedEntryIDs, entryID))
{
if (this.addedEntryIDs == null)
{
this.addedEntryIDs = newDefinedSet();
}
this.addedEntryIDs.add(entryID);
}
}
void deleteEntryID(EntryID entryID)
{
if (!remove(addedEntryIDs, entryID))
{
if (this.deletedEntryIDs == null)
{
this.deletedEntryIDs = newDefinedSet();
}
this.deletedEntryIDs.add(entryID);
}
}
private static boolean remove(EntryIDSet entryIDs, EntryID entryID)
{
return entryIDs != null ? entryIDs.remove(entryID) : false;
}
}
/** A simple class representing a pair of added and deleted VLV values. */
private static class BufferedVLVIndexValues
{
private TreeSet<ByteString> addedSortKeys;
private TreeSet<ByteString> deletedSortKeys;
void addSortKey(ByteString sortKey)
{
if (!remove(deletedSortKeys, sortKey))
{
if (addedSortKeys == null)
{
addedSortKeys = new TreeSet<ByteString>();
}
addedSortKeys.add(sortKey);
}
}
void deleteSortKey(ByteString sortKey)
{
if (!remove(addedSortKeys, sortKey))
{
if (deletedSortKeys == null)
{
deletedSortKeys = new TreeSet<ByteString>();
}
deletedSortKeys.add(sortKey);
}
}
private static boolean remove(TreeSet<ByteString> sortKeys, ByteString sortKey)
{
return sortKeys != null ? sortKeys.remove(sortKey) : false;
}
}
/**
* Construct a new empty index buffer object.
*
* @param entryContainer The database entryContainer using this
* index buffer.
*/
IndexBuffer(EntryContainer entryContainer)
{
this.entryContainer = entryContainer;
}
private BufferedVLVIndexValues createOrGetBufferedVLVIndexValues(VLVIndex vlvIndex)
{
BufferedVLVIndexValues bufferedValues = bufferedVLVIndexes.get(vlvIndex);
if (bufferedValues == null)
{
bufferedValues = new BufferedVLVIndexValues();
bufferedVLVIndexes.put(vlvIndex, bufferedValues);
}
return bufferedValues;
}
private BufferedIndexValues createOrGetBufferedIndexValues(Index index, ByteString keyBytes)
{
BufferedIndexValues values = null;
TreeMap<ByteString, BufferedIndexValues> bufferedOperations = bufferedIndexes.get(index);
if (bufferedOperations == null)
{
bufferedOperations = new TreeMap<ByteString, BufferedIndexValues>();
bufferedIndexes.put(index, bufferedOperations);
}
else
{
values = bufferedOperations.get(keyBytes);
}
if (values == null)
{
values = new BufferedIndexValues();
bufferedOperations.put(keyBytes, values);
}
return values;
}
/**
* Flush the buffered index changes until the given transaction to
* the database.
*
* @param txn a non null database transaction
* @throws StorageRuntimeException If an error occurs in the database.
* @throws DirectoryException If a Directory Server error occurs.
*/
void flush(WriteableTransaction txn) throws StorageRuntimeException, DirectoryException
{
/*
* FIXME: this seems like a surprising way to update the indexes. Why not
* store the buffered changes in a TreeMap in order to have a predictable
* iteration order?
*/
for (AttributeIndex attributeIndex : entryContainer.getAttributeIndexes())
{
for (Index index : attributeIndex.getAllIndexes())
{
flushIndex(index, txn, bufferedIndexes.remove(index));
}
}
for (VLVIndex vlvIndex : entryContainer.getVLVIndexes())
{
BufferedVLVIndexValues bufferedVLVValues = bufferedVLVIndexes.remove(vlvIndex);
if (bufferedVLVValues != null)
{
vlvIndex.updateIndex(txn, bufferedVLVValues.addedSortKeys, bufferedVLVValues.deletedSortKeys);
}
}
final Index id2children = entryContainer.getID2Children();
flushIndex(id2children, txn, bufferedIndexes.remove(id2children));
final Index id2subtree = entryContainer.getID2Subtree();
final TreeMap<ByteString, BufferedIndexValues> bufferedValues = bufferedIndexes.remove(id2subtree);
if (bufferedValues != null)
{
/*
* OPENDJ-1375: add keys in reverse order to be consistent with single
* entry processing in add/delete processing. This is necessary in order
* to avoid deadlocks.
*/
flushIndex(id2subtree, txn, bufferedValues.descendingMap());
}
}
void put(Index index, ByteString key, EntryID entryID)
{
createOrGetBufferedIndexValues(index, key).addEntryID(entryID);
}
void put(VLVIndex index, ByteString sortKey)
{
createOrGetBufferedVLVIndexValues(index).addSortKey(sortKey);
}
void remove(VLVIndex index, ByteString sortKey)
{
createOrGetBufferedVLVIndexValues(index).deleteSortKey(sortKey);
}
void remove(Index index, ByteString key)
{
createOrGetBufferedIndexValues(index, key);
}
void remove(Index index, ByteString key, EntryID entryID)
{
createOrGetBufferedIndexValues(index, key).deleteEntryID(entryID);
}
private void flushIndex(Index index, WriteableTransaction txn,
Map<ByteString, BufferedIndexValues> bufferedValues)
{
if (bufferedValues != null)
{
final Iterator<Map.Entry<ByteString, BufferedIndexValues>> it = bufferedValues.entrySet().iterator();
while (it.hasNext())
{
final Map.Entry<ByteString, BufferedIndexValues> entry = it.next();
final ByteString key = entry.getKey();
final BufferedIndexValues values = entry.getValue();
index.update(txn, key, values.deletedEntryIDs, values.addedEntryIDs);
it.remove();
}
}
}
}