/* * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS HEADER. * * Copyright (c) 1997-2010 Oracle and/or its affiliates. All rights reserved. * * The contents of this file are subject to the terms of either the GNU * General Public License Version 2 only ("GPL") or the Common Development * and Distribution License("CDDL") (collectively, the "License"). You * may not use this file except in compliance with the License. You can * obtain a copy of the License at * https://glassfish.dev.java.net/public/CDDL+GPL_1_1.html * or packager/legal/LICENSE.txt. See the License for the specific * language governing permissions and limitations under the License. * * When distributing the software, include this License Header Notice in each * file and include the License file at packager/legal/LICENSE.txt. * * GPL Classpath Exception: * Oracle designates this particular file as subject to the "Classpath" * exception as provided by Oracle in the GPL Version 2 section of the License * file that accompanied this code. * * Modifications: * If applicable, add the following below the License Header, with the fields * enclosed by brackets [] replaced by your own identifying information: * "Portions Copyright [year] [name of copyright owner]" * * Contributor(s): * If you wish your version of this file to be governed by only the CDDL or * only the GPL Version 2, indicate your decision by adding "[Contributor] * elects to include this software in this distribution under the [CDDL or GPL * Version 2] license." If you don't indicate a single choice of license, a * recipient has the option to distribute your version of this file under * either the CDDL, the GPL Version 2 or to extend the choice of license to * its licensees as provided above. However, if you add GPL Version 2 code * and therefore, elected the GPL Version 2 license, then the option applies * only if the new code is made subject to such option by the copyright * holder. */ package com.sun.appserv.util.cache; import java.text.MessageFormat; import java.util.Properties; import java.util.Map; import java.util.ResourceBundle; /** * MultiLruCache -- in-memory bounded LRU cache with multiple LRU lists * Underlying Hashtable is made into logical segments, with each segment * having its own LRU list. */ public class MultiLruCache extends BaseCache { /* an array of LRU lists; each element in this array is actually * LruCacheItem[2] with LRU list (lru[0] is head and lru[1] the tail */ public static final int LRU_HEAD = 0; public static final int LRU_TAIL = 1; public static final int DEFAULT_HASHTABLE_SEGMENT_SIZE = 4096; int segmentSize; LruCacheItem[][] lists; protected int[] listsLength; int trimCount; int trimIndex; Object trimIndexLk = new Object(); /** * initialize the LRU cache * @param maxCapacity maximum number of entries this cache may hold */ public void init(int maxCapacity, Properties props) throws Exception { super.init(maxCapacity, props); segmentSize = DEFAULT_HASHTABLE_SEGMENT_SIZE; if (props != null) { String prop = props.getProperty("MultiLRUSegmentSize"); if (prop != null) { try { segmentSize = Integer.parseInt(prop); } catch (NumberFormatException nfe) {} } } // create the array of LRU lists int segments = ((maxBuckets / segmentSize) + (((maxBuckets % segmentSize) != 0) ? 1 : 0)); lists = new LruCacheItem[segments][2]; listsLength = new int[lists.length]; for (int i = 0; i < lists.length; i++) { lists[i][LRU_HEAD] = null; lists[i][LRU_TAIL] = null; listsLength[i] = 0; } } /** * get the LRU list associated with the index * @param index into the BaseCache hashtable * @return the LRU list to be used */ private LruCacheItem[] getLRUList(int index) { int segment = (index/segmentSize); return lists[segment]; } /** * create new item * @param hashCode for the entry * @param key Object key * @param value Object value * @param size size in bytes of the item * * subclasses may override to provide their own CacheItem extensions * e.g. one that permits persistence. */ protected CacheItem createItem(int hashCode, Object key, Object value, int size) { return new LruCacheItem(hashCode, key, value, size); } /** * remove an lru item from one of the LRU lists * @param the LRU segment index to trim * @return the item that was successfully trimmed */ protected CacheItem trimLru(int segment) { LruCacheItem[] list = lists[segment]; LruCacheItem l = null; l = list[LRU_TAIL]; list[LRU_TAIL] = l.lPrev; list[LRU_TAIL].lNext = null; l.lPrev = null; listsLength[segment]--; l.isTrimmed = true; trimCount++; return l; } /** * this item is just added to the cache * @param item CacheItem that was created * @return a overflow item; may be null * * Cache bucket is already synchronized by the caller */ protected CacheItem itemAdded(CacheItem item) { CacheItem overflow = null; LruCacheItem lc = (LruCacheItem) item; int index = getIndex(item.hashCode()); int segment = (index/segmentSize); LruCacheItem[] list = lists[segment]; // update the LRU synchronized (list) { if (list[LRU_HEAD] != null) { list[LRU_HEAD].lPrev = lc; lc.lNext = list[LRU_HEAD]; } else list[LRU_TAIL] = lc; list[LRU_HEAD] = lc; listsLength[segment]++; if (isThresholdReached()) { overflow = trimLru(trimIndex); // go round robin for the next trim incrementTrimIndex(); } } return overflow; } /** * this item is accessed * @param item CacheItem accessed * * Cache bucket is already synchronized by the caller */ protected void itemAccessed(CacheItem item) { int index = getIndex(item.hashCode()); int segment = (index/segmentSize); LruCacheItem[] list = lists[segment]; LruCacheItem lc = (LruCacheItem) item; // update the LRU list synchronized (list) { LruCacheItem prev = lc.lPrev; LruCacheItem next = lc.lNext; if (prev != null) { // put the item at the head of LRU list lc.lPrev = null; lc.lNext = list[LRU_HEAD]; list[LRU_HEAD].lPrev = lc; list[LRU_HEAD] = lc; // patch up the previous neighbors prev.lNext = next; if (next != null) next.lPrev = prev; else list[LRU_TAIL] = prev; } } } /** * item value has been refreshed * @param item CacheItem that was refreshed * @param oldSize size of the previous value that was refreshed * Cache bucket is already synchronized by the caller */ protected void itemRefreshed(CacheItem item, int oldSize) { itemAccessed(item); } /** * item value has been removed from the cache * @param item CacheItem that was just removed * * Cache bucket is already synchronized by the caller */ protected void itemRemoved(CacheItem item) { LruCacheItem l = (LruCacheItem) item; int index = getIndex(item.hashCode()); int segment = (index/segmentSize); LruCacheItem[] list = lists[segment]; // remove the item from the LRU list synchronized (list) { // if the item is already trimmed from the LRU list, nothing to do. if (l.isTrimmed) return; LruCacheItem prev = l.lPrev; LruCacheItem next = l.lNext; // patch up the neighbors and make sure head/tail are correct if (prev != null) prev.lNext = next; else list[LRU_HEAD] = next; if (next != null) next.lPrev = prev; else list[LRU_TAIL] = prev; listsLength[segment]--; } } /** * cache has reached threshold so trim its size. subclasses are expected * to provide a robust cache replacement algorithm. */ protected void handleOverflow() { LruCacheItem l = null; } int getListsLength() { return lists.length; } protected void incrementTrimIndex() { synchronized (trimIndexLk) { trimIndex = (trimIndex + 1) % lists.length; } } /** * get generic stats from subclasses */ /** * get the desired statistic counter * @param key to corresponding stat * @return an Object corresponding to the stat * See also: Constant.java for the key */ public Object getStatByName(String key) { Object stat = super.getStatByName(key); if (stat == null && key != null) { if (key.equals(Constants.STAT_MULTILRUCACHE_SEGMENT_SIZE)) stat = Integer.valueOf(segmentSize); else if (key.equals(Constants.STAT_MULTILRUCACHE_TRIM_COUNT)) stat = Integer.valueOf(trimCount); else if (key.equals(Constants.STAT_MULTILRUCACHE_SEGMENT_LIST_LENGTH)) { stat = new Integer[lists.length]; for (int i = 0; i < lists.length; i++) { ((Integer[])stat)[i] = Integer.valueOf(listsLength[i]); } } } return stat; } /** * get the stats snapshot * @return a Map of stats * See also: Constant.java for the keys */ public Map getStats() { Map stats = super.getStats(); stats.put(Constants.STAT_MULTILRUCACHE_SEGMENT_SIZE, Integer.valueOf(segmentSize)); for (int i = 0; i < lists.length; i++) { stats.put(Constants.STAT_MULTILRUCACHE_SEGMENT_LIST_LENGTH + "[" + i + "]:", Integer.valueOf(listsLength[i])); } stats.put(Constants.STAT_MULTILRUCACHE_TRIM_COUNT, Integer.valueOf(trimCount)); return stats; } /** default CacheItem class implementation ***/ static class LruCacheItem extends BaseCache.CacheItem { // double linked LRU list LruCacheItem lNext; LruCacheItem lPrev; boolean isTrimmed; LruCacheItem(int hashCode, Object key, Object value, int size) { super(hashCode, key, value, size); } } }