Skip to content
项目
群组
代码片段
帮助
正在加载...
帮助
为 GitLab 提交贡献
登录/注册
切换导航
H
h2database
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
分枝图
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
计划
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
分枝图
统计图
创建新议题
作业
提交
议题看板
打开侧边栏
Administrator
h2database
Commits
c3170f7d
提交
c3170f7d
authored
10月 16, 2012
作者:
Thomas Mueller
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
LIRS cache: concurrent and concurrent with long key
上级
9dbf9138
显示空白字符变更
内嵌
并排
正在显示
5 个修改的文件
包含
1093 行增加
和
44 行删除
+1093
-44
TestCacheConcurrentLIRS.java
h2/src/test/org/h2/test/store/TestCacheConcurrentLIRS.java
+18
-8
TestCacheLIRS.java
h2/src/test/org/h2/test/store/TestCacheLIRS.java
+12
-0
CacheConcurrentLIRS.java
h2/src/tools/org/h2/dev/store/btree/CacheConcurrentLIRS.java
+111
-26
CacheLIRS.java
h2/src/tools/org/h2/dev/store/btree/CacheLIRS.java
+14
-10
CacheLongKeyLIRS.java
h2/src/tools/org/h2/dev/store/btree/CacheLongKeyLIRS.java
+938
-0
没有找到文件。
h2/src/test/org/h2/test/store/TestCacheConcurrentLIRS.java
浏览文件 @
c3170f7d
...
...
@@ -8,7 +8,7 @@ package org.h2.test.store;
import
java.util.Random
;
import
java.util.concurrent.CountDownLatch
;
import
java.util.concurrent.atomic.AtomicBoolean
;
import
org.h2.dev.store.btree.Cache
Concurrent
LIRS
;
import
org.h2.dev.store.btree.Cache
LongKey
LIRS
;
import
org.h2.test.TestBase
;
import
org.h2.util.Task
;
...
...
@@ -31,12 +31,25 @@ public class TestCacheConcurrentLIRS extends TestBase {
}
private
static
void
testConcurrent
()
{
final
Cache
ConcurrentLIRS
<
Integer
,
Integer
>
test
=
CacheConcurrent
LIRS
.
newInstance
(
100
,
1
);
final
Cache
LongKeyLIRS
<
Integer
>
test
=
CacheLongKey
LIRS
.
newInstance
(
100
,
1
);
int
threadCount
=
8
;
final
CountDownLatch
wait
=
new
CountDownLatch
(
1
);
final
AtomicBoolean
stopped
=
new
AtomicBoolean
();
Task
[]
tasks
=
new
Task
[
threadCount
];
final
int
[]
getCounts
=
new
int
[
threadCount
];
final
int
offset
=
1000000
;
for
(
int
i
=
0
;
i
<
100
;
i
++)
{
test
.
put
(
offset
+
i
,
i
);
}
final
int
[]
keys
=
new
int
[
1000
];
Random
random
=
new
Random
(
1
);
for
(
int
i
=
0
;
i
<
keys
.
length
;
i
++)
{
int
key
;
do
{
key
=
(
int
)
Math
.
abs
(
random
.
nextGaussian
()
*
50
);
}
while
(
key
>
100
);
keys
[
i
]
=
key
;
}
for
(
int
i
=
0
;
i
<
threadCount
;
i
++)
{
final
int
x
=
i
;
Task
t
=
new
Task
()
{
...
...
@@ -46,13 +59,10 @@ public class TestCacheConcurrentLIRS extends TestBase {
wait
.
await
();
int
i
=
0
;
for
(;
!
stopped
.
get
();
i
++)
{
int
key
;
do
{
key
=
(
int
)
Math
.
abs
(
random
.
nextGaussian
()
*
50
);
}
while
(
key
>
100
);
test
.
get
(
key
);
int
key
=
keys
[
random
.
nextInt
(
keys
.
length
)];
test
.
get
(
offset
+
key
);
if
((
i
&
127
)
==
0
)
{
test
.
put
(
random
.
nextInt
(
100
),
random
.
nextInt
());
test
.
put
(
offset
+
random
.
nextInt
(
100
),
random
.
nextInt
());
}
}
getCounts
[
x
]
=
i
;
...
...
h2/src/test/org/h2/test/store/TestCacheLIRS.java
浏览文件 @
c3170f7d
...
...
@@ -82,6 +82,18 @@ public class TestCacheLIRS extends TestBase {
verifyMapSize
(
769
,
2048
);
CacheLIRS
<
Integer
,
Integer
>
test
;
test
=
CacheLIRS
.
newInstance
(
3
,
10
);
test
.
put
(
0
,
0
,
9
);
test
.
put
(
1
,
10
,
9
);
test
.
put
(
2
,
20
,
9
);
test
.
put
(
3
,
30
,
9
);
test
.
put
(
4
,
40
,
9
);
test
=
CacheLIRS
.
newInstance
(
1
,
1
);
test
.
put
(
1
,
10
);
test
.
put
(
0
,
0
);
test
.
get
(
0
);
test
=
CacheLIRS
.
newInstance
(
1000
,
1
);
for
(
int
j
=
0
;
j
<
2000
;
j
++)
{
test
.
put
(
j
,
j
);
...
...
h2/src/tools/org/h2/dev/store/btree/CacheConcurrentLIRS.java
浏览文件 @
c3170f7d
...
...
@@ -13,6 +13,7 @@ import java.util.HashSet;
import
java.util.List
;
import
java.util.Map
;
import
java.util.Set
;
import
java.util.concurrent.ConcurrentMap
;
/**
* A scan resistant cache. It is meant to cache objects that are relatively
...
...
@@ -32,16 +33,18 @@ import java.util.Set;
* prevent unbound memory usage. The maximum size of this queue is at most the
* size of the rest of the stack. About 6.25% of the mapped entries are cold.
* <p>
* Internally, the cache is split into 16 segments, and each segment is an
* individual LIRS cache. Accessed entries are only moved to the top of the
* stack if at least 20 other entries have been moved to the front. Write access
* and moving entries to the top of the stack is synchronized per segment.
* Internally, the cache is split into a number of segments, and each segment is
* an individual LIRS cache.
* <p>
* Accessed entries are only moved to the top of the stack if at least a number
* of other entries have been moved to the front. Write access and moving
* entries to the top of the stack is synchronized per segment.
*
* @author Thomas Mueller
* @param <K> the key type
* @param <V> the value type
*/
public
class
CacheConcurrentLIRS
<
K
,
V
>
extends
AbstractMap
<
K
,
V
>
implements
Map
<
K
,
V
>
{
public
class
CacheConcurrentLIRS
<
K
,
V
>
extends
AbstractMap
<
K
,
V
>
implements
Concurrent
Map
<
K
,
V
>
{
/**
* The maximum memory this cache should use.
...
...
@@ -55,24 +58,29 @@ public class CacheConcurrentLIRS<K, V> extends AbstractMap<K, V> implements Map<
private
Segment
<
K
,
V
>[]
segments
;
private
int
segmentCount
;
private
int
segmentShift
;
private
int
segmentMask
;
private
final
int
stackMoveDistance
;
private
CacheConcurrentLIRS
(
long
maxMemory
,
int
averageMemory
)
{
this
.
maxMemory
=
maxMemory
;
this
.
averageMemory
=
averageMemory
;
private
CacheConcurrentLIRS
(
long
maxMemory
,
int
averageMemory
,
int
segmentCount
,
int
stackMoveDistance
)
{
setMaxMemory
(
maxMemory
);
setAverageMemory
(
averageMemory
);
if
(
Integer
.
bitCount
(
segmentCount
)
!=
1
)
{
throw
new
IllegalArgumentException
(
"The segment count must be a power of 2, is "
+
segmentCount
);
}
this
.
segmentCount
=
segmentCount
;
this
.
stackMoveDistance
=
stackMoveDistance
;
clear
();
}
@SuppressWarnings
(
"unchecked"
)
public
void
clear
()
{
// must be a power of 2
int
count
=
16
;
segmentMask
=
count
-
1
;
segments
=
new
Segment
[
count
];
for
(
int
i
=
0
;
i
<
count
;
i
++)
{
segmentMask
=
segmentCount
-
1
;
segments
=
new
Segment
[
segmentCount
];
for
(
int
i
=
0
;
i
<
segmentCount
;
i
++)
{
segments
[
i
]
=
new
Segment
<
K
,
V
>(
1
+
maxMemory
/
count
,
averageMemory
);
1
+
maxMemory
/
segmentCount
,
averageMemory
,
stackMoveDistance
);
}
segmentShift
=
Integer
.
numberOfTrailingZeros
(
segments
[
0
].
sizeMapArray
());
}
...
...
@@ -132,6 +140,53 @@ public class CacheConcurrentLIRS<K, V> extends AbstractMap<K, V> implements Map<
return
put
(
key
,
value
,
averageMemory
);
}
public
V
putIfAbsent
(
K
key
,
V
value
)
{
int
todo
;
if
(
containsKey
(
key
))
{
return
get
(
key
);
}
return
put
(
key
,
value
);
}
public
boolean
remove
(
Object
key
,
Object
value
)
{
int
todo
;
Entry
<
K
,
V
>
e
=
find
(
key
);
if
(
e
!=
null
)
{
V
x
=
e
.
value
;
if
(
x
!=
null
&&
x
.
equals
(
value
))
{
remove
(
key
);
return
true
;
}
}
return
false
;
}
public
boolean
replace
(
K
key
,
V
oldValue
,
V
newValue
)
{
int
todo
;
Entry
<
K
,
V
>
e
=
find
(
key
);
if
(
e
!=
null
)
{
V
x
=
e
.
value
;
if
(
x
!=
null
&&
x
.
equals
(
oldValue
))
{
put
(
key
,
newValue
);
return
true
;
}
}
return
false
;
}
public
V
replace
(
K
key
,
V
value
)
{
int
todo
;
Entry
<
K
,
V
>
e
=
find
(
key
);
if
(
e
!=
null
)
{
V
x
=
e
.
value
;
if
(
x
!=
null
)
{
put
(
key
,
value
);
return
x
;
}
}
return
null
;
}
/**
* Remove an entry. Both resident and non-resident entries can be removed.
*
...
...
@@ -194,6 +249,24 @@ public class CacheConcurrentLIRS<K, V> extends AbstractMap<K, V> implements Map<
return
x
;
}
/**
* Set the average memory used per entry. It is used to calculate the length
* of the internal array.
*
* @param averageMemory the average memory used (1 or larger)
*/
public
void
setAverageMemory
(
int
averageMemory
)
{
if
(
averageMemory
<=
0
)
{
throw
new
IllegalArgumentException
(
"Average memory must be larger than 0"
);
}
this
.
averageMemory
=
averageMemory
;
if
(
segments
!=
null
)
{
for
(
Segment
<
K
,
V
>
s
:
segments
)
{
s
.
setAverageMemory
(
averageMemory
);
}
}
}
/**
* Set the maximum memory this cache should use. This will not immediately
* cause entries to get removed however; it will only change the limit. To
...
...
@@ -206,10 +279,12 @@ public class CacheConcurrentLIRS<K, V> extends AbstractMap<K, V> implements Map<
throw
new
IllegalArgumentException
(
"Max memory must be larger than 0"
);
}
this
.
maxMemory
=
maxMemory
;
if
(
segments
!=
null
)
{
for
(
Segment
<
K
,
V
>
s
:
segments
)
{
s
.
setMaxMemory
(
1
+
maxMemory
/
segments
.
length
);
}
}
}
/**
* Get the average memory used per entry.
...
...
@@ -236,10 +311,14 @@ public class CacheConcurrentLIRS<K, V> extends AbstractMap<K, V> implements Map<
*
* @param maxMemory the maximum memory to use (1 or larger)
* @param averageMemory the average memory (1 or larger)
* @param segmentCount the number of cache segments (must be a power of 2)
* @param stackMoveDistance how many other item are to be moved to the top
* of the stack before the current item is moved
* @return the cache
*/
public
static
<
K
,
V
>
CacheConcurrentLIRS
<
K
,
V
>
newInstance
(
int
maxMemory
,
int
averageMemory
)
{
return
new
CacheConcurrentLIRS
<
K
,
V
>(
maxMemory
,
averageMemory
);
public
static
<
K
,
V
>
CacheConcurrentLIRS
<
K
,
V
>
newInstance
(
int
maxMemory
,
int
averageMemory
,
int
segmentCount
,
int
stackMoveDistance
)
{
return
new
CacheConcurrentLIRS
<
K
,
V
>(
maxMemory
,
averageMemory
,
segmentCount
,
stackMoveDistance
);
}
/**
...
...
@@ -321,7 +400,7 @@ public class CacheConcurrentLIRS<K, V> extends AbstractMap<K, V> implements Map<
}
/**
* Get the list of keys. This method allows
a view of
the internal state of
* Get the list of keys. This method allows
to read
the internal state of
* the cache.
*
* @param cold if true, only keys for the cold entries are returned
...
...
@@ -344,12 +423,11 @@ public class CacheConcurrentLIRS<K, V> extends AbstractMap<K, V> implements Map<
*/
static
class
Segment
<
K
,
V
>
{
/**
* How many other item are to be moved to the top of the stack before
* the current item is moved.
*/
private
int
stackMoveDistance
=
20
;
private
final
int
stackMoveDistance
;
/**
* The maximum memory this cache should use.
...
...
@@ -423,10 +501,13 @@ public class CacheConcurrentLIRS<K, V> extends AbstractMap<K, V> implements Map<
*
* @param maxMemory the maximum memory to use
* @param averageMemory the average memory usage of an object
* @param stackMoveDistance the number of other entries to be moved to
* the top of the stack before moving an entry to the top
*/
Segment
(
long
maxMemory
,
int
averageMemory
)
{
Segment
(
long
maxMemory
,
int
averageMemory
,
int
stackMoveDistance
)
{
setMaxMemory
(
maxMemory
);
setAverageMemory
(
averageMemory
);
this
.
stackMoveDistance
=
stackMoveDistance
;
clear
();
}
...
...
@@ -621,7 +702,8 @@ public class CacheConcurrentLIRS<K, V> extends AbstractMap<K, V> implements Map<
/**
* Evict cold entries (resident and non-resident) until the memory limit is
* reached.
* reached. The new entry is added as a cold entry, except if it is the only
* entry.
*
* @param newCold a new cold entry
*/
...
...
@@ -632,10 +714,13 @@ public class CacheConcurrentLIRS<K, V> extends AbstractMap<K, V> implements Map<
while
((
queueSize
<<
5
)
<
mapSize
)
{
convertOldestHotToCold
();
}
if
(
stackSize
>
0
)
{
// the new cold entry is at the top of the queue
addToQueue
(
queue
,
newCold
);
}
// the oldest resident cold entries become non-resident
while
(
usedMemory
>
maxMemory
)
{
// but at least one cold entry (the new one) must stay
while
(
usedMemory
>
maxMemory
&&
queueSize
>
1
)
{
Entry
<
K
,
V
>
e
=
queue
.
queuePrev
;
usedMemory
-=
e
.
memory
;
removeFromQueue
(
e
);
...
...
h2/src/tools/org/h2/dev/store/btree/CacheLIRS.java
浏览文件 @
c3170f7d
...
...
@@ -18,7 +18,7 @@ import java.util.Set;
* A scan resistant cache. It is meant to cache objects that are relatively
* costly to acquire, for example file content.
* <p>
* This implementation is not multi-threading sa
v
e. Null keys or null values are
* This implementation is not multi-threading sa
f
e. Null keys or null values are
* not allowed. The map fill factor is at most 75%.
* <p>
* Each entry is assigned a distinct memory size, and the cache will try to use
...
...
@@ -201,7 +201,7 @@ public class CacheLIRS<K, V> extends AbstractMap<K, V> implements Map<K, V> {
return
null
;
}
else
if
(
e
.
isHot
())
{
if
(
e
!=
stack
.
stackNext
)
{
// move a hot entr
ies
to the top of the stack
// move a hot entr
y
to the top of the stack
// unless it is already there
boolean
wasEnd
=
e
==
stack
.
stackPrev
;
removeFromStack
(
e
);
...
...
@@ -343,21 +343,25 @@ public class CacheLIRS<K, V> extends AbstractMap<K, V> implements Map<K, V> {
/**
* Evict cold entries (resident and non-resident) until the memory limit is
* reached.
* reached. The new entry is added as a cold entry, except if it is the only
* entry.
*
* @param new
Cold a new cold
entry
* @param new
Entry a new
entry
*/
private
void
evict
(
Entry
<
K
,
V
>
new
Cold
)
{
private
void
evict
(
Entry
<
K
,
V
>
new
Entry
)
{
// ensure there are not too many hot entries:
// left shift of 5 is multiplication by 32, that means if there are less
// than 1/32 (3.125%) cold entries, a new hot entry needs to become cold
while
((
queueSize
<<
5
)
<
mapSize
)
{
convertOldestHotToCold
();
}
if
(
stackSize
>
0
)
{
// the new cold entry is at the top of the queue
addToQueue
(
queue
,
newCold
);
addToQueue
(
queue
,
newEntry
);
}
// the oldest resident cold entries become non-resident
while
(
usedMemory
>
maxMemory
)
{
// but at least one cold entry (the new one) must stay
while
(
usedMemory
>
maxMemory
&&
queueSize
>
1
)
{
Entry
<
K
,
V
>
e
=
queue
.
queuePrev
;
usedMemory
-=
e
.
memory
;
removeFromQueue
(
e
);
...
...
@@ -459,7 +463,7 @@ public class CacheLIRS<K, V> extends AbstractMap<K, V> implements Map<K, V> {
}
/**
* Get the list of keys. This method allows to
view
the internal state of
* Get the list of keys. This method allows to
read
the internal state of
* the cache.
*
* @param cold if true, only keys for the cold entries are returned
...
...
@@ -492,7 +496,7 @@ public class CacheLIRS<K, V> extends AbstractMap<K, V> implements Map<K, V> {
/**
* Check whether there is a resident entry for the given key. This method
* does not adjust
s
the internal state of the cache.
* does not adjust the internal state of the cache.
*
* @param key the key (may not be null)
* @return true if there is a resident entry
...
...
h2/src/tools/org/h2/dev/store/btree/CacheLongKeyLIRS.java
0 → 100644
浏览文件 @
c3170f7d
/*
* Copyright 2004-2011 H2 Group. Multiple-Licensed under the H2 License,
* Version 1.0, and under the Eclipse Public License, Version 1.0
* (http://h2database.com/html/license.html).
* Initial Developer: H2 Group
*/
package
org
.
h2
.
dev
.
store
.
btree
;
import
java.util.ArrayList
;
import
java.util.HashMap
;
import
java.util.HashSet
;
import
java.util.List
;
import
java.util.Map
;
import
java.util.Set
;
/**
* A scan resistant cache that uses keys of type long. It is meant to cache
* objects that are relatively costly to acquire, for example file content.
* <p>
* This implementation is multi-threading safe and supports concurrent access.
* Null keys or null values are not allowed. The map fill factor is at most 75%.
* <p>
* Each entry is assigned a distinct memory size, and the cache will try to use
* at most the specified amount of memory. The memory unit is not relevant,
* however it is suggested to use bytes as the unit.
* <p>
* This class implements an approximation of the the LIRS replacement algorithm
* invented by Xiaodong Zhang and Song Jiang as described in
* http://www.cse.ohio-state.edu/~zhang/lirs-sigmetrics-02.html with a few
* smaller changes: An additional queue for non-resident entries is used, to
* prevent unbound memory usage. The maximum size of this queue is at most the
* size of the rest of the stack. About 6.25% of the mapped entries are cold.
* <p>
* Internally, the cache is split into 16 segments, and each segment is an
* individual LIRS cache. Accessed entries are only moved to the top of the
* stack if at least 20 other entries have been moved to the front. Write access
* and moving entries to the top of the stack is synchronized per segment.
*
* @author Thomas Mueller
* @param <V> the value type
*/
public
class
CacheLongKeyLIRS
<
V
>
{
/**
* The maximum memory this cache should use.
*/
private
long
maxMemory
;
/**
* The average memory used by one entry.
*/
private
int
averageMemory
;
private
Segment
<
V
>[]
segments
;
private
int
segmentShift
;
private
int
segmentMask
;
private
CacheLongKeyLIRS
(
long
maxMemory
,
int
averageMemory
)
{
this
.
maxMemory
=
maxMemory
;
this
.
averageMemory
=
averageMemory
;
clear
();
}
@SuppressWarnings
(
"unchecked"
)
public
void
clear
()
{
// must be a power of 2
int
count
=
16
;
segmentMask
=
count
-
1
;
segments
=
new
Segment
[
count
];
long
max
=
1
+
maxMemory
/
segments
.
length
;
for
(
int
i
=
0
;
i
<
count
;
i
++)
{
segments
[
i
]
=
new
Segment
<
V
>(
max
,
averageMemory
);
}
segmentShift
=
Integer
.
numberOfTrailingZeros
(
segments
[
0
].
sizeMapArray
());
}
private
Entry
<
V
>
find
(
long
key
)
{
int
hash
=
getHash
(
key
);
return
getSegment
(
hash
).
find
(
key
,
hash
);
}
/**
* Check whether there is a resident entry for the given key. This method
* does not adjust the internal state of the cache.
*
* @param key the key (may not be null)
* @return true if there is a resident entry
*/
public
boolean
containsKey
(
long
key
)
{
int
hash
=
getHash
(
key
);
return
getSegment
(
hash
).
containsKey
(
key
,
hash
);
}
/**
* Get the value for the given key if the entry is cached. This method does
* not modify the internal state.
*
* @param key the key (may not be null)
* @return the value, or null if there is no resident entry
*/
public
V
peek
(
long
key
)
{
Entry
<
V
>
e
=
find
(
key
);
return
e
==
null
?
null
:
e
.
value
;
}
/**
* Add an entry to the cache. The entry may or may not exist in the cache
* yet. This method will usually mark unknown entries as cold and known
* entries as hot.
*
* @param key the key (may not be null)
* @param value the value (may not be null)
* @param memory the memory used for the given entry
* @return the old value, or null if there is no resident entry
*/
public
V
put
(
long
key
,
V
value
,
int
memory
)
{
int
hash
=
getHash
(
key
);
return
getSegment
(
hash
).
put
(
key
,
value
,
hash
,
memory
);
}
/**
* Add an entry to the cache using the average memory size.
*
* @param key the key (may not be null)
* @param value the value (may not be null)
* @return the old value, or null if there is no resident entry
*/
public
V
put
(
long
key
,
V
value
)
{
return
put
(
key
,
value
,
averageMemory
);
}
/**
* Remove an entry. Both resident and non-resident entries can be removed.
*
* @param key the key (may not be null)
* @return the old value, or null if there is no resident entry
*/
public
synchronized
V
remove
(
long
key
)
{
int
hash
=
getHash
(
key
);
return
getSegment
(
hash
).
remove
(
key
,
hash
);
}
/**
* Get the memory used for the given key.
*
* @param key the key (may not be null)
* @return the memory, or 0 if there is no resident entry
*/
public
int
getMemory
(
long
key
)
{
int
hash
=
getHash
(
key
);
return
getSegment
(
hash
).
getMemory
(
key
,
hash
);
}
/**
* Get the value for the given key if the entry is cached. This method
* adjusts the internal state of the cache, to ensure commonly used entries
* stay in the cache.
*
* @param key the key (may not be null)
* @return the value, or null if there is no resident entry
*/
public
V
get
(
long
key
)
{
int
hash
=
getHash
(
key
);
return
getSegment
(
hash
).
get
(
key
,
hash
);
}
private
Segment
<
V
>
getSegment
(
int
hash
)
{
int
segmentIndex
=
(
hash
>>>
segmentShift
)
&
segmentMask
;
return
segments
[
segmentIndex
];
}
static
int
getHash
(
long
key
)
{
int
hash
=
(
int
)
((
key
>>>
32
)
^
key
);
// Doug Lea's supplemental secondaryHash function (inlined)
// to protect against hash codes that don't differ in low order bits
hash
^=
(
hash
>>>
20
)
^
(
hash
>>>
12
);
hash
^=
(
hash
>>>
7
)
^
(
hash
>>>
4
);
return
hash
;
}
/**
* Get the currently used memory.
*
* @return the used memory
*/
public
long
getUsedMemory
()
{
long
x
=
0
;
for
(
Segment
<
V
>
s
:
segments
)
{
x
+=
s
.
getUsedMemory
();
}
return
x
;
}
/**
* Set the maximum memory this cache should use. This will not immediately
* cause entries to get removed however; it will only change the limit. To
* resize the internal array, call the clear method.
*
* @param maxMemory the maximum size (1 or larger)
*/
public
void
setMaxMemory
(
long
maxMemory
)
{
if
(
maxMemory
<=
0
)
{
throw
new
IllegalArgumentException
(
"Max memory must be larger than 0"
);
}
this
.
maxMemory
=
maxMemory
;
long
max
=
1
+
maxMemory
/
segments
.
length
;
for
(
Segment
<
V
>
s
:
segments
)
{
s
.
setMaxMemory
(
max
);
}
}
/**
* Set the average memory used per entry. It is used to calculate the length
* of the internal array.
*
* @param averageMemory the average memory used (1 or larger)
*/
public
void
setAverageMemory
(
int
averageMemory
)
{
if
(
averageMemory
<=
0
)
{
throw
new
IllegalArgumentException
(
"Average memory must be larger than 0"
);
}
this
.
averageMemory
=
averageMemory
;
for
(
Segment
<
V
>
s
:
segments
)
{
s
.
setAverageMemory
(
averageMemory
);
}
}
/**
* Get the average memory used per entry.
*
* @return the average memory
*/
public
int
getAverageMemory
()
{
return
averageMemory
;
}
/**
* Get the maximum memory to use.
*
* @return the maximum memory
*/
public
long
getMaxMemory
()
{
return
maxMemory
;
}
/**
* Create a new cache with the given memory size. To just limit the number
* of entries, use the required number as the maximum memory, and an average
* size of 1.
*
* @param maxMemory the maximum memory to use (1 or larger)
* @param averageMemory the average memory (1 or larger)
* @return the cache
*/
public
static
<
V
>
CacheLongKeyLIRS
<
V
>
newInstance
(
int
maxMemory
,
int
averageMemory
)
{
return
new
CacheLongKeyLIRS
<
V
>(
maxMemory
,
averageMemory
);
}
/**
* Get the number of non-resident entries in the cache.
*
* @return the number of non-resident entries
*/
public
int
sizeNonResident
()
{
int
x
=
0
;
for
(
Segment
<
V
>
s
:
segments
)
{
x
+=
s
.
sizeNonResident
();
}
return
x
;
}
/**
* Get the length of the internal map array.
*
* @return the size of the array
*/
public
int
sizeMapArray
()
{
int
x
=
0
;
for
(
Segment
<
V
>
s
:
segments
)
{
x
+=
s
.
sizeMapArray
();
}
return
x
;
}
/**
* Get the entry set for all resident entries.
*
* @return the entry set
*/
public
Set
<
Long
>
keySet
()
{
HashSet
<
Long
>
set
=
new
HashSet
<
Long
>();
for
(
Segment
<
V
>
s
:
segments
)
{
set
.
addAll
(
s
.
keySet
());
}
return
set
;
}
/**
* Get the values for all resident entries.
*
* @return the entry set
*/
public
List
<
V
>
values
()
{
ArrayList
<
V
>
list
=
new
ArrayList
<
V
>();
for
(
long
k
:
keySet
())
{
V
value
=
find
(
k
).
value
;
if
(
value
!=
null
)
{
list
.
add
(
value
);
}
}
return
list
;
}
public
boolean
isEmpty
()
{
return
size
()
==
0
;
}
/**
* Get the entry set for all resident entries.
*
* @return the entry set
*/
public
Set
<
Map
.
Entry
<
Long
,
V
>>
entrySet
()
{
return
getMap
().
entrySet
();
}
public
boolean
containsValue
(
Object
value
)
{
return
getMap
().
containsValue
(
value
);
}
public
Map
<
Long
,
V
>
getMap
()
{
HashMap
<
Long
,
V
>
map
=
new
HashMap
<
Long
,
V
>();
for
(
long
k
:
keySet
())
{
V
x
=
find
(
k
).
value
;
if
(
x
!=
null
)
{
map
.
put
(
k
,
x
);
}
}
return
map
;
}
/**
* Get the number of hot entries in the cache.
*
* @return the number of hot entries
*/
public
int
sizeHot
()
{
int
x
=
0
;
for
(
Segment
<
V
>
s
:
segments
)
{
x
+=
s
.
sizeHot
();
}
return
x
;
}
/**
* Get the number of resident entries.
*
* @return the number of entries
*/
public
int
size
()
{
int
x
=
0
;
for
(
Segment
<
V
>
s
:
segments
)
{
x
+=
s
.
size
();
}
return
x
;
}
public
void
putAll
(
Map
<
Long
,
?
extends
V
>
m
)
{
for
(
Map
.
Entry
<
Long
,
?
extends
V
>
e
:
m
.
entrySet
())
{
// copy only non-null entries
put
(
e
.
getKey
(),
e
.
getValue
());
}
}
/**
* Get the list of keys. This method allows to read the internal state of
* the cache.
*
* @param cold if true, only keys for the cold entries are returned
* @param nonResident true for non-resident entries
* @return the key list
*/
public
synchronized
List
<
Long
>
keys
(
boolean
cold
,
boolean
nonResident
)
{
ArrayList
<
Long
>
keys
=
new
ArrayList
<
Long
>();
for
(
Segment
<
V
>
s
:
segments
)
{
keys
.
addAll
(
s
.
keys
(
cold
,
nonResident
));
}
return
keys
;
}
/**
* A cache segment
*
* @param <V> the value type
*/
static
class
Segment
<
V
>
{
/**
* How many other item are to be moved to the top of the stack before
* the current item is moved.
*/
private
int
stackMoveDistance
=
20
;
/**
* The maximum memory this cache should use.
*/
private
long
maxMemory
;
/**
* The average memory used by one entry.
*/
private
int
averageMemory
;
/**
* The currently used memory.
*/
private
long
usedMemory
;
/**
* The number of (hot, cold, and non-resident) entries in the map.
*/
private
int
mapSize
;
/**
* The bit mask that is applied to the key hash code to get the index in the
* map array. The mask is the length of the array minus one.
*/
private
int
mask
;
/**
* The LIRS stack size.
*/
private
int
stackSize
;
/**
* The size of the LIRS queue for resident cold entries.
*/
private
int
queueSize
;
/**
* The size of the LIRS queue for non-resident cold entries.
*/
private
int
queue2Size
;
/**
* The map array. The size is always a power of 2.
*/
private
Entry
<
V
>[]
entries
;
/**
* The stack of recently referenced elements. This includes all hot entries,
* the recently referenced cold entries, and all non-resident cold entries.
*/
private
Entry
<
V
>
stack
;
/**
* The queue of resident cold entries.
*/
private
Entry
<
V
>
queue
;
/**
* The queue of non-resident cold entries.
*/
private
Entry
<
V
>
queue2
;
/**
* The number of times any item was moved to the top of the stack.
*/
private
int
stackMoveCounter
;
/**
* Create a new cache.
*
* @param maxMemory the maximum memory to use
* @param averageMemory the average memory usage of an object
*/
Segment
(
long
maxMemory
,
int
averageMemory
)
{
setMaxMemory
(
maxMemory
);
setAverageMemory
(
averageMemory
);
clear
();
}
synchronized
void
clear
()
{
// calculate the size of the map array
// assume a fill factor of at most 80%
long
maxLen
=
(
long
)
(
maxMemory
/
averageMemory
/
0.75
);
// the size needs to be a power of 2
long
l
=
8
;
while
(
l
<
maxLen
)
{
l
+=
l
;
}
// the array size is at most 2^31 elements
int
len
=
(
int
)
Math
.
min
(
1L
<<
31
,
l
);
// the bit mask has all bits set
mask
=
len
-
1
;
// initialize the stack and queue heads
stack
=
new
Entry
<
V
>();
stack
.
stackPrev
=
stack
.
stackNext
=
stack
;
queue
=
new
Entry
<
V
>();
queue
.
queuePrev
=
queue
.
queueNext
=
queue
;
queue2
=
new
Entry
<
V
>();
queue2
.
queuePrev
=
queue2
.
queueNext
=
queue2
;
// first set to null - avoiding out of memory
entries
=
null
;
@SuppressWarnings
(
"unchecked"
)
Entry
<
V
>[]
e
=
new
Entry
[
len
];
entries
=
e
;
mapSize
=
0
;
usedMemory
=
0
;
stackSize
=
queueSize
=
queue2Size
=
0
;
}
V
peek
(
long
key
,
int
hash
)
{
Entry
<
V
>
e
=
find
(
key
,
hash
);
return
e
==
null
?
null
:
e
.
value
;
}
int
getMemory
(
long
key
,
int
hash
)
{
Entry
<
V
>
e
=
find
(
key
,
hash
);
return
e
==
null
?
0
:
e
.
memory
;
}
V
get
(
long
key
,
int
hash
)
{
Entry
<
V
>
e
=
find
(
key
,
hash
);
if
(
e
==
null
)
{
// the entry was not found
return
null
;
}
V
value
=
e
.
value
;
if
(
value
==
null
)
{
// it was a non-resident entry
return
null
;
}
if
(
e
.
isHot
())
{
if
(
e
!=
stack
.
stackNext
)
{
if
(
stackMoveDistance
==
0
||
stackMoveCounter
-
e
.
topMove
>
stackMoveDistance
)
{
access
(
key
,
hash
);
}
}
}
else
{
access
(
key
,
hash
);
}
return
value
;
}
/**
* Access an item, moving the entry to the top of the stack or front of the
* queue if found.
*
* @param key the key
*/
private
synchronized
void
access
(
long
key
,
int
hash
)
{
Entry
<
V
>
e
=
find
(
key
,
hash
);
if
(
e
==
null
||
e
.
value
==
null
)
{
return
;
}
if
(
e
.
isHot
())
{
if
(
e
!=
stack
.
stackNext
)
{
if
(
stackMoveDistance
==
0
||
stackMoveCounter
-
e
.
topMove
>
stackMoveDistance
)
{
// move a hot entry to the top of the stack
// unless it is already there
boolean
wasEnd
=
e
==
stack
.
stackPrev
;
removeFromStack
(
e
);
if
(
wasEnd
)
{
// if moving the last entry, the last entry
// could not be cold, which is not allowed
pruneStack
();
}
addToStack
(
e
);
}
}
}
else
{
removeFromQueue
(
e
);
if
(
e
.
stackNext
!=
null
)
{
// resident cold entries become hot
// if they are on the stack
removeFromStack
(
e
);
// which means a hot entry needs to become cold
convertOldestHotToCold
();
}
else
{
// cold entries that are not on the stack
// move to the front of the queue
addToQueue
(
queue
,
e
);
}
// in any case, the cold entry is moved to the top of the stack
addToStack
(
e
);
}
}
V
put
(
long
key
,
V
value
,
int
hash
)
{
return
put
(
key
,
value
,
hash
,
averageMemory
);
}
synchronized
V
put
(
long
key
,
V
value
,
int
hash
,
int
memory
)
{
if
(
value
==
null
)
{
throw
new
NullPointerException
();
}
V
old
;
Entry
<
V
>
e
=
find
(
key
,
hash
);
if
(
e
==
null
)
{
old
=
null
;
}
else
{
old
=
e
.
value
;
remove
(
key
,
hash
);
}
e
=
new
Entry
<
V
>();
e
.
key
=
key
;
e
.
value
=
value
;
e
.
memory
=
memory
;
int
index
=
hash
&
mask
;
e
.
mapNext
=
entries
[
index
];
entries
[
index
]
=
e
;
usedMemory
+=
memory
;
if
(
usedMemory
>
maxMemory
&&
mapSize
>
0
)
{
// an old entry needs to be removed
evict
(
e
);
}
mapSize
++;
// added entries are always added to the stack
addToStack
(
e
);
return
old
;
}
synchronized
V
remove
(
long
key
,
int
hash
)
{
int
index
=
hash
&
mask
;
Entry
<
V
>
e
=
entries
[
index
];
if
(
e
==
null
)
{
return
null
;
}
V
old
;
if
(
e
.
key
==
key
)
{
old
=
e
.
value
;
entries
[
index
]
=
e
.
mapNext
;
}
else
{
Entry
<
V
>
last
;
do
{
last
=
e
;
e
=
e
.
mapNext
;
if
(
e
==
null
)
{
return
null
;
}
}
while
(
e
.
key
!=
key
);
old
=
e
.
value
;
last
.
mapNext
=
e
.
mapNext
;
}
mapSize
--;
usedMemory
-=
e
.
memory
;
if
(
e
.
stackNext
!=
null
)
{
removeFromStack
(
e
);
}
if
(
e
.
isHot
())
{
// when removing a hot entry, the newest cold entry gets hot,
// so the number of hot entries does not change
e
=
queue
.
queueNext
;
if
(
e
!=
queue
)
{
removeFromQueue
(
e
);
if
(
e
.
stackNext
==
null
)
{
addToStackBottom
(
e
);
}
}
}
else
{
removeFromQueue
(
e
);
}
pruneStack
();
return
old
;
}
/**
* Evict cold entries (resident and non-resident) until the memory limit is
* reached. The new entry is added as a cold entry, except if it is the only
* entry.
*
* @param newCold a new cold entry
*/
private
void
evict
(
Entry
<
V
>
newCold
)
{
// ensure there are not too many hot entries:
// left shift of 5 is multiplication by 32, that means if there are less
// than 1/32 (3.125%) cold entries, a new hot entry needs to become cold
while
((
queueSize
<<
5
)
<
mapSize
)
{
convertOldestHotToCold
();
}
if
(
stackSize
>
0
)
{
// the new cold entry is at the top of the queue
addToQueue
(
queue
,
newCold
);
}
// the oldest resident cold entries become non-resident
// but at least one cold entry (the new one) must stay
while
(
usedMemory
>
maxMemory
&&
queueSize
>
1
)
{
Entry
<
V
>
e
=
queue
.
queuePrev
;
usedMemory
-=
e
.
memory
;
removeFromQueue
(
e
);
e
.
value
=
null
;
e
.
memory
=
0
;
addToQueue
(
queue2
,
e
);
// the size of the non-resident-cold entries needs to be limited
while
(
queue2Size
+
queue2Size
>
stackSize
)
{
e
=
queue2
.
queuePrev
;
int
hash
=
getHash
(
e
.
key
);
remove
(
e
.
key
,
hash
);
}
}
}
private
void
convertOldestHotToCold
()
{
// the last entry of the stack is known to be hot
Entry
<
V
>
last
=
stack
.
stackPrev
;
// remove from stack - which is done anyway in the stack pruning, but we
// can do it here as well
removeFromStack
(
last
);
// adding an entry to the queue will make it cold
addToQueue
(
queue
,
last
);
pruneStack
();
}
/**
* Ensure the last entry of the stack is cold.
*/
private
void
pruneStack
()
{
while
(
true
)
{
Entry
<
V
>
last
=
stack
.
stackPrev
;
if
(
last
==
stack
||
last
.
isHot
())
{
break
;
}
// the cold entry is still in the queue
removeFromStack
(
last
);
}
}
/**
* Try to find an entry in the map.
*
* @param key the key
* @return the entry (might be a non-resident)
*/
Entry
<
V
>
find
(
long
key
,
int
hash
)
{
int
index
=
hash
&
mask
;
Entry
<
V
>
e
=
entries
[
index
];
while
(
e
!=
null
&&
e
.
key
!=
key
)
{
e
=
e
.
mapNext
;
}
return
e
;
}
private
void
addToStack
(
Entry
<
V
>
e
)
{
e
.
stackPrev
=
stack
;
e
.
stackNext
=
stack
.
stackNext
;
e
.
stackNext
.
stackPrev
=
e
;
stack
.
stackNext
=
e
;
stackSize
++;
e
.
topMove
=
stackMoveCounter
++;
}
private
void
addToStackBottom
(
Entry
<
V
>
e
)
{
e
.
stackNext
=
stack
;
e
.
stackPrev
=
stack
.
stackPrev
;
e
.
stackPrev
.
stackNext
=
e
;
stack
.
stackPrev
=
e
;
stackSize
++;
}
private
void
removeFromStack
(
Entry
<
V
>
e
)
{
e
.
stackPrev
.
stackNext
=
e
.
stackNext
;
e
.
stackNext
.
stackPrev
=
e
.
stackPrev
;
e
.
stackPrev
=
e
.
stackNext
=
null
;
stackSize
--;
}
private
void
addToQueue
(
Entry
<
V
>
q
,
Entry
<
V
>
e
)
{
e
.
queuePrev
=
q
;
e
.
queueNext
=
q
.
queueNext
;
e
.
queueNext
.
queuePrev
=
e
;
q
.
queueNext
=
e
;
if
(
e
.
value
!=
null
)
{
queueSize
++;
}
else
{
queue2Size
++;
}
}
private
void
removeFromQueue
(
Entry
<
V
>
e
)
{
e
.
queuePrev
.
queueNext
=
e
.
queueNext
;
e
.
queueNext
.
queuePrev
=
e
.
queuePrev
;
e
.
queuePrev
=
e
.
queueNext
=
null
;
if
(
e
.
value
!=
null
)
{
queueSize
--;
}
else
{
queue2Size
--;
}
}
synchronized
List
<
Long
>
keys
(
boolean
cold
,
boolean
nonResident
)
{
ArrayList
<
Long
>
keys
=
new
ArrayList
<
Long
>();
if
(
cold
)
{
Entry
<
V
>
start
=
nonResident
?
queue2
:
queue
;
for
(
Entry
<
V
>
e
=
start
.
queueNext
;
e
!=
start
;
e
=
e
.
queueNext
)
{
keys
.
add
(
e
.
key
);
}
}
else
{
for
(
Entry
<
V
>
e
=
stack
.
stackNext
;
e
!=
stack
;
e
=
e
.
stackNext
)
{
keys
.
add
(
e
.
key
);
}
}
return
keys
;
}
int
size
()
{
return
mapSize
-
queue2Size
;
}
boolean
containsKey
(
long
key
,
int
hash
)
{
Entry
<
V
>
e
=
find
(
key
,
hash
);
return
e
!=
null
&&
e
.
value
!=
null
;
}
Set
<
Long
>
keySet
()
{
HashSet
<
Long
>
set
=
new
HashSet
<
Long
>();
for
(
Entry
<
V
>
e
=
stack
.
stackNext
;
e
!=
stack
;
e
=
e
.
stackNext
)
{
set
.
add
(
e
.
key
);
}
for
(
Entry
<
V
>
e
=
queue
.
queueNext
;
e
!=
queue
;
e
=
e
.
queueNext
)
{
set
.
add
(
e
.
key
);
}
return
set
;
}
int
sizeHot
()
{
return
mapSize
-
queueSize
-
queue2Size
;
}
int
sizeNonResident
()
{
return
queue2Size
;
}
int
sizeMapArray
()
{
return
entries
.
length
;
}
long
getUsedMemory
()
{
return
usedMemory
;
}
void
setMaxMemory
(
long
maxMemory
)
{
if
(
maxMemory
<=
0
)
{
throw
new
IllegalArgumentException
(
"Max memory must be larger than 0"
);
}
this
.
maxMemory
=
maxMemory
;
}
long
getMaxMemory
()
{
return
maxMemory
;
}
void
setAverageMemory
(
int
averageMemory
)
{
if
(
averageMemory
<=
0
)
{
throw
new
IllegalArgumentException
(
"Average memory must be larger than 0"
);
}
this
.
averageMemory
=
averageMemory
;
}
}
/**
* A cache entry. Each entry is either hot (low inter-reference recency;
* LIR), cold (high inter-reference recency; HIR), or non-resident-cold. Hot
* entries are in the stack only. Cold entries are in the queue, and may be
* in the stack. Non-resident-cold entries have their value set to null and
* are in the stack and in the non-resident queue.
*
* @param <K> the key type
* @param <V> the value type
*/
static
class
Entry
<
V
>
{
/**
* The key.
*/
long
key
;
/**
* The value. Set to null for non-resident-cold entries.
*/
V
value
;
/**
* The estimated memory used.
*/
int
memory
;
/**
* When the item was last moved to the top of the stack.
*/
int
topMove
;
/**
* The next entry in the stack.
*/
Entry
<
V
>
stackNext
;
/**
* The previous entry in the stack.
*/
Entry
<
V
>
stackPrev
;
/**
* The next entry in the queue (either the resident queue or the
* non-resident queue).
*/
Entry
<
V
>
queueNext
;
/**
* The previous entry in the queue.
*/
Entry
<
V
>
queuePrev
;
/**
* The next entry in the map
*/
Entry
<
V
>
mapNext
;
/**
* Whether this entry is hot. Cold entries are in one of the two queues.
*
* @return whether the entry is hot
*/
boolean
isHot
()
{
return
queueNext
==
null
;
}
}
}
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论