Skip to content
项目
群组
代码片段
帮助
正在加载...
帮助
为 GitLab 提交贡献
登录/注册
切换导航
H
h2database
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
分枝图
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
计划
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
分枝图
统计图
创建新议题
作业
提交
议题看板
打开侧边栏
Administrator
h2database
Commits
670d256c
提交
670d256c
authored
11月 06, 2012
作者:
Thomas Mueller
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
A persistent multi-version map: compact based on estimated size, not page count
上级
3a92ea32
显示空白字符变更
内嵌
并排
正在显示
5 个修改的文件
包含
166 行增加
和
166 行删除
+166
-166
TestMVStore.java
h2/src/test/org/h2/test/store/TestMVStore.java
+1
-1
Chunk.java
h2/src/tools/org/h2/dev/store/btree/Chunk.java
+38
-12
Dump.java
h2/src/tools/org/h2/dev/store/btree/Dump.java
+6
-2
MVStore.java
h2/src/tools/org/h2/dev/store/btree/MVStore.java
+109
-126
Page.java
h2/src/tools/org/h2/dev/store/btree/Page.java
+12
-25
没有找到文件。
h2/src/test/org/h2/test/store/TestMVStore.java
浏览文件 @
670d256c
...
@@ -577,7 +577,7 @@ public class TestMVStore extends TestBase {
...
@@ -577,7 +577,7 @@ public class TestMVStore extends TestBase {
m
.
put
(
j
+
i
,
"Hello "
+
j
);
m
.
put
(
j
+
i
,
"Hello "
+
j
);
}
}
s
.
store
();
s
.
store
();
s
.
compact
(
80
);
//
s.compact(80);
s
.
close
();
s
.
close
();
long
len
=
FileUtils
.
size
(
fileName
);
long
len
=
FileUtils
.
size
(
fileName
);
// System.out.println(" len:" + len);
// System.out.println(" len:" + len);
...
...
h2/src/tools/org/h2/dev/store/btree/Chunk.java
浏览文件 @
670d256c
...
@@ -6,6 +6,7 @@
...
@@ -6,6 +6,7 @@
*/
*/
package
org
.
h2
.
dev
.
store
.
btree
;
package
org
.
h2
.
dev
.
store
.
btree
;
import
java.nio.ByteBuffer
;
import
java.util.HashMap
;
import
java.util.HashMap
;
/**
/**
...
@@ -19,6 +20,7 @@ import java.util.HashMap;
...
@@ -19,6 +20,7 @@ import java.util.HashMap;
* 4 bytes: length
* 4 bytes: length
* 4 bytes: chunk id (an incrementing number)
* 4 bytes: chunk id (an incrementing number)
* 8 bytes: metaRootPos
* 8 bytes: metaRootPos
* 8 bytes: maxLengthLive
* [ Page ] *
* [ Page ] *
*/
*/
public
class
Chunk
{
public
class
Chunk
{
...
@@ -34,19 +36,19 @@ public class Chunk {
...
@@ -34,19 +36,19 @@ public class Chunk {
long
start
;
long
start
;
/**
/**
* The length in bytes
(may be larger than the actual value)
.
* The length in bytes.
*/
*/
long
length
;
int
length
;
/**
/**
* The
entry count
.
* The
sum of the max length of all pages
.
*/
*/
int
entryCount
;
long
maxLength
;
/**
/**
* The
number of life (non-garbage) objects
.
* The
sum of the max length of all pages that are in use
.
*/
*/
int
liveCount
;
long
maxLengthLive
;
/**
/**
* The garbage collection priority.
* The garbage collection priority.
...
@@ -67,6 +69,30 @@ public class Chunk {
...
@@ -67,6 +69,30 @@ public class Chunk {
this
.
id
=
id
;
this
.
id
=
id
;
}
}
public
static
Chunk
fromHeader
(
ByteBuffer
buff
,
long
start
)
{
if
(
buff
.
get
()
!=
'c'
)
{
throw
new
RuntimeException
(
"File corrupt"
);
}
int
length
=
buff
.
getInt
();
int
chunkId
=
buff
.
getInt
();
long
metaRootPos
=
buff
.
getLong
();
long
maxLengthLive
=
buff
.
getLong
();
Chunk
c
=
new
Chunk
(
chunkId
);
c
.
length
=
length
;
c
.
start
=
start
;
c
.
metaRootPos
=
metaRootPos
;
c
.
maxLengthLive
=
maxLengthLive
;
return
c
;
}
void
writeHeader
(
ByteBuffer
buff
)
{
buff
.
put
((
byte
)
'c'
);
buff
.
putInt
(
length
);
buff
.
putInt
(
id
);
buff
.
putLong
(
metaRootPos
);
buff
.
putLong
(
maxLengthLive
);
}
/**
/**
* Build a block from the given string.
* Build a block from the given string.
*
*
...
@@ -78,16 +104,16 @@ public class Chunk {
...
@@ -78,16 +104,16 @@ public class Chunk {
int
id
=
Integer
.
parseInt
(
map
.
get
(
"id"
));
int
id
=
Integer
.
parseInt
(
map
.
get
(
"id"
));
Chunk
c
=
new
Chunk
(
id
);
Chunk
c
=
new
Chunk
(
id
);
c
.
start
=
Long
.
parseLong
(
map
.
get
(
"start"
));
c
.
start
=
Long
.
parseLong
(
map
.
get
(
"start"
));
c
.
length
=
Long
.
parseLong
(
map
.
get
(
"length"
));
c
.
length
=
Integer
.
parseInt
(
map
.
get
(
"length"
));
c
.
entryCount
=
Integer
.
parseInt
(
map
.
get
(
"entryCount
"
));
c
.
maxLength
=
Long
.
parseLong
(
map
.
get
(
"maxLength
"
));
c
.
liveCount
=
Integer
.
parseInt
(
map
.
get
(
"liveCount
"
));
c
.
maxLengthLive
=
Long
.
parseLong
(
map
.
get
(
"maxLengthLive
"
));
c
.
metaRootPos
=
Long
.
parseLong
(
map
.
get
(
"metaRoot"
));
c
.
metaRootPos
=
Long
.
parseLong
(
map
.
get
(
"metaRoot"
));
c
.
version
=
Long
.
parseLong
(
map
.
get
(
"version"
));
c
.
version
=
Long
.
parseLong
(
map
.
get
(
"version"
));
return
c
;
return
c
;
}
}
public
int
getFillRate
()
{
public
int
getFillRate
()
{
return
entryCount
==
0
?
0
:
100
*
liveCount
/
entryCount
;
return
(
int
)
(
maxLength
==
0
?
0
:
100
*
maxLengthLive
/
maxLength
)
;
}
}
public
int
hashCode
()
{
public
int
hashCode
()
{
...
@@ -103,8 +129,8 @@ public class Chunk {
...
@@ -103,8 +129,8 @@ public class Chunk {
"id:"
+
id
+
","
+
"id:"
+
id
+
","
+
"start:"
+
start
+
","
+
"start:"
+
start
+
","
+
"length:"
+
length
+
","
+
"length:"
+
length
+
","
+
"
entryCount:"
+
entryCount
+
","
+
"
maxLength:"
+
maxLength
+
","
+
"
liveCount:"
+
liveCount
+
","
+
"
maxLengthLive:"
+
maxLengthLive
+
","
+
"metaRoot:"
+
metaRootPos
+
","
+
"metaRoot:"
+
metaRootPos
+
","
+
"version:"
+
version
;
"version:"
+
version
;
}
}
...
...
h2/src/tools/org/h2/dev/store/btree/Dump.java
浏览文件 @
670d256c
...
@@ -80,8 +80,12 @@ public class Dump {
...
@@ -80,8 +80,12 @@ public class Dump {
int
chunkLength
=
block
.
getInt
();
int
chunkLength
=
block
.
getInt
();
int
chunkId
=
block
.
getInt
();
int
chunkId
=
block
.
getInt
();
long
metaRootPos
=
block
.
getLong
();
long
metaRootPos
=
block
.
getLong
();
writer
.
println
(
" chunk "
+
chunkId
+
" at "
+
pos
+
long
maxLengthLive
=
block
.
getLong
();
" length "
+
chunkLength
+
" root "
+
metaRootPos
);
writer
.
println
(
" chunk "
+
chunkId
+
" at "
+
pos
+
" length "
+
chunkLength
+
" root "
+
metaRootPos
+
" maxLengthLive "
+
maxLengthLive
);
ByteBuffer
chunk
=
ByteBuffer
.
allocate
(
chunkLength
);
ByteBuffer
chunk
=
ByteBuffer
.
allocate
(
chunkLength
);
file
.
position
(
pos
);
file
.
position
(
pos
);
FileUtils
.
readFully
(
file
,
chunk
);
FileUtils
.
readFully
(
file
,
chunk
);
...
...
h2/src/tools/org/h2/dev/store/btree/MVStore.java
浏览文件 @
670d256c
...
@@ -35,10 +35,7 @@ header:
...
@@ -35,10 +35,7 @@ header:
H:3,blockSize=4096,...
H:3,blockSize=4096,...
TODO:
TODO:
- compact: use total max length instead of page count (liveCount)
- support background writes (store old version)
- support database version / app version
- support database version / app version
- limited support for writing to old versions (branches)
- atomic test-and-set (when supporting concurrent writes)
- atomic test-and-set (when supporting concurrent writes)
- possibly split chunk data into immutable and mutable
- possibly split chunk data into immutable and mutable
- support stores that span multiple files (chunks stored in other files)
- support stores that span multiple files (chunks stored in other files)
...
@@ -56,13 +53,17 @@ TODO:
...
@@ -56,13 +53,17 @@ TODO:
- recovery: ensure data is not overwritten for 1 minute
- recovery: ensure data is not overwritten for 1 minute
- pluggable caching (specially for in-memory file systems)
- pluggable caching (specially for in-memory file systems)
- file locking
- file locking
- allocate memory
Utils.newBytes
- allocate memory
with Utils.newBytes and so on
- unified exception handling
- unified exception handling
- check if locale specific string comparison can make data disappear
- check if locale specific string comparison can make data disappear
- concurrent map; avoid locking during IO (pre-load pages)
- concurrent map; avoid locking during IO (pre-load pages)
- maybe split database into multiple files, to speed up compact operation
- maybe split database into multiple files, to speed up compact operation
- automated 'kill process' and 'power failure' test
- automated 'kill process' and 'power failure' test
- implement table engine for H2
- implement table engine for H2
- auto-compact from time to time and on close
- test and possibly improve compact operation (for large dbs)
- support background writes (concurrent modification & store)
- limited support for writing to old versions (branches)
*/
*/
...
@@ -384,6 +385,7 @@ public class MVStore {
...
@@ -384,6 +385,7 @@ public class MVStore {
c
.
start
=
header
.
start
;
c
.
start
=
header
.
start
;
c
.
length
=
header
.
length
;
c
.
length
=
header
.
length
;
c
.
metaRootPos
=
header
.
metaRootPos
;
c
.
metaRootPos
=
header
.
metaRootPos
;
c
.
maxLengthLive
=
header
.
maxLengthLive
;
}
}
lastChunkId
=
Math
.
max
(
c
.
id
,
lastChunkId
);
lastChunkId
=
Math
.
max
(
c
.
id
,
lastChunkId
);
chunks
.
put
(
c
.
id
,
c
);
chunks
.
put
(
c
.
id
,
c
);
...
@@ -499,28 +501,26 @@ public class MVStore {
...
@@ -499,28 +501,26 @@ public class MVStore {
return
currentVersion
;
return
currentVersion
;
}
}
// the last chunk
might have been changed
in the last save()
// the last chunk
was not completely correct
in the last save()
// this needs to be updated now (it's better not to update right after
,
// this needs to be updated now (it's better not to update right after
// s
ave()
, because that would modify the meta map again)
// s
toring
, because that would modify the meta map again)
Chunk
c
=
chunks
.
get
(
lastChunkId
);
Chunk
c
=
chunks
.
get
(
lastChunkId
);
if
(
c
!=
null
)
{
if
(
c
!=
null
)
{
meta
.
put
(
"chunk."
+
c
.
id
,
c
.
toString
());
meta
.
put
(
"chunk."
+
c
.
id
,
c
.
toString
());
}
}
int
chunkId
=
++
lastChunkId
;
c
=
new
Chunk
(++
lastChunkId
);
c
.
maxLength
=
Long
.
MAX_VALUE
;
c
=
new
Chunk
(
chunkId
);
c
.
maxLengthLive
=
Long
.
MAX_VALUE
;
c
.
entryCount
=
Integer
.
MAX_VALUE
;
c
.
liveCount
=
Integer
.
MAX_VALUE
;
c
.
start
=
Long
.
MAX_VALUE
;
c
.
start
=
Long
.
MAX_VALUE
;
c
.
length
=
Long
.
MAX_VALUE
;
c
.
length
=
Integer
.
MAX_VALUE
;
c
.
version
=
currentVersion
;
c
.
version
=
currentVersion
;
chunks
.
put
(
c
.
id
,
c
);
chunks
.
put
(
c
.
id
,
c
);
meta
.
put
(
"chunk."
+
c
.
id
,
c
.
toString
());
meta
.
put
(
"chunk."
+
c
.
id
,
c
.
toString
());
applyFreedChunks
();
applyFreedChunks
();
ArrayList
<
Integer
>
removedChunks
=
New
.
arrayList
();
ArrayList
<
Integer
>
removedChunks
=
New
.
arrayList
();
for
(
Chunk
x
:
chunks
.
values
())
{
for
(
Chunk
x
:
chunks
.
values
())
{
if
(
x
.
liveCount
==
0
&&
(
retainChunk
==
-
1
||
x
.
id
<
retainChunk
))
{
if
(
x
.
maxLengthLive
==
0
&&
(
retainChunk
==
-
1
||
x
.
id
<
retainChunk
))
{
meta
.
remove
(
"chunk."
+
x
.
id
);
meta
.
remove
(
"chunk."
+
x
.
id
);
removedChunks
.
add
(
x
.
id
);
removedChunks
.
add
(
x
.
id
);
}
else
{
}
else
{
...
@@ -528,7 +528,6 @@ public class MVStore {
...
@@ -528,7 +528,6 @@ public class MVStore {
}
}
applyFreedChunks
();
applyFreedChunks
();
}
}
int
count
=
0
;
int
maxLength
=
1
+
4
+
4
+
8
;
int
maxLength
=
1
+
4
+
4
+
8
;
for
(
MVMap
<?,
?>
m
:
mapsChanged
.
values
())
{
for
(
MVMap
<?,
?>
m
:
mapsChanged
.
values
())
{
if
(
m
==
meta
||
!
m
.
hasUnsavedChanges
())
{
if
(
m
==
meta
||
!
m
.
hasUnsavedChanges
())
{
...
@@ -539,36 +538,32 @@ public class MVStore {
...
@@ -539,36 +538,32 @@ public class MVStore {
meta
.
put
(
"root."
+
m
.
getId
(),
"0"
);
meta
.
put
(
"root."
+
m
.
getId
(),
"0"
);
}
else
{
}
else
{
maxLength
+=
p
.
getMaxLengthTempRecursive
();
maxLength
+=
p
.
getMaxLengthTempRecursive
();
count
+=
p
.
countTempRecursive
();
meta
.
put
(
"root."
+
m
.
getId
(),
String
.
valueOf
(
Long
.
MAX_VALUE
));
meta
.
put
(
"root."
+
m
.
getId
(),
String
.
valueOf
(
Long
.
MAX_VALUE
));
}
}
}
}
maxLength
+=
meta
.
getRoot
().
getMaxLengthTempRecursive
();
maxLength
+=
meta
.
getRoot
().
getMaxLengthTempRecursive
();
count
+=
meta
.
getRoot
().
countTempRecursive
();
ByteBuffer
buff
=
ByteBuffer
.
allocate
(
maxLength
);
ByteBuffer
buff
=
ByteBuffer
.
allocate
(
maxLength
);
// need to patch the header later
// need to patch the header later
buff
.
put
((
byte
)
'c'
);
c
.
writeHeader
(
buff
);
buff
.
putInt
(
0
);
c
.
maxLength
=
0
;
buff
.
putInt
(
0
);
c
.
maxLengthLive
=
0
;
buff
.
putLong
(
0
);
for
(
MVMap
<?,
?>
m
:
mapsChanged
.
values
())
{
for
(
MVMap
<?,
?>
m
:
mapsChanged
.
values
())
{
if
(
m
==
meta
||
!
m
.
hasUnsavedChanges
())
{
if
(
m
==
meta
||
!
m
.
hasUnsavedChanges
())
{
continue
;
continue
;
}
}
Page
p
=
m
.
getRoot
();
Page
p
=
m
.
getRoot
();
if
(
p
.
getTotalCount
()
>
0
)
{
if
(
p
.
getTotalCount
()
>
0
)
{
long
root
=
p
.
writeTempRecursive
(
buff
,
chunkId
);
long
root
=
p
.
writeTempRecursive
(
c
,
buff
);
meta
.
put
(
"root."
+
m
.
getId
(),
""
+
root
);
meta
.
put
(
"root."
+
m
.
getId
(),
""
+
root
);
}
}
}
}
// fix metadata
c
.
entryCount
=
count
;
c
.
liveCount
=
count
;
meta
.
put
(
"chunk."
+
c
.
id
,
c
.
toString
());
meta
.
put
(
"chunk."
+
c
.
id
,
c
.
toString
());
meta
.
getRoot
().
writeTempRecursive
(
buff
,
chunkId
);
// this will modify maxLengthLive, but
// the correct value is written in the chunk header
meta
.
getRoot
().
writeTempRecursive
(
c
,
buff
);
buff
.
flip
();
buff
.
flip
();
int
length
=
buff
.
limit
();
int
length
=
buff
.
limit
();
...
@@ -576,17 +571,18 @@ public class MVStore {
...
@@ -576,17 +571,18 @@ public class MVStore {
// need to keep old chunks
// need to keep old chunks
// until they are are no longer referenced
// until they are are no longer referenced
// by a old version
// by a
n
old version
// so empty space is not reused too early
// so empty space is not reused too early
for
(
int
x
:
removedChunks
)
{
for
(
int
x
:
removedChunks
)
{
chunks
.
remove
(
x
);
chunks
.
remove
(
x
);
}
}
buff
.
rewind
();
buff
.
rewind
();
buff
.
put
((
byte
)
'c'
);
buff
.
putInt
(
length
);
c
.
start
=
filePos
;
buff
.
putInt
(
chunkId
);
c
.
length
=
length
;
buff
.
putLong
(
meta
.
getRoot
().
getPos
());
c
.
metaRootPos
=
meta
.
getRoot
().
getPos
();
c
.
writeHeader
(
buff
);
buff
.
rewind
();
buff
.
rewind
();
try
{
try
{
writeCount
++;
writeCount
++;
...
@@ -597,9 +593,6 @@ public class MVStore {
...
@@ -597,9 +593,6 @@ public class MVStore {
}
}
rootChunkStart
=
filePos
;
rootChunkStart
=
filePos
;
revertTemp
();
revertTemp
();
// update the start position and length
c
.
start
=
filePos
;
c
.
length
=
length
;
long
version
=
incrementVersion
();
long
version
=
incrementVersion
();
// write the new version (after the commit)
// write the new version (after the commit)
...
@@ -609,11 +602,13 @@ public class MVStore {
...
@@ -609,11 +602,13 @@ public class MVStore {
}
}
private
void
applyFreedChunks
()
{
private
void
applyFreedChunks
()
{
// apply liveCount changes
for
(
HashMap
<
Integer
,
Chunk
>
freed
:
freedChunks
.
values
())
{
for
(
HashMap
<
Integer
,
Chunk
>
freed
:
freedChunks
.
values
())
{
for
(
Chunk
f
:
freed
.
values
())
{
for
(
Chunk
f
:
freed
.
values
())
{
Chunk
c
=
chunks
.
get
(
f
.
id
);
Chunk
c
=
chunks
.
get
(
f
.
id
);
c
.
liveCount
+=
f
.
liveCount
;
c
.
maxLengthLive
+=
f
.
maxLengthLive
;
if
(
c
.
maxLengthLive
<
0
)
{
throw
new
RuntimeException
(
"Corrupt max length"
);
}
}
}
}
}
freedChunks
.
clear
();
freedChunks
.
clear
();
...
@@ -714,17 +709,7 @@ public class MVStore {
...
@@ -714,17 +709,7 @@ public class MVStore {
ByteBuffer
buff
=
ByteBuffer
.
wrap
(
new
byte
[
32
]);
ByteBuffer
buff
=
ByteBuffer
.
wrap
(
new
byte
[
32
]);
DataUtils
.
readFully
(
file
,
buff
);
DataUtils
.
readFully
(
file
,
buff
);
buff
.
rewind
();
buff
.
rewind
();
if
(
buff
.
get
()
!=
'c'
)
{
return
Chunk
.
fromHeader
(
buff
,
start
);
throw
new
RuntimeException
(
"File corrupt"
);
}
int
length
=
buff
.
getInt
();
int
chunkId
=
buff
.
getInt
();
long
metaRootPos
=
buff
.
getLong
();
Chunk
c
=
new
Chunk
(
chunkId
);
c
.
start
=
start
;
c
.
length
=
length
;
c
.
metaRootPos
=
metaRootPos
;
return
c
;
}
catch
(
IOException
e
)
{
}
catch
(
IOException
e
)
{
throw
new
RuntimeException
(
e
);
throw
new
RuntimeException
(
e
);
}
}
...
@@ -743,23 +728,22 @@ public class MVStore {
...
@@ -743,23 +728,22 @@ public class MVStore {
// avoid division by 0
// avoid division by 0
return
false
;
return
false
;
}
}
long
liveCountTotal
=
0
,
entryCountTotal
=
0
;
long
maxLengthSum
=
0
,
maxLengthLiveSum
=
0
;
for
(
Chunk
c
:
chunks
.
values
())
{
for
(
Chunk
c
:
chunks
.
values
())
{
entryCountTotal
+=
c
.
entryCount
;
maxLengthSum
+=
c
.
maxLength
;
liveCountTotal
+=
c
.
liveCount
;
maxLengthLiveSum
+=
c
.
maxLengthLive
;
}
}
if
(
entryCountTotal
<=
0
)
{
if
(
maxLengthSum
<=
0
)
{
// avoid division by 0
// avoid division by 0
entryCountTotal
=
1
;
maxLengthSum
=
1
;
}
}
int
percentTotal
=
(
int
)
(
100
*
liveCountTotal
/
entryCountTotal
);
int
percentTotal
=
(
int
)
(
100
*
maxLengthLiveSum
/
maxLengthSum
);
if
(
percentTotal
>
fillRate
)
{
if
(
percentTotal
>
fillRate
)
{
return
false
;
return
false
;
}
}
// calculate how many entries a chunk has on average
// calculate the average max length
// TODO use the max size instead of the count
int
averageMaxLength
=
(
int
)
(
maxLengthSum
/
chunks
.
size
());
int
averageEntryCount
=
(
int
)
(
entryCountTotal
/
chunks
.
size
());
// the 'old' list contains the chunks we want to free up
// the 'old' list contains the chunks we want to free up
ArrayList
<
Chunk
>
old
=
New
.
arrayList
();
ArrayList
<
Chunk
>
old
=
New
.
arrayList
();
...
@@ -781,16 +765,16 @@ public class MVStore {
...
@@ -781,16 +765,16 @@ public class MVStore {
}
}
});
});
// find out up to were we need to move
// find out up to were
in the old list
we need to move
// try to move one (average sized) chunk
// try to move one (average sized) chunk
int
moveCount
=
0
;
long
moved
=
0
;
Chunk
move
=
null
;
Chunk
move
=
null
;
for
(
Chunk
c
:
old
)
{
for
(
Chunk
c
:
old
)
{
if
(
move
Count
+
c
.
liveCount
>
averageEntryCount
)
{
if
(
move
!=
null
&&
moved
+
c
.
maxLengthLive
>
averageMaxLength
)
{
break
;
break
;
}
}
log
(
" chunk "
+
c
.
id
+
" "
+
c
.
getFillRate
()
+
"% full; prio="
+
c
.
collectPriority
);
log
(
" chunk "
+
c
.
id
+
" "
+
c
.
getFillRate
()
+
"% full; prio="
+
c
.
collectPriority
);
move
Count
+=
c
.
liveCount
;
move
d
+=
c
.
maxLengthLive
;
move
=
c
;
move
=
c
;
}
}
...
@@ -804,57 +788,43 @@ public class MVStore {
...
@@ -804,57 +788,43 @@ public class MVStore {
it
.
remove
();
it
.
remove
();
}
}
}
}
while
(!
isKnownVersion
(
move
.
version
))
{
int
id
=
move
.
id
;
// iterate over all the pages in the old pages
while
(
true
)
{
for
(
Chunk
c
:
old
)
{
Chunk
m
=
chunks
.
get
(++
id
);
copyLive
(
c
,
old
);
if
(
id
>
lastChunkId
)
{
// no known version
return
false
;
}
if
(
m
!=
null
)
{
move
=
m
;
break
;
}
}
}
}
// the metaRootPos might not be set
store
();
move
=
readChunkHeader
(
move
.
start
)
;
return
true
;
log
(
" meta:"
+
move
.
id
+
"/"
+
move
.
metaRootPos
+
" start: "
+
move
.
start
);
}
// change at least one entry in the map
private
void
copyLive
(
Chunk
chunk
,
ArrayList
<
Chunk
>
old
)
{
// to ensure a chunk will be written
ByteBuffer
buff
=
ByteBuffer
.
allocate
(
chunk
.
length
);
// (even if there is nothing to move)
try
{
meta
.
put
(
"chunk."
+
move
.
id
,
move
.
toString
());
file
.
position
(
chunk
.
start
);
MVMap
<
String
,
String
>
oldMeta
=
new
MVMap
<
String
,
String
>(
this
,
0
,
"old-meta"
,
STRING_TYPE
,
STRING_TYPE
,
0
);
DataUtils
.
readFully
(
file
,
buff
);
oldMeta
.
setRootPos
(
move
.
metaRootPos
);
}
catch
(
IOException
e
)
{
Iterator
<
String
>
it
=
oldMeta
.
keyIterator
(
"map."
);
throw
new
RuntimeException
(
e
);
while
(
it
.
hasNext
())
{
String
k
=
it
.
next
();
if
(!
k
.
startsWith
(
"map."
))
{
break
;
}
}
String
s
=
oldMeta
.
get
(
k
);
Chunk
.
fromHeader
(
buff
,
chunk
.
start
);
k
=
k
.
substring
(
"map."
.
length
());
int
chunkLength
=
chunk
.
length
;
while
(
buff
.
position
()
<
chunkLength
)
{
int
start
=
buff
.
position
();
int
pageLength
=
buff
.
getInt
();
buff
.
getShort
();
int
mapId
=
DataUtils
.
readVarInt
(
buff
);
@SuppressWarnings
(
"unchecked"
)
@SuppressWarnings
(
"unchecked"
)
MVMap
<
Object
,
Object
>
data
=
(
MVMap
<
Object
,
Object
>)
maps
.
get
(
k
);
MVMap
<
Object
,
Object
>
map
=
(
MVMap
<
Object
,
Object
>)
getMap
(
mapId
);
if
(
data
==
null
)
{
if
(
map
==
null
)
{
buff
.
position
(
start
+
pageLength
);
continue
;
continue
;
}
}
log
(
" "
+
k
+
" "
+
s
.
replace
(
'\n'
,
' '
));
buff
.
position
(
start
);
String
[]
idTypeList
=
StringUtils
.
arraySplit
(
s
,
'/'
,
false
);
Page
page
=
new
Page
(
map
,
0
);
int
id
=
Integer
.
parseInt
(
idTypeList
[
0
]);
page
.
read
(
buff
,
chunk
.
id
,
buff
.
position
(),
chunk
.
length
);
DataType
kt
=
buildDataType
(
idTypeList
[
3
]);
for
(
int
i
=
0
;
i
<
page
.
getKeyCount
();
i
++)
{
DataType
vt
=
buildDataType
(
idTypeList
[
4
]);
Object
k
=
page
.
getKey
(
i
);
long
oldDataRoot
=
Long
.
parseLong
(
oldMeta
.
get
(
"root."
+
id
));
Page
p
=
map
.
getPage
(
k
);
if
(
oldDataRoot
!=
0
)
{
MVMap
<?,
?>
oldData
=
new
MVMap
<
Object
,
Object
>(
this
,
id
,
"old-"
+
k
,
kt
,
vt
,
0
);
oldData
.
setRootPos
(
oldDataRoot
);
Iterator
<?>
dataIt
=
oldData
.
keyIterator
(
null
);
while
(
dataIt
.
hasNext
())
{
Object
o
=
dataIt
.
next
();
Page
p
=
data
.
getPage
(
o
);
if
(
p
==
null
)
{
if
(
p
==
null
)
{
// was removed later - ignore
// was removed later - ignore
// or the chunk no longer exists
// or the chunk no longer exists
...
@@ -864,17 +834,25 @@ public class MVStore {
...
@@ -864,17 +834,25 @@ public class MVStore {
}
else
{
}
else
{
Chunk
c
=
getChunk
(
p
.
getPos
());
Chunk
c
=
getChunk
(
p
.
getPos
());
if
(
old
.
contains
(
c
))
{
if
(
old
.
contains
(
c
))
{
log
(
" move key:"
+
o
+
" chunk:"
+
c
.
id
);
log
(
" move key:"
+
k
+
" chunk:"
+
c
.
id
);
Object
value
=
data
.
get
(
o
);
Object
value
=
map
.
remove
(
k
);
data
.
remove
(
o
);
map
.
put
(
k
,
value
);
data
.
put
(
o
,
value
);
}
}
}
}
}
}
}
}
}
}
store
();
return
true
;
private
MVMap
<?,
?>
getMap
(
int
mapId
)
{
if
(
mapId
==
0
)
{
return
meta
;
}
for
(
MVMap
<?,
?>
m
:
maps
.
values
())
{
if
(
m
.
getId
()
==
mapId
)
{
return
m
;
}
}
return
null
;
}
}
/**
/**
...
@@ -909,9 +887,6 @@ public class MVStore {
...
@@ -909,9 +887,6 @@ public class MVStore {
// but we don't optimize for rollback
// but we don't optimize for rollback
cache
.
remove
(
pos
);
cache
.
remove
(
pos
);
Chunk
c
=
getChunk
(
pos
);
Chunk
c
=
getChunk
(
pos
);
if
(
c
.
liveCount
==
0
)
{
throw
new
RuntimeException
(
"Negative live count: "
+
pos
);
}
HashMap
<
Integer
,
Chunk
>
freed
=
freedChunks
.
get
(
currentVersion
);
HashMap
<
Integer
,
Chunk
>
freed
=
freedChunks
.
get
(
currentVersion
);
if
(
freed
==
null
)
{
if
(
freed
==
null
)
{
freed
=
New
.
hashMap
();
freed
=
New
.
hashMap
();
...
@@ -922,7 +897,7 @@ public class MVStore {
...
@@ -922,7 +897,7 @@ public class MVStore {
f
=
new
Chunk
(
c
.
id
);
f
=
new
Chunk
(
c
.
id
);
freed
.
put
(
c
.
id
,
f
);
freed
.
put
(
c
.
id
,
f
);
}
}
f
.
liveCount
--
;
f
.
maxLengthLive
-=
DataUtils
.
getPageMaxLength
(
pos
)
;
}
}
}
}
...
@@ -993,6 +968,14 @@ public class MVStore {
...
@@ -993,6 +968,14 @@ public class MVStore {
return
retainVersion
;
return
retainVersion
;
}
}
/**
* Check whether all data can be read from this version. This requires that
* all chunks referenced by this version are still available (not
* overwritten).
*
* @param version the version
* @return true if all data can be read
*/
private
boolean
isKnownVersion
(
long
version
)
{
private
boolean
isKnownVersion
(
long
version
)
{
if
(
version
>
currentVersion
||
version
<
0
)
{
if
(
version
>
currentVersion
||
version
<
0
)
{
return
false
;
return
false
;
...
@@ -1006,7 +989,7 @@ public class MVStore {
...
@@ -1006,7 +989,7 @@ public class MVStore {
if
(
c
==
null
)
{
if
(
c
==
null
)
{
return
false
;
return
false
;
}
}
// also, all ch
eck
referenced by this version
// also, all ch
unks
referenced by this version
// need to be available in the file
// need to be available in the file
MVMap
<
String
,
String
>
oldMeta
=
getMetaMap
(
version
);
MVMap
<
String
,
String
>
oldMeta
=
getMetaMap
(
version
);
if
(
oldMeta
==
null
)
{
if
(
oldMeta
==
null
)
{
...
...
h2/src/tools/org/h2/dev/store/btree/Page.java
浏览文件 @
670d256c
...
@@ -57,7 +57,7 @@ public class Page {
...
@@ -57,7 +57,7 @@ public class Page {
private
Page
[]
childrenPages
;
private
Page
[]
childrenPages
;
private
long
[]
counts
;
private
long
[]
counts
;
private
Page
(
MVMap
<?,
?>
map
,
long
version
)
{
Page
(
MVMap
<?,
?>
map
,
long
version
)
{
this
.
map
=
map
;
this
.
map
=
map
;
this
.
version
=
version
;
this
.
version
=
version
;
}
}
...
@@ -596,7 +596,7 @@ public class Page {
...
@@ -596,7 +596,7 @@ public class Page {
}
}
}
}
private
void
read
(
ByteBuffer
buff
,
int
chunkId
,
int
offset
,
int
maxLength
)
{
void
read
(
ByteBuffer
buff
,
int
chunkId
,
int
offset
,
int
maxLength
)
{
int
start
=
buff
.
position
();
int
start
=
buff
.
position
();
int
pageLength
=
buff
.
getInt
();
int
pageLength
=
buff
.
getInt
();
if
(
pageLength
>
maxLength
)
{
if
(
pageLength
>
maxLength
)
{
...
@@ -661,10 +661,10 @@ public class Page {
...
@@ -661,10 +661,10 @@ public class Page {
/**
/**
* Store the page and update the position.
* Store the page and update the position.
*
*
* @param chunk the chunk
* @param buff the target buffer
* @param buff the target buffer
* @param chunkId the chunk id
*/
*/
private
void
write
(
ByteBuffer
buff
,
int
chunkId
)
{
private
void
write
(
Chunk
chunk
,
ByteBuffer
buff
)
{
int
start
=
buff
.
position
();
int
start
=
buff
.
position
();
buff
.
putInt
(
0
);
buff
.
putInt
(
0
);
buff
.
putShort
((
byte
)
0
);
buff
.
putShort
((
byte
)
0
);
...
@@ -707,11 +707,15 @@ public class Page {
...
@@ -707,11 +707,15 @@ public class Page {
}
}
int
pageLength
=
buff
.
position
()
-
start
;
int
pageLength
=
buff
.
position
()
-
start
;
buff
.
putInt
(
start
,
pageLength
);
buff
.
putInt
(
start
,
pageLength
);
int
chunkId
=
chunk
.
id
;
int
check
=
DataUtils
.
getCheckValue
(
chunkId
)
int
check
=
DataUtils
.
getCheckValue
(
chunkId
)
^
DataUtils
.
getCheckValue
(
start
)
^
DataUtils
.
getCheckValue
(
start
)
^
DataUtils
.
getCheckValue
(
pageLength
);
^
DataUtils
.
getCheckValue
(
pageLength
);
buff
.
putShort
(
start
+
4
,
(
short
)
check
);
buff
.
putShort
(
start
+
4
,
(
short
)
check
);
this
.
pos
=
DataUtils
.
getPagePos
(
chunkId
,
start
,
pageLength
,
type
);
this
.
pos
=
DataUtils
.
getPagePos
(
chunkId
,
start
,
pageLength
,
type
);
long
max
=
DataUtils
.
getPageMaxLength
(
pos
);
chunk
.
maxLength
+=
max
;
chunk
.
maxLengthLive
+=
max
;
}
}
/**
/**
...
@@ -748,42 +752,25 @@ public class Page {
...
@@ -748,42 +752,25 @@ public class Page {
* Store this page and all children that are changed, in reverse order, and
* Store this page and all children that are changed, in reverse order, and
* update the position and the children.
* update the position and the children.
*
*
* @param chunk the chunk
* @param buff the target buffer
* @param buff the target buffer
* @param chunkId the chunk id
* @return the page id
* @return the page id
*/
*/
long
writeTempRecursive
(
ByteBuffer
buff
,
int
chunkId
)
{
long
writeTempRecursive
(
Chunk
chunk
,
ByteBuffer
buff
)
{
if
(!
isLeaf
())
{
if
(!
isLeaf
())
{
int
len
=
children
.
length
;
int
len
=
children
.
length
;
for
(
int
i
=
0
;
i
<
len
;
i
++)
{
for
(
int
i
=
0
;
i
<
len
;
i
++)
{
Page
p
=
childrenPages
[
i
];
Page
p
=
childrenPages
[
i
];
if
(
p
!=
null
)
{
if
(
p
!=
null
)
{
children
[
i
]
=
p
.
writeTempRecursive
(
buff
,
chunkId
);
children
[
i
]
=
p
.
writeTempRecursive
(
chunk
,
buff
);
childrenPages
[
i
]
=
null
;
childrenPages
[
i
]
=
null
;
}
}
}
}
}
}
write
(
buff
,
chunkId
);
write
(
chunk
,
buff
);
return
pos
;
return
pos
;
}
}
/**
* Count the temporary pages recursively.
*
* @return the number of temporary pages
*/
int
countTempRecursive
()
{
int
count
=
1
;
if
(!
isLeaf
())
{
for
(
Page
p
:
childrenPages
)
{
if
(
p
!=
null
)
{
count
+=
p
.
countTempRecursive
();
}
}
}
return
count
;
}
long
getVersion
()
{
long
getVersion
()
{
return
version
;
return
version
;
}
}
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论