Skip to content
项目
群组
代码片段
帮助
正在加载...
帮助
为 GitLab 提交贡献
登录/注册
切换导航
H
h2database
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
分枝图
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
计划
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
分枝图
统计图
创建新议题
作业
提交
议题看板
打开侧边栏
Administrator
h2database
Commits
d3bd6b1d
提交
d3bd6b1d
authored
7月 22, 2015
作者:
Thomas Mueller
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
MVStore: power failure could corrupt the store, if writes were re-ordered.
上级
ba429036
显示空白字符变更
内嵌
并排
正在显示
6 个修改的文件
包含
288 行增加
和
152 行删除
+288
-152
changelog.html
h2/src/docsrc/html/changelog.html
+2
-0
MVStore.java
h2/src/main/org/h2/mvstore/MVStore.java
+224
-128
TestAll.java
h2/src/test/org/h2/test/TestAll.java
+1
-0
TestReorderWrites.java
h2/src/test/org/h2/test/poweroff/TestReorderWrites.java
+47
-14
TestMVStore.java
h2/src/test/org/h2/test/store/TestMVStore.java
+7
-4
FilePathReorderWrites.java
h2/src/test/org/h2/test/utils/FilePathReorderWrites.java
+7
-6
没有找到文件。
h2/src/docsrc/html/changelog.html
浏览文件 @
d3bd6b1d
...
@@ -21,6 +21,8 @@ Change Log
...
@@ -21,6 +21,8 @@ Change Log
<h2>
Next Version (unreleased)
</h2>
<h2>
Next Version (unreleased)
</h2>
<ul>
<ul>
<li>
MVStore: power failure could corrupt the store, if writes were re-ordered.
</li>
<li>
For compatibility with other databases, support for (double and float)
<li>
For compatibility with other databases, support for (double and float)
-0.0 has been removed. 0.0 is used instead.
-0.0 has been removed. 0.0 is used instead.
</li>
</li>
...
...
h2/src/main/org/h2/mvstore/MVStore.java
浏览文件 @
d3bd6b1d
...
@@ -22,9 +22,9 @@ import java.util.concurrent.ConcurrentHashMap;
...
@@ -22,9 +22,9 @@ import java.util.concurrent.ConcurrentHashMap;
import
org.h2.compress.CompressDeflate
;
import
org.h2.compress.CompressDeflate
;
import
org.h2.compress.CompressLZF
;
import
org.h2.compress.CompressLZF
;
import
org.h2.compress.Compressor
;
import
org.h2.compress.Compressor
;
import
org.h2.mvstore.Page.PageChildren
;
import
org.h2.mvstore.cache.CacheLongKeyLIRS
;
import
org.h2.mvstore.cache.CacheLongKeyLIRS
;
import
org.h2.mvstore.type.StringDataType
;
import
org.h2.mvstore.type.StringDataType
;
import
org.h2.mvstore.Page.PageChildren
;
import
org.h2.util.MathUtils
;
import
org.h2.util.MathUtils
;
import
org.h2.util.New
;
import
org.h2.util.New
;
...
@@ -240,6 +240,13 @@ public class MVStore {
...
@@ -240,6 +240,13 @@ public class MVStore {
* The time the store was created, in milliseconds since 1970.
* The time the store was created, in milliseconds since 1970.
*/
*/
private
long
creationTime
;
private
long
creationTime
;
/**
* How long to retain old, persisted chunks, in milliseconds. For larger or
* equal to zero, a chunk is never directly overwritten if unused, but
* instead, the unused field is set. If smaller zero, chunks are directly
* overwritten if unused.
*/
private
int
retentionTime
;
private
int
retentionTime
;
private
long
lastCommitTime
;
private
long
lastCommitTime
;
...
@@ -270,6 +277,8 @@ public class MVStore {
...
@@ -270,6 +277,8 @@ public class MVStore {
private
IllegalStateException
panicException
;
private
IllegalStateException
panicException
;
private
long
lastTimeAbsolute
;
/**
/**
* Create and open the store.
* Create and open the store.
*
*
...
@@ -337,7 +346,7 @@ public class MVStore {
...
@@ -337,7 +346,7 @@ public class MVStore {
fileStore
.
open
(
fileName
,
readOnly
,
encryptionKey
);
fileStore
.
open
(
fileName
,
readOnly
,
encryptionKey
);
}
}
if
(
fileStore
.
size
()
==
0
)
{
if
(
fileStore
.
size
()
==
0
)
{
creationTime
=
getTime
();
creationTime
=
getTime
Absolute
();
lastCommitTime
=
creationTime
;
lastCommitTime
=
creationTime
;
storeHeader
.
put
(
"H"
,
2
);
storeHeader
.
put
(
"H"
,
2
);
storeHeader
.
put
(
"blockSize"
,
BLOCK_SIZE
);
storeHeader
.
put
(
"blockSize"
,
BLOCK_SIZE
);
...
@@ -354,7 +363,7 @@ public class MVStore {
...
@@ -354,7 +363,7 @@ public class MVStore {
Arrays
.
fill
(
encryptionKey
,
(
char
)
0
);
Arrays
.
fill
(
encryptionKey
,
(
char
)
0
);
}
}
}
}
lastCommitTime
=
getTime
();
lastCommitTime
=
getTime
SinceCreation
();
// setAutoCommitDelay starts the thread, but only if
// setAutoCommitDelay starts the thread, but only if
// the parameter is different from the old value
// the parameter is different from the old value
...
@@ -520,13 +529,15 @@ public class MVStore {
...
@@ -520,13 +529,15 @@ public class MVStore {
}
}
private
Chunk
getChunkForVersion
(
long
version
)
{
private
Chunk
getChunkForVersion
(
long
version
)
{
Chunk
c
=
lastChunk
;
Chunk
newest
=
null
;
while
(
true
)
{
for
(
Chunk
c
:
chunks
.
values
())
{
if
(
c
==
null
||
c
.
version
<=
version
)
{
if
(
c
.
version
<=
version
)
{
return
c
;
if
(
newest
==
null
||
c
.
id
>
newest
.
id
)
{
newest
=
c
;
}
}
}
c
=
chunks
.
get
(
c
.
id
-
1
);
}
}
return
newest
;
}
}
/**
/**
...
@@ -546,10 +557,9 @@ public class MVStore {
...
@@ -546,10 +557,9 @@ public class MVStore {
}
}
private
synchronized
void
readStoreHeader
()
{
private
synchronized
void
readStoreHeader
()
{
boolean
validHeader
=
false
;
Chunk
newest
=
null
;
// we don't know yet which chunk and version are the newest
boolean
validStoreHeader
=
false
;
long
newestVersion
=
-
1
;
// find out which chunk and version are the newest
long
chunkBlock
=
-
1
;
// read the first two blocks
// read the first two blocks
ByteBuffer
fileHeaderBlocks
=
fileStore
.
readFully
(
0
,
2
*
BLOCK_SIZE
);
ByteBuffer
fileHeaderBlocks
=
fileStore
.
readFully
(
0
,
2
*
BLOCK_SIZE
);
byte
[]
buff
=
new
byte
[
BLOCK_SIZE
];
byte
[]
buff
=
new
byte
[
BLOCK_SIZE
];
...
@@ -578,18 +588,22 @@ public class MVStore {
...
@@ -578,18 +588,22 @@ public class MVStore {
continue
;
continue
;
}
}
long
version
=
DataUtils
.
readHexLong
(
m
,
"version"
,
0
);
long
version
=
DataUtils
.
readHexLong
(
m
,
"version"
,
0
);
if
(
version
>
newestV
ersion
)
{
if
(
newest
==
null
||
version
>
newest
.
v
ersion
)
{
newestVersion
=
version
;
validStoreHeader
=
true
;
storeHeader
.
putAll
(
m
);
storeHeader
.
putAll
(
m
);
chunkBlock
=
DataUtils
.
readHexLong
(
m
,
"block"
,
0
);
creationTime
=
DataUtils
.
readHexLong
(
m
,
"created"
,
0
);
creationTime
=
DataUtils
.
readHexLong
(
m
,
"created"
,
0
);
validHeader
=
true
;
int
chunkId
=
DataUtils
.
readHexInt
(
m
,
"chunk"
,
0
);
long
block
=
DataUtils
.
readHexLong
(
m
,
"block"
,
0
);
Chunk
test
=
readChunkHeaderAndFooter
(
block
);
if
(
test
!=
null
&&
test
.
id
==
chunkId
)
{
newest
=
test
;
}
}
}
}
catch
(
Exception
e
)
{
}
catch
(
Exception
e
)
{
continue
;
continue
;
}
}
}
}
if
(!
validHeader
)
{
if
(!
valid
Store
Header
)
{
throw
DataUtils
.
newIllegalStateException
(
throw
DataUtils
.
newIllegalStateException
(
DataUtils
.
ERROR_FILE_CORRUPT
,
DataUtils
.
ERROR_FILE_CORRUPT
,
"Store header is corrupt: {0}"
,
fileStore
);
"Store header is corrupt: {0}"
,
fileStore
);
...
@@ -629,58 +643,51 @@ public class MVStore {
...
@@ -629,58 +643,51 @@ public class MVStore {
creationTime
=
now
;
creationTime
=
now
;
storeHeader
.
put
(
"created"
,
creationTime
);
storeHeader
.
put
(
"created"
,
creationTime
);
}
}
Chunk
test
=
readChunkFooter
(
fileStore
.
size
());
Chunk
footer
=
readChunkFooter
(
fileStore
.
size
());
if
(
test
!=
null
)
{
if
(
footer
!=
null
)
{
test
=
readChunkHeaderAndFooter
(
test
.
block
);
if
(
footer
.
version
>
newestVersion
)
{
if
(
test
!=
null
)
{
newestVersion
=
footer
.
version
;
if
(
newest
==
null
||
test
.
version
>
newest
.
version
)
{
chunkBlock
=
footer
.
block
;
newest
=
test
;
}
}
}
}
if
(
chunkBlock
<=
0
)
{
}
if
(
newest
==
null
)
{
// no chunk
// no chunk
return
;
return
;
}
}
// read the chunk header and footer,
// read the chunk header and footer,
// and follow the chain of next chunks
// and follow the chain of next chunks
lastChunk
=
null
;
while
(
true
)
{
while
(
true
)
{
Chunk
header
;
if
(
newest
.
next
==
0
||
try
{
newest
.
next
>=
fileStore
.
size
()
/
BLOCK_SIZE
)
{
header
=
readChunkHeader
(
chunkBlock
);
// no (valid) next
}
catch
(
Exception
e
)
{
// invalid chunk header: ignore, but stop
break
;
break
;
}
}
if
(
header
.
version
<
newestVersion
)
{
test
=
readChunkHeaderAndFooter
(
newest
.
next
);
// we have reached the end
if
(
test
==
null
||
test
.
id
<=
newest
.
id
)
{
break
;
break
;
}
}
footer
=
readChunkFooter
((
chunkBlock
+
header
.
len
)
*
BLOCK_SIZE
);
newest
=
test
;
if
(
footer
==
null
||
footer
.
id
!=
header
.
id
)
{
// invalid chunk footer, or the wrong one
break
;
}
}
lastChunk
=
header
;
setLastChunk
(
newest
);
newestVersion
=
header
.
version
;
loadChunkMeta
();
if
(
header
.
next
==
0
||
// read all chunk headers and footers within the retention time,
header
.
next
>=
fileStore
.
size
()
/
BLOCK_SIZE
)
{
// to detect unwritten data after a power failure
// no (valid) next
verifyLastChunks
();
break
;
// build the free space list
for
(
Chunk
c
:
chunks
.
values
())
{
if
(
c
.
pageCountLive
==
0
)
{
// remove this chunk in the next save operation
registerFreePage
(
currentVersion
,
c
.
id
,
0
,
0
);
}
}
chunkBlock
=
header
.
next
;
long
start
=
c
.
block
*
BLOCK_SIZE
;
int
length
=
c
.
len
*
BLOCK_SIZE
;
fileStore
.
markUsed
(
start
,
length
);
}
}
if
(
lastChunk
==
null
)
{
// no valid chunk
return
;
}
}
lastMapId
=
lastChunk
.
mapId
;
currentVersion
=
lastChunk
.
version
;
setWriteVersion
(
currentVersion
);
chunks
.
put
(
lastChunk
.
id
,
lastChunk
);
meta
.
setRootPos
(
lastChunk
.
metaRootPos
,
-
1
);
private
void
loadChunkMeta
()
{
// load the chunk metadata: we can load in any order,
// load the chunk metadata: we can load in any order,
// because loading chunk metadata might recursively load another chunk
// because loading chunk metadata might recursively load another chunk
for
(
Iterator
<
String
>
it
=
meta
.
keyIterator
(
"chunk."
);
it
.
hasNext
();)
{
for
(
Iterator
<
String
>
it
=
meta
.
keyIterator
(
"chunk."
);
it
.
hasNext
();)
{
...
@@ -699,18 +706,83 @@ public class MVStore {
...
@@ -699,18 +706,83 @@ public class MVStore {
chunks
.
put
(
c
.
id
,
c
);
chunks
.
put
(
c
.
id
,
c
);
}
}
}
}
// build the free space list
for
(
Chunk
c
:
chunks
.
values
())
{
if
(
c
.
pageCountLive
==
0
)
{
// remove this chunk in the next save operation
registerFreePage
(
currentVersion
,
c
.
id
,
0
,
0
);
}
}
long
start
=
c
.
block
*
BLOCK_SIZE
;
int
length
=
c
.
len
*
BLOCK_SIZE
;
private
void
setLastChunk
(
Chunk
last
)
{
fileStore
.
markUsed
(
start
,
length
);
lastChunk
=
last
;
if
(
last
==
null
)
{
// no valid chunk
lastMapId
=
0
;
currentVersion
=
0
;
meta
.
setRootPos
(
0
,
-
1
);
}
else
{
lastMapId
=
last
.
mapId
;
currentVersion
=
last
.
version
;
chunks
.
put
(
last
.
id
,
last
);
meta
.
setRootPos
(
last
.
metaRootPos
,
-
1
);
}
setWriteVersion
(
currentVersion
);
}
private
void
verifyLastChunks
()
{
long
time
=
getTimeSinceCreation
();
ArrayList
<
Integer
>
ids
=
new
ArrayList
<
Integer
>(
chunks
.
keySet
());
Collections
.
sort
(
ids
);
int
newestValidChunk
=
-
1
;
Chunk
old
=
null
;
for
(
Integer
chunkId
:
ids
)
{
Chunk
c
=
chunks
.
get
(
chunkId
);
if
(
old
!=
null
&&
c
.
time
<
old
.
time
)
{
// old chunk (maybe leftover from a previous crash)
break
;
}
old
=
c
;
if
(
c
.
time
+
retentionTime
<
time
)
{
// old chunk, no need to verify
newestValidChunk
=
c
.
id
;
continue
;
}
Chunk
test
=
readChunkHeaderAndFooter
(
c
.
block
);
if
(
test
==
null
||
test
.
id
!=
c
.
id
)
{
break
;
}
newestValidChunk
=
chunkId
;
}
Chunk
newest
=
chunks
.
get
(
newestValidChunk
);
if
(
newest
!=
lastChunk
)
{
// to avoid re-using newer chunks later on, we could clear
// the headers and footers of those, but we might not know about all
// of them, so that could be incomplete - but we check that newer
// chunks are written after older chunks, so we are safe
rollbackTo
(
newest
==
null
?
0
:
newest
.
version
);
}
}
}
}
/**
* Read a chunk header and footer, and verify the stored data is consistent.
*
* @param block the block
* @return the chunk, or null if the header or footer don't match or are not
* consistent
*/
private
Chunk
readChunkHeaderAndFooter
(
long
block
)
{
Chunk
header
;
try
{
header
=
readChunkHeader
(
block
);
}
catch
(
Exception
e
)
{
// invalid chunk header: ignore, but stop
return
null
;
}
if
(
header
==
null
)
{
return
null
;
}
Chunk
footer
=
readChunkFooter
((
block
+
header
.
len
)
*
BLOCK_SIZE
);
if
(
footer
==
null
||
footer
.
id
!=
header
.
id
)
{
return
null
;
}
return
header
;
}
/**
/**
* Try to read a chunk footer.
* Try to read a chunk footer.
*
*
...
@@ -782,7 +854,8 @@ public class MVStore {
...
@@ -782,7 +854,8 @@ public class MVStore {
if
(
closed
)
{
if
(
closed
)
{
return
;
return
;
}
}
if
(
fileStore
!=
null
&&
!
fileStore
.
isReadOnly
())
{
FileStore
f
=
fileStore
;
if
(
f
!=
null
&&
!
f
.
isReadOnly
())
{
stopBackgroundThread
();
stopBackgroundThread
();
if
(
hasUnsavedChanges
())
{
if
(
hasUnsavedChanges
())
{
commitAndSave
();
commitAndSave
();
...
@@ -979,11 +1052,12 @@ public class MVStore {
...
@@ -979,11 +1052,12 @@ public class MVStore {
private
long
storeNowTry
()
{
private
long
storeNowTry
()
{
freeUnusedChunks
();
freeUnusedChunks
();
int
currentUnsavedPageCount
=
unsavedMemory
;
int
currentUnsavedPageCount
=
unsavedMemory
;
long
storeVersion
=
currentStoreVersion
;
long
storeVersion
=
currentStoreVersion
;
long
version
=
++
currentVersion
;
long
version
=
++
currentVersion
;
setWriteVersion
(
version
);
setWriteVersion
(
version
);
long
time
=
getTime
();
long
time
=
getTime
SinceCreation
();
lastCommitTime
=
time
;
lastCommitTime
=
time
;
retainChunk
=
null
;
retainChunk
=
null
;
...
@@ -1166,7 +1240,6 @@ public class MVStore {
...
@@ -1166,7 +1240,6 @@ public class MVStore {
// may only shrink after the store header was written
// may only shrink after the store header was written
shrinkFileIfPossible
(
1
);
shrinkFileIfPossible
(
1
);
}
}
for
(
MVMap
<?,
?>
m
:
changed
)
{
for
(
MVMap
<?,
?>
m
:
changed
)
{
Page
p
=
m
.
getRoot
();
Page
p
=
m
.
getRoot
();
if
(
p
.
getTotalCount
()
>
0
)
{
if
(
p
.
getTotalCount
()
>
0
)
{
...
@@ -1192,7 +1265,7 @@ public class MVStore {
...
@@ -1192,7 +1265,7 @@ public class MVStore {
}
}
Set
<
Integer
>
referenced
=
collectReferencedChunks
();
Set
<
Integer
>
referenced
=
collectReferencedChunks
();
ArrayList
<
Chunk
>
free
=
New
.
arrayList
();
ArrayList
<
Chunk
>
free
=
New
.
arrayList
();
long
time
=
getTime
();
long
time
=
getTime
SinceCreation
();
for
(
Chunk
c
:
chunks
.
values
())
{
for
(
Chunk
c
:
chunks
.
values
())
{
if
(!
referenced
.
contains
(
c
.
id
))
{
if
(!
referenced
.
contains
(
c
.
id
))
{
free
.
add
(
c
);
free
.
add
(
c
);
...
@@ -1347,12 +1420,14 @@ public class MVStore {
...
@@ -1347,12 +1420,14 @@ public class MVStore {
}
}
private
boolean
canOverwriteChunk
(
Chunk
c
,
long
time
)
{
private
boolean
canOverwriteChunk
(
Chunk
c
,
long
time
)
{
if
(
retentionTime
>=
0
)
{
if
(
c
.
time
+
retentionTime
>
time
)
{
if
(
c
.
time
+
retentionTime
>
time
)
{
return
false
;
return
false
;
}
}
if
(
c
.
unused
==
0
||
c
.
unused
+
retentionTime
/
2
>
time
)
{
if
(
c
.
unused
==
0
||
c
.
unused
+
retentionTime
/
2
>
time
)
{
return
false
;
return
false
;
}
}
}
Chunk
r
=
retainChunk
;
Chunk
r
=
retainChunk
;
if
(
r
!=
null
&&
c
.
version
>
r
.
version
)
{
if
(
r
!=
null
&&
c
.
version
>
r
.
version
)
{
return
false
;
return
false
;
...
@@ -1360,8 +1435,21 @@ public class MVStore {
...
@@ -1360,8 +1435,21 @@ public class MVStore {
return
true
;
return
true
;
}
}
private
long
getTime
()
{
private
long
getTimeSinceCreation
()
{
return
System
.
currentTimeMillis
()
-
creationTime
;
return
Math
.
max
(
0
,
getTimeAbsolute
()
-
creationTime
);
}
private
long
getTimeAbsolute
()
{
long
now
=
System
.
currentTimeMillis
();
if
(
lastTimeAbsolute
!=
0
&&
now
<
lastTimeAbsolute
)
{
// time seems to have run backwards - this can happen
// when the system time is adjusted, for example
// on a leap second
now
=
lastTimeAbsolute
;
}
else
{
lastTimeAbsolute
=
now
;
}
return
now
;
}
}
/**
/**
...
@@ -1394,20 +1482,12 @@ public class MVStore {
...
@@ -1394,20 +1482,12 @@ public class MVStore {
c
.
maxLenLive
+=
f
.
maxLenLive
;
c
.
maxLenLive
+=
f
.
maxLenLive
;
c
.
pageCountLive
+=
f
.
pageCountLive
;
c
.
pageCountLive
+=
f
.
pageCountLive
;
if
(
c
.
pageCountLive
<
0
&&
c
.
pageCountLive
>
-
MARKED_FREE
)
{
if
(
c
.
pageCountLive
<
0
&&
c
.
pageCountLive
>
-
MARKED_FREE
)
{
throw
DataUtils
.
newIllegalStateException
(
// can happen after a rollback
DataUtils
.
ERROR_INTERNAL
,
c
.
pageCountLive
=
0
;
"Corrupt page count {0}"
,
c
.
pageCountLive
);
}
}
if
(
c
.
maxLenLive
<
0
&&
c
.
maxLenLive
>
-
MARKED_FREE
)
{
if
(
c
.
maxLenLive
<
0
&&
c
.
maxLenLive
>
-
MARKED_FREE
)
{
throw
DataUtils
.
newIllegalStateException
(
// can happen after a rollback
DataUtils
.
ERROR_INTERNAL
,
c
.
maxLenLive
=
0
;
"Corrupt max length {0}"
,
c
.
maxLenLive
);
}
if
(
c
.
pageCountLive
<=
0
&&
c
.
maxLenLive
>
0
||
c
.
maxLenLive
<=
0
&&
c
.
pageCountLive
>
0
)
{
throw
DataUtils
.
newIllegalStateException
(
DataUtils
.
ERROR_INTERNAL
,
"Corrupt max length {0}"
,
c
.
maxLenLive
);
}
}
modified
.
add
(
c
);
modified
.
add
(
c
);
}
}
...
@@ -1441,6 +1521,9 @@ public class MVStore {
...
@@ -1441,6 +1521,9 @@ public class MVStore {
if
(
savedPercent
<
minPercent
)
{
if
(
savedPercent
<
minPercent
)
{
return
;
return
;
}
}
if
(!
closed
)
{
sync
();
}
fileStore
.
truncate
(
end
);
fileStore
.
truncate
(
end
);
}
}
...
@@ -1548,7 +1631,7 @@ public class MVStore {
...
@@ -1548,7 +1631,7 @@ public class MVStore {
int
oldRetentionTime
=
retentionTime
;
int
oldRetentionTime
=
retentionTime
;
boolean
oldReuse
=
reuseSpace
;
boolean
oldReuse
=
reuseSpace
;
try
{
try
{
retentionTime
=
0
;
retentionTime
=
-
1
;
freeUnusedChunks
();
freeUnusedChunks
();
if
(
fileStore
.
getFillRate
()
>
targetFillRate
)
{
if
(
fileStore
.
getFillRate
()
>
targetFillRate
)
{
return
false
;
return
false
;
...
@@ -1618,7 +1701,7 @@ public class MVStore {
...
@@ -1618,7 +1701,7 @@ public class MVStore {
buff
.
position
(
0
);
buff
.
position
(
0
);
c
.
writeChunkHeader
(
buff
,
chunkHeaderLen
);
c
.
writeChunkHeader
(
buff
,
chunkHeaderLen
);
buff
.
position
(
length
-
Chunk
.
FOOTER_LENGTH
);
buff
.
position
(
length
-
Chunk
.
FOOTER_LENGTH
);
buff
.
put
(
lastChunk
.
getFooterBytes
());
buff
.
put
(
c
.
getFooterBytes
());
buff
.
position
(
0
);
buff
.
position
(
0
);
write
(
end
,
buff
.
getBuffer
());
write
(
end
,
buff
.
getBuffer
());
releaseWriteBuffer
(
buff
);
releaseWriteBuffer
(
buff
);
...
@@ -1629,7 +1712,6 @@ public class MVStore {
...
@@ -1629,7 +1712,6 @@ public class MVStore {
// update the metadata (store at the end of the file)
// update the metadata (store at the end of the file)
reuseSpace
=
false
;
reuseSpace
=
false
;
commitAndSave
();
commitAndSave
();
sync
();
sync
();
// now re-use the empty space
// now re-use the empty space
...
@@ -1655,7 +1737,7 @@ public class MVStore {
...
@@ -1655,7 +1737,7 @@ public class MVStore {
c
.
block
=
pos
/
BLOCK_SIZE
;
c
.
block
=
pos
/
BLOCK_SIZE
;
c
.
writeChunkHeader
(
buff
,
chunkHeaderLen
);
c
.
writeChunkHeader
(
buff
,
chunkHeaderLen
);
buff
.
position
(
length
-
Chunk
.
FOOTER_LENGTH
);
buff
.
position
(
length
-
Chunk
.
FOOTER_LENGTH
);
buff
.
put
(
lastChunk
.
getFooterBytes
());
buff
.
put
(
c
.
getFooterBytes
());
buff
.
position
(
0
);
buff
.
position
(
0
);
write
(
pos
,
buff
.
getBuffer
());
write
(
pos
,
buff
.
getBuffer
());
releaseWriteBuffer
(
buff
);
releaseWriteBuffer
(
buff
);
...
@@ -1674,7 +1756,11 @@ public class MVStore {
...
@@ -1674,7 +1756,11 @@ public class MVStore {
* implementation calls FileChannel.force(true).
* implementation calls FileChannel.force(true).
*/
*/
public
void
sync
()
{
public
void
sync
()
{
fileStore
.
sync
();
checkOpen
();
FileStore
f
=
fileStore
;
if
(
f
!=
null
)
{
f
.
sync
();
}
}
}
/**
/**
...
@@ -1723,7 +1809,7 @@ public class MVStore {
...
@@ -1723,7 +1809,7 @@ public class MVStore {
long
maxLengthSum
=
0
;
long
maxLengthSum
=
0
;
long
maxLengthLiveSum
=
0
;
long
maxLengthLiveSum
=
0
;
long
time
=
getTime
();
long
time
=
getTime
SinceCreation
();
for
(
Chunk
c
:
chunks
.
values
())
{
for
(
Chunk
c
:
chunks
.
values
())
{
// ignore young chunks, because we don't optimize those
// ignore young chunks, because we don't optimize those
...
@@ -2061,6 +2147,7 @@ public class MVStore {
...
@@ -2061,6 +2147,7 @@ public class MVStore {
if
(
oldMeta
==
null
)
{
if
(
oldMeta
==
null
)
{
return
false
;
return
false
;
}
}
try
{
for
(
Iterator
<
String
>
it
=
oldMeta
.
keyIterator
(
"chunk."
);
for
(
Iterator
<
String
>
it
=
oldMeta
.
keyIterator
(
"chunk."
);
it
.
hasNext
();)
{
it
.
hasNext
();)
{
String
chunkKey
=
it
.
next
();
String
chunkKey
=
it
.
next
();
...
@@ -2068,8 +2155,19 @@ public class MVStore {
...
@@ -2068,8 +2155,19 @@ public class MVStore {
break
;
break
;
}
}
if
(!
meta
.
containsKey
(
chunkKey
))
{
if
(!
meta
.
containsKey
(
chunkKey
))
{
String
s
=
oldMeta
.
get
(
chunkKey
);
Chunk
c2
=
Chunk
.
fromString
(
s
);
Chunk
test
=
readChunkHeaderAndFooter
(
c2
.
block
);
if
(
test
==
null
||
test
.
id
!=
c2
.
id
)
{
return
false
;
return
false
;
}
}
// we store this chunk
chunks
.
put
(
c2
.
id
,
c2
);
}
}
}
catch
(
IllegalStateException
e
)
{
// the chunk missing where the metadata is stored
return
false
;
}
}
return
true
;
return
true
;
}
}
...
@@ -2183,43 +2281,43 @@ public class MVStore {
...
@@ -2183,43 +2281,43 @@ public class MVStore {
meta
.
rollbackTo
(
version
);
meta
.
rollbackTo
(
version
);
metaChanged
=
false
;
metaChanged
=
false
;
boolean
loadFromFile
=
false
;
boolean
loadFromFile
=
false
;
// get the largest chunk with a version
// find out which chunks to remove,
// higher or equal the requested version
// and which is the newest chunk to keep
Chunk
removeChunksNewerThan
=
null
;
// (the chunk list can have gaps)
Chunk
c
=
lastChunk
;
ArrayList
<
Integer
>
remove
=
new
ArrayList
<
Integer
>();
while
(
true
)
{
Chunk
keep
=
null
;
if
(
c
==
null
||
c
.
version
<
version
)
{
for
(
Chunk
c
:
chunks
.
values
())
{
break
;
if
(
c
.
version
>
version
)
{
remove
.
add
(
c
.
id
);
}
else
if
(
keep
==
null
||
keep
.
id
<
c
.
id
)
{
keep
=
c
;
}
}
removeChunksNewerThan
=
c
;
c
=
chunks
.
get
(
c
.
id
-
1
);
}
}
Chunk
last
=
lastChunk
;
if
(
remove
.
size
()
>
0
)
{
if
(
removeChunksNewerThan
!=
null
&&
// remove the youngest first, so we don't create gaps
last
.
version
>
removeChunksNewerThan
.
version
)
{
// (in case we remove many chunks)
Collections
.
sort
(
remove
,
Collections
.
reverseOrder
());
revertTemp
(
version
);
revertTemp
(
version
);
loadFromFile
=
true
;
loadFromFile
=
true
;
while
(
true
)
{
for
(
int
id
:
remove
)
{
last
=
lastChunk
;
Chunk
c
=
chunks
.
remove
(
id
);
if
(
last
==
null
)
{
long
start
=
c
.
block
*
BLOCK_SIZE
;
break
;
int
length
=
c
.
len
*
BLOCK_SIZE
;
}
else
if
(
last
.
version
<=
removeChunksNewerThan
.
version
)
{
break
;
}
chunks
.
remove
(
lastChunk
.
id
);
long
start
=
last
.
block
*
BLOCK_SIZE
;
int
length
=
last
.
len
*
BLOCK_SIZE
;
fileStore
.
free
(
start
,
length
);
fileStore
.
free
(
start
,
length
);
//
need to
overwrite the chunk,
// overwrite the chunk,
// so it
can not be used
// so it
is not be used later on
WriteBuffer
buff
=
getWriteBuffer
();
WriteBuffer
buff
=
getWriteBuffer
();
buff
.
limit
(
length
);
buff
.
limit
(
length
);
// buff.clear() does not set the data
// buff.clear() does not set the data
Arrays
.
fill
(
buff
.
getBuffer
().
array
(),
(
byte
)
0
);
Arrays
.
fill
(
buff
.
getBuffer
().
array
(),
(
byte
)
0
);
write
(
start
,
buff
.
getBuffer
());
write
(
start
,
buff
.
getBuffer
());
releaseWriteBuffer
(
buff
);
releaseWriteBuffer
(
buff
);
lastChunk
=
chunks
.
get
(
lastChunk
.
id
-
1
);
// only really needed if we remove many chunks, when writes are
// re-ordered - but we do it always, because rollback is not
// performance criticial
sync
();
}
}
lastChunk
=
keep
;
writeStoreHeader
();
writeStoreHeader
();
readStoreHeader
();
readStoreHeader
();
}
}
...
@@ -2233,12 +2331,10 @@ public class MVStore {
...
@@ -2233,12 +2331,10 @@ public class MVStore {
m
.
setRootPos
(
getRootPos
(
meta
,
id
),
-
1
);
m
.
setRootPos
(
getRootPos
(
meta
,
id
),
-
1
);
}
}
}
}
}
}
// rollback might have rolled back the stored chunk metadata as well
// rollback might have rolled back the stored chunk metadata as well
if
(
lastChunk
!=
null
)
{
if
(
lastChunk
!=
null
)
{
c
=
chunks
.
get
(
lastChunk
.
id
-
1
);
for
(
Chunk
c
:
chunks
.
values
())
{
if
(
c
!=
null
)
{
meta
.
put
(
Chunk
.
getMetaKey
(
c
.
id
),
c
.
asString
());
meta
.
put
(
Chunk
.
getMetaKey
(
c
.
id
),
c
.
asString
());
}
}
}
}
...
@@ -2371,7 +2467,7 @@ public class MVStore {
...
@@ -2371,7 +2467,7 @@ public class MVStore {
// could also commit when there are many unsaved pages,
// could also commit when there are many unsaved pages,
// but according to a test it doesn't really help
// but according to a test it doesn't really help
long
time
=
getTime
();
long
time
=
getTime
SinceCreation
();
if
(
time
<=
lastCommitTime
+
autoCommitDelay
)
{
if
(
time
<=
lastCommitTime
+
autoCommitDelay
)
{
return
;
return
;
}
}
...
...
h2/src/test/org/h2/test/TestAll.java
浏览文件 @
d3bd6b1d
...
@@ -795,6 +795,7 @@ kill -9 `jps -l | grep "org.h2.test." | cut -d " " -f 1`
...
@@ -795,6 +795,7 @@ kill -9 `jps -l | grep "org.h2.test." | cut -d " " -f 1`
addTest
(
new
TestMVTableEngine
());
addTest
(
new
TestMVTableEngine
());
addTest
(
new
TestObjectDataType
());
addTest
(
new
TestObjectDataType
());
addTest
(
new
TestRandomMapOps
());
addTest
(
new
TestRandomMapOps
());
addTest
(
new
TestReorderWrites
());
addTest
(
new
TestSpinLock
());
addTest
(
new
TestSpinLock
());
addTest
(
new
TestStreamStore
());
addTest
(
new
TestStreamStore
());
addTest
(
new
TestTransactionStore
());
addTest
(
new
TestTransactionStore
());
...
...
h2/src/test/org/h2/test/poweroff/TestReorderWrites.java
浏览文件 @
d3bd6b1d
...
@@ -13,6 +13,7 @@ import java.util.Map;
...
@@ -13,6 +13,7 @@ import java.util.Map;
import
java.util.Random
;
import
java.util.Random
;
import
org.h2.mvstore.MVStore
;
import
org.h2.mvstore.MVStore
;
import
org.h2.mvstore.MVStoreTool
;
import
org.h2.store.fs.FilePath
;
import
org.h2.store.fs.FilePath
;
import
org.h2.store.fs.FileUtils
;
import
org.h2.store.fs.FileUtils
;
import
org.h2.test.TestBase
;
import
org.h2.test.TestBase
;
...
@@ -24,6 +25,8 @@ import org.h2.test.utils.FilePathReorderWrites;
...
@@ -24,6 +25,8 @@ import org.h2.test.utils.FilePathReorderWrites;
*/
*/
public
class
TestReorderWrites
extends
TestBase
{
public
class
TestReorderWrites
extends
TestBase
{
private
static
final
boolean
LOG
=
false
;
/**
/**
* Run just this test.
* Run just this test.
*
*
...
@@ -35,16 +38,16 @@ public class TestReorderWrites extends TestBase {
...
@@ -35,16 +38,16 @@ public class TestReorderWrites extends TestBase {
@Override
@Override
public
void
test
()
throws
Exception
{
public
void
test
()
throws
Exception
{
testMVStore
();
testFileSystem
();
testFileSystem
();
// testMVStore();
}
}
private
void
testMVStore
()
{
private
void
testMVStore
()
{
FilePathReorderWrites
fs
=
FilePathReorderWrites
.
register
();
FilePathReorderWrites
fs
=
FilePathReorderWrites
.
register
();
String
fileName
=
"reorder:memFS:test.mv"
;
String
fileName
=
"reorder:memFS:test.mv"
;
Random
r
=
new
Random
(
1
);
for
(
int
i
=
0
;
i
<
1000
;
i
++)
{
for
(
int
i
=
0
;
i
<
100
;
i
++)
{
log
(
i
+
" --------------------------------"
);
System
.
out
.
println
(
i
+
" tst --------------------------------"
);
Random
r
=
new
Random
(
i
);
fs
.
setPowerOffCountdown
(
100
,
i
);
fs
.
setPowerOffCountdown
(
100
,
i
);
FileUtils
.
delete
(
fileName
);
FileUtils
.
delete
(
fileName
);
MVStore
store
=
new
MVStore
.
Builder
().
MVStore
store
=
new
MVStore
.
Builder
().
...
@@ -52,12 +55,12 @@ public class TestReorderWrites extends TestBase {
...
@@ -52,12 +55,12 @@ public class TestReorderWrites extends TestBase {
autoCommitDisabled
().
open
();
autoCommitDisabled
().
open
();
// store.setRetentionTime(10);
// store.setRetentionTime(10);
Map
<
Integer
,
byte
[]>
map
=
store
.
openMap
(
"data"
);
Map
<
Integer
,
byte
[]>
map
=
store
.
openMap
(
"data"
);
map
.
put
(
0
,
new
byte
[
1
]);
map
.
put
(
-
1
,
new
byte
[
1
]);
store
.
commit
();
store
.
commit
();
// if (r.nextBoolean()) {
store
.
getFileStore
().
sync
();
store
.
getFileStore
().
sync
();
//}
int
stop
=
4
+
r
.
nextInt
(
20
);
fs
.
setPowerOffCountdown
(
4
+
r
.
nextInt
(
20
),
i
);
log
(
"synched start"
);
fs
.
setPowerOffCountdown
(
stop
,
i
);
try
{
try
{
for
(
int
j
=
1
;
j
<
100
;
j
++)
{
for
(
int
j
=
1
;
j
<
100
;
j
++)
{
Map
<
Integer
,
Integer
>
newMap
=
store
.
openMap
(
"d"
+
j
);
Map
<
Integer
,
Integer
>
newMap
=
store
.
openMap
(
"d"
+
j
);
...
@@ -69,31 +72,61 @@ public class TestReorderWrites extends TestBase {
...
@@ -69,31 +72,61 @@ public class TestReorderWrites extends TestBase {
}
else
{
}
else
{
map
.
put
(
key
,
new
byte
[
len
]);
map
.
put
(
key
,
new
byte
[
len
]);
}
}
log
(
"op "
+
j
+
": "
);
store
.
commit
();
store
.
commit
();
switch
(
r
.
nextInt
(
10
))
{
case
0
:
log
(
"op compact"
);
store
.
compact
(
100
,
10
*
1024
);
break
;
case
1
:
log
(
"op compactMoveChunks"
);
store
.
compactMoveChunks
();
log
(
"op compactMoveChunks done"
);
break
;
}
}
}
// write has to fail at some point
// write has to fail at some point
fail
();
fail
();
}
catch
(
IllegalStateException
e
)
{
}
catch
(
IllegalStateException
e
)
{
log
(
"stop "
+
e
);
// expected
// expected
}
}
try
{
store
.
close
();
store
.
close
();
System
.
out
.
println
(
"-------------------------------- test"
);
}
catch
(
IllegalStateException
e
)
{
// expected
store
.
closeImmediately
();
}
log
(
"verify"
);
fs
.
setPowerOffCountdown
(
100
,
0
);
fs
.
setPowerOffCountdown
(
100
,
0
);
System
.
out
.
println
(
"file size: "
+
FileUtils
.
size
(
fileName
));
if
(
LOG
)
{
MVStoreTool
.
dump
(
fileName
,
true
);
}
store
=
new
MVStore
.
Builder
().
store
=
new
MVStore
.
Builder
().
fileName
(
fileName
).
fileName
(
fileName
).
autoCommitDisabled
().
open
();
autoCommitDisabled
().
open
();
map
=
store
.
openMap
(
"data"
);
map
=
store
.
openMap
(
"data"
);
assertEquals
(
1
,
map
.
get
(
0
).
length
);
if
(!
map
.
containsKey
(-
1
))
{
fail
(
"key not found, size="
+
map
.
size
()
+
" i="
+
i
);
}
else
{
assertEquals
(
"i="
+
i
,
1
,
map
.
get
(-
1
).
length
);
}
for
(
int
j
=
0
;
j
<
100
;
j
++)
{
for
(
int
j
=
0
;
j
<
100
;
j
++)
{
Map
<
Integer
,
Integer
>
newMap
=
store
.
openMap
(
"d"
+
j
);
Map
<
Integer
,
Integer
>
newMap
=
store
.
openMap
(
"d"
+
j
);
newMap
.
get
(
j
);
newMap
.
get
(
j
);
}
}
//
map.keySet();
map
.
keySet
();
store
.
close
();
store
.
close
();
}
}
}
}
private
static
void
log
(
String
message
)
{
if
(
LOG
)
{
System
.
out
.
println
(
message
);
}
}
private
void
testFileSystem
()
throws
IOException
{
private
void
testFileSystem
()
throws
IOException
{
FilePathReorderWrites
fs
=
FilePathReorderWrites
.
register
();
FilePathReorderWrites
fs
=
FilePathReorderWrites
.
register
();
String
fileName
=
"reorder:memFS:test"
;
String
fileName
=
"reorder:memFS:test"
;
...
...
h2/src/test/org/h2/test/store/TestMVStore.java
浏览文件 @
d3bd6b1d
...
@@ -948,6 +948,10 @@ public class TestMVStore extends TestBase {
...
@@ -948,6 +948,10 @@ public class TestMVStore extends TestBase {
break
;
break
;
}
}
}
}
// the last chunk is at the end
s
.
setReuseSpace
(
false
);
map
=
s
.
openMap
(
"test2"
);
map
.
put
(
1
,
new
byte
[
1000
]);
s
.
close
();
s
.
close
();
FilePath
f
=
FilePath
.
get
(
fileName
);
FilePath
f
=
FilePath
.
get
(
fileName
);
int
blockSize
=
4
*
1024
;
int
blockSize
=
4
*
1024
;
...
@@ -976,6 +980,8 @@ public class TestMVStore extends TestBase {
...
@@ -976,6 +980,8 @@ public class TestMVStore extends TestBase {
s
=
openStore
(
fileName
);
s
=
openStore
(
fileName
);
map
=
s
.
openMap
(
"test"
);
map
=
s
.
openMap
(
"test"
);
assertEquals
(
100
,
map
.
get
(
0
).
length
);
assertEquals
(
100
,
map
.
get
(
0
).
length
);
map
=
s
.
openMap
(
"test2"
);
assertFalse
(
map
.
containsKey
(
1
));
s
.
close
();
s
.
close
();
}
else
{
}
else
{
// both headers are corrupt
// both headers are corrupt
...
@@ -1414,7 +1420,7 @@ public class TestMVStore extends TestBase {
...
@@ -1414,7 +1420,7 @@ public class TestMVStore extends TestBase {
assertEquals
(
0
,
m
.
size
());
assertEquals
(
0
,
m
.
size
());
s
.
commit
();
s
.
commit
();
// ensure only nodes are read, but not leaves
// ensure only nodes are read, but not leaves
assertEquals
(
4
1
,
s
.
getFileStore
().
getReadCount
());
assertEquals
(
4
5
,
s
.
getFileStore
().
getReadCount
());
assertTrue
(
s
.
getFileStore
().
getWriteCount
()
<
5
);
assertTrue
(
s
.
getFileStore
().
getWriteCount
()
<
5
);
s
.
close
();
s
.
close
();
}
}
...
@@ -1579,7 +1585,6 @@ public class TestMVStore extends TestBase {
...
@@ -1579,7 +1585,6 @@ public class TestMVStore extends TestBase {
data
.
put
(
"2"
,
"World"
);
data
.
put
(
"2"
,
"World"
);
s
.
commit
();
s
.
commit
();
assertEquals
(
1
,
s
.
getCurrentVersion
());
assertEquals
(
1
,
s
.
getCurrentVersion
());
assertFalse
(
m
.
containsKey
(
"chunk.2"
));
assertEquals
(
"[data]"
,
s
.
getMapNames
().
toString
());
assertEquals
(
"[data]"
,
s
.
getMapNames
().
toString
());
assertEquals
(
"data"
,
s
.
getMapName
(
data
.
getId
()));
assertEquals
(
"data"
,
s
.
getMapName
(
data
.
getId
()));
...
@@ -1599,8 +1604,6 @@ public class TestMVStore extends TestBase {
...
@@ -1599,8 +1604,6 @@ public class TestMVStore extends TestBase {
s
.
rollbackTo
(
1
);
s
.
rollbackTo
(
1
);
assertEquals
(
"Hello"
,
data
.
get
(
"1"
));
assertEquals
(
"Hello"
,
data
.
get
(
"1"
));
assertEquals
(
"World"
,
data
.
get
(
"2"
));
assertEquals
(
"World"
,
data
.
get
(
"2"
));
assertFalse
(
m
.
containsKey
(
"chunk.1"
));
assertFalse
(
m
.
containsKey
(
"chunk.2"
));
s
.
close
();
s
.
close
();
}
}
...
...
h2/src/test/org/h2/test/utils/FilePathReorderWrites.java
浏览文件 @
d3bd6b1d
...
@@ -105,10 +105,11 @@ public class FilePathReorderWrites extends FilePathWrapper {
...
@@ -105,10 +105,11 @@ public class FilePathReorderWrites extends FilePathWrapper {
if
(
powerFailureCountdown
==
0
)
{
if
(
powerFailureCountdown
==
0
)
{
return
;
return
;
}
}
if
(
--
powerFailureCountdown
>
0
)
{
if
(
powerFailureCountdown
<
0
)
{
return
;
throw
POWER_FAILURE
;
}
}
if
(
powerFailureCountdown
>=
-
1
)
{
powerFailureCountdown
--;
if
(
powerFailureCountdown
==
0
)
{
powerFailureCountdown
--;
powerFailureCountdown
--;
throw
POWER_FAILURE
;
throw
POWER_FAILURE
;
}
}
...
@@ -122,7 +123,7 @@ public class FilePathReorderWrites extends FilePathWrapper {
...
@@ -122,7 +123,7 @@ public class FilePathReorderWrites extends FilePathWrapper {
IOUtils
.
copy
(
in
,
out
);
IOUtils
.
copy
(
in
,
out
);
FileChannel
base
=
getBase
().
open
(
mode
);
FileChannel
base
=
getBase
().
open
(
mode
);
FileChannel
readBase
=
copy
.
open
(
mode
);
FileChannel
readBase
=
copy
.
open
(
mode
);
return
new
File
PowerFailure
(
this
,
base
,
readBase
);
return
new
File
ReorderWrites
(
this
,
base
,
readBase
);
}
}
@Override
@Override
...
@@ -140,7 +141,7 @@ public class FilePathReorderWrites extends FilePathWrapper {
...
@@ -140,7 +141,7 @@ public class FilePathReorderWrites extends FilePathWrapper {
/**
/**
* An file that checks for errors before each write operation.
* An file that checks for errors before each write operation.
*/
*/
class
File
PowerFailure
extends
FileBase
{
class
File
ReorderWrites
extends
FileBase
{
private
final
FilePathReorderWrites
file
;
private
final
FilePathReorderWrites
file
;
/**
/**
...
@@ -162,7 +163,7 @@ class FilePowerFailure extends FileBase {
...
@@ -162,7 +163,7 @@ class FilePowerFailure extends FileBase {
private
int
id
;
private
int
id
;
File
PowerFailure
(
FilePathReorderWrites
file
,
FileChannel
base
,
FileChannel
readBase
)
{
File
ReorderWrites
(
FilePathReorderWrites
file
,
FileChannel
base
,
FileChannel
readBase
)
{
this
.
file
=
file
;
this
.
file
=
file
;
this
.
base
=
base
;
this
.
base
=
base
;
this
.
readBase
=
readBase
;
this
.
readBase
=
readBase
;
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论