Skip to content
项目
群组
代码片段
帮助
正在加载...
帮助
为 GitLab 提交贡献
登录/注册
切换导航
H
h2database
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
分枝图
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
计划
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
分枝图
统计图
创建新议题
作业
提交
议题看板
打开侧边栏
Administrator
h2database
Commits
2ed0a04b
提交
2ed0a04b
authored
2月 07, 2014
作者:
Thomas Mueller
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
MVStore: the file format was changed slightly.
上级
b9633b12
隐藏空白字符变更
内嵌
并排
正在显示
10 个修改的文件
包含
334 行增加
和
220 行删除
+334
-220
Chunk.java
h2/src/main/org/h2/mvstore/Chunk.java
+33
-4
MVStore.java
h2/src/main/org/h2/mvstore/MVStore.java
+238
-172
MVStoreTool.java
h2/src/main/org/h2/mvstore/MVStoreTool.java
+1
-1
OffHeapStore.java
h2/src/main/org/h2/mvstore/OffHeapStore.java
+2
-2
Page.java
h2/src/main/org/h2/mvstore/Page.java
+1
-1
MVTableEngine.java
h2/src/main/org/h2/mvstore/db/MVTableEngine.java
+3
-3
TestBenchmark.java
h2/src/test/org/h2/test/store/TestBenchmark.java
+2
-2
TestKillProcessWhileWriting.java
...c/test/org/h2/test/store/TestKillProcessWhileWriting.java
+1
-1
TestMVStore.java
h2/src/test/org/h2/test/store/TestMVStore.java
+46
-28
TestRandomMapOps.java
h2/src/test/org/h2/test/store/TestRandomMapOps.java
+7
-6
没有找到文件。
h2/src/main/org/h2/mvstore/Chunk.java
浏览文件 @
2ed0a04b
...
@@ -82,6 +82,12 @@ public class Chunk {
...
@@ -82,6 +82,12 @@ public class Chunk {
* The last used map id.
* The last used map id.
*/
*/
public
int
mapId
;
public
int
mapId
;
/**
* The predicted position of the next chunk.
*/
public
long
next
;
public
long
nextSize
;
Chunk
(
int
id
)
{
Chunk
(
int
id
)
{
this
.
id
=
id
;
this
.
id
=
id
;
...
@@ -94,7 +100,7 @@ public class Chunk {
...
@@ -94,7 +100,7 @@ public class Chunk {
* @param start the start of the chunk in the file
* @param start the start of the chunk in the file
* @return the chunk
* @return the chunk
*/
*/
static
Chunk
from
Header
(
ByteBuffer
buff
,
long
start
)
{
static
Chunk
readChunk
Header
(
ByteBuffer
buff
,
long
start
)
{
int
pos
=
buff
.
position
();
int
pos
=
buff
.
position
();
byte
[]
data
=
new
byte
[
Math
.
min
(
buff
.
remaining
(),
MAX_HEADER_LENGTH
)];
byte
[]
data
=
new
byte
[
Math
.
min
(
buff
.
remaining
(),
MAX_HEADER_LENGTH
)];
buff
.
get
(
data
);
buff
.
get
(
data
);
...
@@ -120,7 +126,7 @@ public class Chunk {
...
@@ -120,7 +126,7 @@ public class Chunk {
*
*
* @param buff the target buffer
* @param buff the target buffer
*/
*/
void
writeHeader
(
WriteBuffer
buff
,
int
minLength
)
{
void
write
Chunk
Header
(
WriteBuffer
buff
,
int
minLength
)
{
long
pos
=
buff
.
position
();
long
pos
=
buff
.
position
();
buff
.
put
(
asString
().
getBytes
(
DataUtils
.
LATIN
));
buff
.
put
(
asString
().
getBytes
(
DataUtils
.
LATIN
));
while
(
buff
.
position
()
-
pos
<
minLength
-
1
)
{
while
(
buff
.
position
()
-
pos
<
minLength
-
1
)
{
...
@@ -152,7 +158,8 @@ public class Chunk {
...
@@ -152,7 +158,8 @@ public class Chunk {
c
.
maxLenLive
=
DataUtils
.
readHexLong
(
map
,
"liveMax"
,
c
.
maxLength
);
c
.
maxLenLive
=
DataUtils
.
readHexLong
(
map
,
"liveMax"
,
c
.
maxLength
);
c
.
metaRootPos
=
DataUtils
.
readHexLong
(
map
,
"root"
,
0
);
c
.
metaRootPos
=
DataUtils
.
readHexLong
(
map
,
"root"
,
0
);
c
.
time
=
DataUtils
.
readHexLong
(
map
,
"time"
,
0
);
c
.
time
=
DataUtils
.
readHexLong
(
map
,
"time"
,
0
);
c
.
version
=
DataUtils
.
readHexLong
(
map
,
"version"
,
0
);
c
.
version
=
DataUtils
.
readHexLong
(
map
,
"version"
,
id
);
c
.
next
=
DataUtils
.
readHexLong
(
map
,
"next"
,
0
);
return
c
;
return
c
;
}
}
...
@@ -188,12 +195,34 @@ public class Chunk {
...
@@ -188,12 +195,34 @@ public class Chunk {
}
}
DataUtils
.
appendMap
(
buff
,
"map"
,
mapId
);
DataUtils
.
appendMap
(
buff
,
"map"
,
mapId
);
DataUtils
.
appendMap
(
buff
,
"max"
,
maxLength
);
DataUtils
.
appendMap
(
buff
,
"max"
,
maxLength
);
if
(
next
!=
0
)
{
DataUtils
.
appendMap
(
buff
,
"next"
,
next
);
}
DataUtils
.
appendMap
(
buff
,
"pages"
,
pageCount
);
DataUtils
.
appendMap
(
buff
,
"pages"
,
pageCount
);
DataUtils
.
appendMap
(
buff
,
"root"
,
metaRootPos
);
DataUtils
.
appendMap
(
buff
,
"root"
,
metaRootPos
);
DataUtils
.
appendMap
(
buff
,
"time"
,
time
);
DataUtils
.
appendMap
(
buff
,
"time"
,
time
);
DataUtils
.
appendMap
(
buff
,
"version"
,
version
);
if
(
version
!=
id
)
{
DataUtils
.
appendMap
(
buff
,
"version"
,
version
);
}
return
buff
.
toString
();
return
buff
.
toString
();
}
}
byte
[]
getFooterBytes
()
{
StringBuilder
buff
=
new
StringBuilder
();
DataUtils
.
appendMap
(
buff
,
"chunk"
,
id
);
DataUtils
.
appendMap
(
buff
,
"block"
,
block
);
if
(
version
!=
id
)
{
DataUtils
.
appendMap
(
buff
,
"version"
,
version
);
}
byte
[]
bytes
=
buff
.
toString
().
getBytes
(
DataUtils
.
LATIN
);
int
checksum
=
DataUtils
.
getFletcher32
(
bytes
,
bytes
.
length
/
2
*
2
);
DataUtils
.
appendMap
(
buff
,
"fletcher"
,
checksum
);
while
(
buff
.
length
()
<
MVStore
.
CHUNK_FOOTER_LENGTH
-
1
)
{
buff
.
append
(
' '
);
}
buff
.
append
(
"\n"
);
return
buff
.
toString
().
getBytes
(
DataUtils
.
LATIN
);
}
@Override
@Override
public
String
toString
()
{
public
String
toString
()
{
...
...
h2/src/main/org/h2/mvstore/MVStore.java
浏览文件 @
2ed0a04b
...
@@ -50,6 +50,8 @@ Documentation
...
@@ -50,6 +50,8 @@ Documentation
- better document that writes are in background thread
- better document that writes are in background thread
- better document how to do non-unique indexes
- better document how to do non-unique indexes
- document pluggable store and OffHeapStore
- document pluggable store and OffHeapStore
- document file format
- the chunk id is normally the version
MVTableEngine:
MVTableEngine:
- verify that tests don't use the PageStore
- verify that tests don't use the PageStore
...
@@ -64,10 +66,12 @@ TransactionStore:
...
@@ -64,10 +66,12 @@ TransactionStore:
MVStore:
MVStore:
- maybe reduce size of store header
- maybe let a chunk point to a list of potential next chunks
- maybe let a chunk point to a list of potential next chunks
(so no fixed location header is needed), similar to a skip list
(so no fixed location header is needed), similar to a skip list
(options: just after the current chunk, list of next)
- document and review the file format
- document and review the file format
- the newest chunk is the one with the newest version
(not necessarily the one with the highest chunk id)
- automated 'kill process' and 'power failure' test
- automated 'kill process' and 'power failure' test
- update checkstyle
- update checkstyle
...
@@ -166,8 +170,6 @@ public class MVStore {
...
@@ -166,8 +170,6 @@ public class MVStore {
private
final
int
pageSplitSize
;
private
final
int
pageSplitSize
;
private
long
lastChunkBlock
;
/**
/**
* The page cache. The default size is 16 MB, and the average size is 2 KB.
* The page cache. The default size is 16 MB, and the average size is 2 KB.
* It is split in 16 segments. The stack move distance is 2% of the expected
* It is split in 16 segments. The stack move distance is 2% of the expected
...
@@ -175,7 +177,10 @@ public class MVStore {
...
@@ -175,7 +177,10 @@ public class MVStore {
*/
*/
private
CacheLongKeyLIRS
<
Page
>
cache
;
private
CacheLongKeyLIRS
<
Page
>
cache
;
private
int
lastChunkId
;
/**
* The newest chunk. If nothing was stored yet, this field is not set.
*/
private
Chunk
lastChunk
;
/**
/**
* The map of chunks.
* The map of chunks.
...
@@ -219,7 +224,7 @@ public class MVStore {
...
@@ -219,7 +224,7 @@ public class MVStore {
private
long
currentVersion
;
private
long
currentVersion
;
/**
/**
* The version of the last stored chunk.
* The version of the last stored chunk
, or -1 if nothing was stored so far
.
*/
*/
private
long
lastStoredVersion
;
private
long
lastStoredVersion
;
...
@@ -323,24 +328,6 @@ public class MVStore {
...
@@ -323,24 +328,6 @@ public class MVStore {
writeFileHeader
();
writeFileHeader
();
}
else
{
}
else
{
readFileHeader
();
readFileHeader
();
long
format
=
DataUtils
.
readHexLong
(
fileHeader
,
"format"
,
1
);
if
(
format
>
FORMAT_WRITE
&&
!
fileStore
.
isReadOnly
())
{
throw
DataUtils
.
newIllegalStateException
(
DataUtils
.
ERROR_UNSUPPORTED_FORMAT
,
"The write format {0} is larger than the supported format {1}, "
+
"and the file was not opened in read-only mode"
,
format
,
FORMAT_WRITE
);
}
format
=
DataUtils
.
readHexLong
(
fileHeader
,
"formatRead"
,
format
);
if
(
format
>
FORMAT_READ
)
{
throw
DataUtils
.
newIllegalStateException
(
DataUtils
.
ERROR_UNSUPPORTED_FORMAT
,
"The read format {0} is larger than the supported format {1}"
,
format
,
FORMAT_READ
);
}
if
(
lastChunkBlock
>
0
)
{
readMeta
();
}
}
}
}
catch
(
IllegalStateException
e
)
{
}
catch
(
IllegalStateException
e
)
{
try
{
try
{
...
@@ -510,13 +497,12 @@ public class MVStore {
...
@@ -510,13 +497,12 @@ public class MVStore {
}
}
private
Chunk
getChunkForVersion
(
long
version
)
{
private
Chunk
getChunkForVersion
(
long
version
)
{
for
(
int
chunkId
=
lastChunkId
;;
chunkId
--)
{
Chunk
c
=
lastChunk
;
Chunk
x
=
chunks
.
get
(
chunkId
);
while
(
true
)
{
if
(
x
==
null
)
{
if
(
c
==
null
||
c
.
version
<=
version
)
{
return
null
;
return
c
;
}
else
if
(
x
.
version
<=
version
)
{
return
x
;
}
}
c
=
chunks
.
get
(
c
.
id
-
1
);
}
}
}
}
...
@@ -536,56 +522,11 @@ public class MVStore {
...
@@ -536,56 +522,11 @@ public class MVStore {
metaChanged
=
true
;
metaChanged
=
true
;
}
}
private
synchronized
void
readMeta
()
{
private
synchronized
void
readFileHeader
()
{
chunks
.
clear
();
Chunk
header
=
readChunkHeader
(
lastChunkBlock
);
if
(
header
.
block
==
Long
.
MAX_VALUE
)
{
throw
DataUtils
.
newIllegalStateException
(
DataUtils
.
ERROR_FILE_CORRUPT
,
"Chunk {0} is invalid"
,
header
.
id
);
}
lastChunkId
=
header
.
id
;
lastMapId
=
header
.
mapId
;
currentVersion
=
header
.
version
;
setWriteVersion
(
currentVersion
);
chunks
.
put
(
header
.
id
,
header
);
meta
.
setRootPos
(
header
.
metaRootPos
,
-
1
);
chunks
.
put
(
header
.
id
,
header
);
// we can load the chunk in any order,
// because loading chunk metadata
// might recursively load another chunk
for
(
Iterator
<
String
>
it
=
meta
.
keyIterator
(
"chunk."
);
it
.
hasNext
();)
{
String
s
=
it
.
next
();
if
(!
s
.
startsWith
(
"chunk."
))
{
break
;
}
s
=
meta
.
get
(
s
);
Chunk
c
=
Chunk
.
fromString
(
s
);
if
(!
chunks
.
containsKey
(
c
.
id
))
{
if
(
c
.
block
==
Long
.
MAX_VALUE
)
{
throw
DataUtils
.
newIllegalStateException
(
DataUtils
.
ERROR_FILE_CORRUPT
,
"Chunk {0} is invalid"
,
c
.
id
);
}
chunks
.
put
(
c
.
id
,
c
);
}
}
// build the free space list
for
(
Chunk
c
:
chunks
.
values
())
{
if
(
c
.
pageCountLive
==
0
)
{
// remove this chunk in the next save operation
registerFreePage
(
currentVersion
,
c
.
id
,
0
,
0
);
}
long
start
=
c
.
block
*
BLOCK_SIZE
;
int
length
=
c
.
len
*
BLOCK_SIZE
;
fileStore
.
markUsed
(
start
,
length
);
}
}
private
void
readFileHeader
()
{
boolean
validHeader
=
false
;
boolean
validHeader
=
false
;
// we don't know yet which chunk is the newest
// we don't know yet which chunk and version are the newest
long
newestChunk
=
-
1
;
long
newestVersion
=
-
1
;
long
newestChunkBlock
=
-
1
;
// read the last block of the file, and then the two first blocks
// read the last block of the file, and then the two first blocks
ByteBuffer
fileHeaderBlocks
=
fileStore
.
readFully
(
0
,
2
*
BLOCK_SIZE
);
ByteBuffer
fileHeaderBlocks
=
fileStore
.
readFully
(
0
,
2
*
BLOCK_SIZE
);
byte
[]
buff
=
new
byte
[
BLOCK_SIZE
];
byte
[]
buff
=
new
byte
[
BLOCK_SIZE
];
...
@@ -603,11 +544,12 @@ public class MVStore {
...
@@ -603,11 +544,12 @@ public class MVStore {
if
(
check
!=
checksum
)
{
if
(
check
!=
checksum
)
{
continue
;
continue
;
}
}
long
chunk
=
DataUtils
.
readHexLong
(
m
,
"chunk"
,
0
);
int
chunk
=
DataUtils
.
readHexInt
(
m
,
"chunk"
,
0
);
if
(
chunk
>
newestChunk
)
{
long
version
=
DataUtils
.
readHexLong
(
m
,
"version"
,
chunk
);
newestChunk
=
chunk
;
if
(
version
>
newestVersion
)
{
newestVersion
=
version
;
fileHeader
.
putAll
(
m
);
fileHeader
.
putAll
(
m
);
la
stChunkBlock
=
DataUtils
.
readHexLong
(
m
,
"block"
,
0
);
newe
stChunkBlock
=
DataUtils
.
readHexLong
(
m
,
"block"
,
0
);
creationTime
=
DataUtils
.
readHexLong
(
m
,
"created"
,
0
);
creationTime
=
DataUtils
.
readHexLong
(
m
,
"created"
,
0
);
validHeader
=
true
;
validHeader
=
true
;
}
}
...
@@ -620,6 +562,23 @@ public class MVStore {
...
@@ -620,6 +562,23 @@ public class MVStore {
DataUtils
.
ERROR_FILE_CORRUPT
,
DataUtils
.
ERROR_FILE_CORRUPT
,
"Store header is corrupt: {0}"
,
fileStore
);
"Store header is corrupt: {0}"
,
fileStore
);
}
}
long
format
=
DataUtils
.
readHexLong
(
fileHeader
,
"format"
,
1
);
if
(
format
>
FORMAT_WRITE
&&
!
fileStore
.
isReadOnly
())
{
throw
DataUtils
.
newIllegalStateException
(
DataUtils
.
ERROR_UNSUPPORTED_FORMAT
,
"The write format {0} is larger than the supported format {1}, "
+
"and the file was not opened in read-only mode"
,
format
,
FORMAT_WRITE
);
}
format
=
DataUtils
.
readHexLong
(
fileHeader
,
"formatRead"
,
format
);
if
(
format
>
FORMAT_READ
)
{
throw
DataUtils
.
newIllegalStateException
(
DataUtils
.
ERROR_UNSUPPORTED_FORMAT
,
"The read format {0} is larger than the supported format {1}"
,
format
,
FORMAT_READ
);
}
lastStoredVersion
=
-
1
;
ByteBuffer
lastBlock
=
fileStore
.
readFully
(
ByteBuffer
lastBlock
=
fileStore
.
readFully
(
fileStore
.
size
()
-
CHUNK_FOOTER_LENGTH
,
CHUNK_FOOTER_LENGTH
);
fileStore
.
size
()
-
CHUNK_FOOTER_LENGTH
,
CHUNK_FOOTER_LENGTH
);
buff
=
new
byte
[
CHUNK_FOOTER_LENGTH
];
buff
=
new
byte
[
CHUNK_FOOTER_LENGTH
];
...
@@ -634,23 +593,79 @@ public class MVStore {
...
@@ -634,23 +593,79 @@ public class MVStore {
byte
[]
bytes
=
s
.
getBytes
(
DataUtils
.
LATIN
);
byte
[]
bytes
=
s
.
getBytes
(
DataUtils
.
LATIN
);
int
checksum
=
DataUtils
.
getFletcher32
(
bytes
,
bytes
.
length
/
2
*
2
);
int
checksum
=
DataUtils
.
getFletcher32
(
bytes
,
bytes
.
length
/
2
*
2
);
if
(
check
==
checksum
)
{
if
(
check
==
checksum
)
{
long
chunk
=
DataUtils
.
readHexLong
(
m
,
"chunk"
,
0
);
int
chunk
=
DataUtils
.
readHexInt
(
m
,
"chunk"
,
0
);
if
(
chunk
>
newestChunk
)
{
long
version
=
DataUtils
.
readHexLong
(
m
,
"version"
,
chunk
);
if
(
version
>
newestVersion
)
{
fileHeader
.
putAll
(
m
);
fileHeader
.
putAll
(
m
);
lastChunkBlock
=
DataUtils
.
readHexLong
(
m
,
"block"
,
0
);
newestVersion
=
version
;
newestChunkBlock
=
DataUtils
.
readHexLong
(
m
,
"block"
,
0
);
validHeader
=
true
;
validHeader
=
true
;
}
}
}
}
}
catch
(
Exception
e
)
{
}
catch
(
Exception
e
)
{
// ignore
// ignore
}
}
lastStoredVersion
=
-
1
;
if
(
newestChunkBlock
>
0
)
{
readMeta
(
newestChunkBlock
);
}
}
private
void
readMeta
(
long
chunkBlock
)
{
chunks
.
clear
();
Chunk
header
=
readChunkHeader
(
chunkBlock
);
if
(
header
.
block
==
Long
.
MAX_VALUE
)
{
throw
DataUtils
.
newIllegalStateException
(
DataUtils
.
ERROR_FILE_CORRUPT
,
"Chunk {0} is invalid"
,
header
.
id
);
}
lastChunk
=
header
;
lastMapId
=
header
.
mapId
;
currentVersion
=
header
.
version
;
setWriteVersion
(
currentVersion
);
chunks
.
put
(
header
.
id
,
header
);
meta
.
setRootPos
(
header
.
metaRootPos
,
-
1
);
// we can load the chunk in any order,
// because loading chunk metadata
// might recursively load another chunk
for
(
Iterator
<
String
>
it
=
meta
.
keyIterator
(
"chunk."
);
it
.
hasNext
();)
{
String
s
=
it
.
next
();
if
(!
s
.
startsWith
(
"chunk."
))
{
break
;
}
s
=
meta
.
get
(
s
);
Chunk
c
=
Chunk
.
fromString
(
s
);
if
(!
chunks
.
containsKey
(
c
.
id
))
{
if
(
c
.
block
==
Long
.
MAX_VALUE
)
{
throw
DataUtils
.
newIllegalStateException
(
DataUtils
.
ERROR_FILE_CORRUPT
,
"Chunk {0} is invalid"
,
c
.
id
);
}
chunks
.
put
(
c
.
id
,
c
);
}
}
// build the free space list
for
(
Chunk
c
:
chunks
.
values
())
{
if
(
c
.
pageCountLive
==
0
)
{
// remove this chunk in the next save operation
registerFreePage
(
currentVersion
,
c
.
id
,
0
,
0
);
}
long
start
=
c
.
block
*
BLOCK_SIZE
;
int
length
=
c
.
len
*
BLOCK_SIZE
;
fileStore
.
markUsed
(
start
,
length
);
}
}
}
private
byte
[]
getFileHeaderBytes
()
{
private
byte
[]
getFileHeaderBytes
()
{
StringBuilder
buff
=
new
StringBuilder
(
"H:2"
);
StringBuilder
buff
=
new
StringBuilder
(
"H:2"
);
fileHeader
.
put
(
"block"
,
lastChunkBlock
);
if
(
lastChunk
!=
null
)
{
fileHeader
.
put
(
"chunk"
,
lastChunkId
);
fileHeader
.
put
(
"chunk"
,
lastChunk
.
id
);
if
(
lastChunk
.
version
!=
lastChunk
.
id
)
{
fileHeader
.
put
(
"version"
,
lastChunk
.
version
);
}
else
{
fileHeader
.
remove
(
"version"
);
}
fileHeader
.
put
(
"block"
,
lastChunk
.
block
);
}
DataUtils
.
appendMap
(
buff
,
fileHeader
);
DataUtils
.
appendMap
(
buff
,
fileHeader
);
byte
[]
bytes
=
buff
.
toString
().
getBytes
(
DataUtils
.
LATIN
);
byte
[]
bytes
=
buff
.
toString
().
getBytes
(
DataUtils
.
LATIN
);
int
checksum
=
DataUtils
.
getFletcher32
(
bytes
,
bytes
.
length
/
2
*
2
);
int
checksum
=
DataUtils
.
getFletcher32
(
bytes
,
bytes
.
length
/
2
*
2
);
...
@@ -659,22 +674,6 @@ public class MVStore {
...
@@ -659,22 +674,6 @@ public class MVStore {
return
buff
.
toString
().
getBytes
(
DataUtils
.
LATIN
);
return
buff
.
toString
().
getBytes
(
DataUtils
.
LATIN
);
}
}
private
byte
[]
getChunkFooterBytes
()
{
StringBuilder
buff
=
new
StringBuilder
();
fileHeader
.
put
(
"chunk"
,
lastChunkId
);
fileHeader
.
put
(
"block"
,
lastChunkBlock
);
DataUtils
.
appendMap
(
buff
,
"chunk"
,
lastChunkId
);
DataUtils
.
appendMap
(
buff
,
"block"
,
lastChunkBlock
);
byte
[]
bytes
=
buff
.
toString
().
getBytes
(
DataUtils
.
LATIN
);
int
checksum
=
DataUtils
.
getFletcher32
(
bytes
,
bytes
.
length
/
2
*
2
);
DataUtils
.
appendMap
(
buff
,
"fletcher"
,
checksum
);
while
(
buff
.
length
()
<
CHUNK_FOOTER_LENGTH
-
1
)
{
buff
.
append
(
' '
);
}
buff
.
append
(
"\n"
);
return
buff
.
toString
().
getBytes
(
DataUtils
.
LATIN
);
}
private
void
writeFileHeader
()
{
private
void
writeFileHeader
()
{
byte
[]
bytes
=
getFileHeaderBytes
();
byte
[]
bytes
=
getFileHeaderBytes
();
ByteBuffer
header
=
ByteBuffer
.
allocate
(
2
*
BLOCK_SIZE
);
ByteBuffer
header
=
ByteBuffer
.
allocate
(
2
*
BLOCK_SIZE
);
...
@@ -872,14 +871,18 @@ public class MVStore {
...
@@ -872,14 +871,18 @@ public class MVStore {
// the last chunk was not stored before and needs to be set now (it's
// the last chunk was not stored before and needs to be set now (it's
// better not to update right after storing, because that would modify
// better not to update right after storing, because that would modify
// the meta map again)
// the meta map again)
Chunk
lastChunk
=
chunks
.
get
(
lastChunkId
);
int
lastChunkId
;
if
(
lastChunk
!=
null
)
{
if
(
lastChunk
==
null
)
{
meta
.
put
(
Chunk
.
getMetaKey
(
lastChunk
.
id
),
lastChunk
.
asString
());
lastChunkId
=
0
;
}
else
{
lastChunkId
=
lastChunk
.
id
;
meta
.
put
(
Chunk
.
getMetaKey
(
lastChunkId
),
lastChunk
.
asString
());
// never go backward in time
// never go backward in time
time
=
Math
.
max
(
lastChunk
.
time
,
time
);
time
=
Math
.
max
(
lastChunk
.
time
,
time
);
}
}
Chunk
c
;
Chunk
c
;
c
=
new
Chunk
(++
lastChunkId
);
c
=
new
Chunk
(
lastChunkId
+
1
);
c
.
pageCount
=
Integer
.
MAX_VALUE
;
c
.
pageCount
=
Integer
.
MAX_VALUE
;
c
.
pageCountLive
=
Integer
.
MAX_VALUE
;
c
.
pageCountLive
=
Integer
.
MAX_VALUE
;
c
.
maxLength
=
Long
.
MAX_VALUE
;
c
.
maxLength
=
Long
.
MAX_VALUE
;
...
@@ -890,6 +893,7 @@ public class MVStore {
...
@@ -890,6 +893,7 @@ public class MVStore {
c
.
time
=
time
;
c
.
time
=
time
;
c
.
version
=
version
;
c
.
version
=
version
;
c
.
mapId
=
lastMapId
;
c
.
mapId
=
lastMapId
;
c
.
next
=
Long
.
MAX_VALUE
;
chunks
.
put
(
c
.
id
,
c
);
chunks
.
put
(
c
.
id
,
c
);
// force a metadata update
// force a metadata update
meta
.
put
(
Chunk
.
getMetaKey
(
c
.
id
),
c
.
asString
());
meta
.
put
(
Chunk
.
getMetaKey
(
c
.
id
),
c
.
asString
());
...
@@ -923,7 +927,7 @@ public class MVStore {
...
@@ -923,7 +927,7 @@ public class MVStore {
Set
<
Chunk
>
removedChunks
=
applyFreedSpace
(
storeVersion
,
time
);
Set
<
Chunk
>
removedChunks
=
applyFreedSpace
(
storeVersion
,
time
);
WriteBuffer
buff
=
getWriteBuffer
();
WriteBuffer
buff
=
getWriteBuffer
();
// need to patch the header later
// need to patch the header later
c
.
writeHeader
(
buff
,
0
);
c
.
write
Chunk
Header
(
buff
,
0
);
int
headerLength
=
buff
.
position
();
int
headerLength
=
buff
.
position
();
c
.
pageCount
=
0
;
c
.
pageCount
=
0
;
c
.
pageCountLive
=
0
;
c
.
pageCountLive
=
0
;
...
@@ -950,13 +954,6 @@ public class MVStore {
...
@@ -950,13 +954,6 @@ public class MVStore {
CHUNK_FOOTER_LENGTH
,
BLOCK_SIZE
);
CHUNK_FOOTER_LENGTH
,
BLOCK_SIZE
);
buff
.
limit
(
length
);
buff
.
limit
(
length
);
// free up the space of unused chunks now
for
(
Chunk
x
:
removedChunks
)
{
long
start
=
x
.
block
*
BLOCK_SIZE
;
int
len
=
x
.
len
*
BLOCK_SIZE
;
fileStore
.
free
(
start
,
len
);
}
// the length of the file that is still in use
// the length of the file that is still in use
// (not necessarily the end of the file)
// (not necessarily the end of the file)
long
end
=
getFileLengthInUse
();
long
end
=
getFileLengthInUse
();
...
@@ -965,29 +962,85 @@ public class MVStore {
...
@@ -965,29 +962,85 @@ public class MVStore {
filePos
=
fileStore
.
allocate
(
length
);
filePos
=
fileStore
.
allocate
(
length
);
}
else
{
}
else
{
filePos
=
end
;
filePos
=
end
;
fileStore
.
markUsed
(
end
,
length
);
}
}
// end is not necessarily the end of the file
// end is not necessarily the end of the file
boolean
storeAtEndOfFile
=
filePos
+
length
>=
fileStore
.
size
();
boolean
storeAtEndOfFile
=
filePos
+
length
>=
fileStore
.
size
();
// free up the space of unused chunks now
// (after allocating space for this chunk, so that
// no data is lost if writing this chunk fails)
for
(
Chunk
x
:
removedChunks
)
{
long
start
=
x
.
block
*
BLOCK_SIZE
;
int
len
=
x
.
len
*
BLOCK_SIZE
;
fileStore
.
free
(
start
,
len
);
}
if
(!
reuseSpace
)
{
// we can not mark it earlier, because it
// might have been allocated by one of the
// removed chunks
fileStore
.
markUsed
(
end
,
length
);
}
c
.
block
=
filePos
/
BLOCK_SIZE
;
c
.
block
=
filePos
/
BLOCK_SIZE
;
c
.
len
=
length
/
BLOCK_SIZE
;
c
.
len
=
length
/
BLOCK_SIZE
;
c
.
metaRootPos
=
metaRoot
.
getPos
();
c
.
metaRootPos
=
metaRoot
.
getPos
();
// calculate and set the likely next position
if
(
reuseSpace
)
{
int
predictBlocks
=
c
.
len
;
c
.
nextSize
=
predictBlocks
;
long
predictedNextStart
=
fileStore
.
allocate
(
predictBlocks
*
BLOCK_SIZE
);
fileStore
.
free
(
predictedNextStart
,
predictBlocks
*
BLOCK_SIZE
);
c
.
next
=
predictedNextStart
/
BLOCK_SIZE
;
if
(
c
.
next
==
c
.
block
+
c
.
len
)
{
// just after this chunk
c
.
next
=
0
;
}
}
else
{
// just after this chunk
c
.
next
=
0
;
}
buff
.
position
(
0
);
buff
.
position
(
0
);
c
.
writeHeader
(
buff
,
headerLength
);
c
.
writeChunkHeader
(
buff
,
headerLength
);
lastChunkBlock
=
filePos
/
BLOCK_SIZE
;
revertTemp
(
storeVersion
);
revertTemp
(
storeVersion
);
buff
.
position
(
buff
.
limit
()
-
CHUNK_FOOTER_LENGTH
);
buff
.
position
(
buff
.
limit
()
-
CHUNK_FOOTER_LENGTH
);
buff
.
put
(
getChunk
FooterBytes
());
buff
.
put
(
c
.
get
FooterBytes
());
buff
.
position
(
0
);
buff
.
position
(
0
);
write
(
filePos
,
buff
.
getBuffer
());
write
(
filePos
,
buff
.
getBuffer
());
releaseWriteBuffer
(
buff
);
releaseWriteBuffer
(
buff
);
// overwrite the header if required
// whether to overwrite the file header
boolean
needHeader
=
false
;
// whether to reset the list of next chunks
boolean
resetNextChain
=
false
;
if
(!
storeAtEndOfFile
)
{
if
(!
storeAtEndOfFile
)
{
if
(
lastChunk
==
null
)
{
needHeader
=
true
;
}
else
if
(
lastChunk
.
block
+
lastChunk
.
len
==
c
.
block
)
{
// just after the last chunk
}
else
if
(
lastChunk
.
next
==
c
.
block
)
{
// the last matched
}
else
{
needHeader
=
true
;
}
;
// TODO check the length of the list - if too large
// write anyway (lets assume 11 is too large)
;
// TODO verify that no chunk between the last known one
// and the current chunk were removed
;
// TODO remove this once we have implemented
// reading and once we have tests
needHeader
=
true
;
}
lastChunk
=
c
;
if
(
needHeader
)
{
writeFileHeader
();
writeFileHeader
();
}
if
(!
storeAtEndOfFile
)
{
// may only shrink after the file header was written
shrinkFileIfPossible
(
1
);
shrinkFileIfPossible
(
1
);
}
}
...
@@ -1187,7 +1240,7 @@ public class MVStore {
...
@@ -1187,7 +1240,7 @@ public class MVStore {
private
Chunk
readChunkHeader
(
long
block
)
{
private
Chunk
readChunkHeader
(
long
block
)
{
long
p
=
block
*
BLOCK_SIZE
;
long
p
=
block
*
BLOCK_SIZE
;
ByteBuffer
buff
=
fileStore
.
readFully
(
p
,
Chunk
.
MAX_HEADER_LENGTH
);
ByteBuffer
buff
=
fileStore
.
readFully
(
p
,
Chunk
.
MAX_HEADER_LENGTH
);
return
Chunk
.
from
Header
(
buff
,
p
);
return
Chunk
.
readChunk
Header
(
buff
,
p
);
}
}
/**
/**
...
@@ -1200,7 +1253,7 @@ public class MVStore {
...
@@ -1200,7 +1253,7 @@ public class MVStore {
*/
*/
public
synchronized
boolean
compactMoveChunks
()
{
public
synchronized
boolean
compactMoveChunks
()
{
checkOpen
();
checkOpen
();
if
(
chunks
.
size
()
==
0
)
{
if
(
lastChunk
==
null
)
{
// nothing to do
// nothing to do
return
false
;
return
false
;
}
}
...
@@ -1239,7 +1292,7 @@ public class MVStore {
...
@@ -1239,7 +1292,7 @@ public class MVStore {
int
length
=
c
.
len
*
BLOCK_SIZE
;
int
length
=
c
.
len
*
BLOCK_SIZE
;
buff
.
limit
(
length
);
buff
.
limit
(
length
);
ByteBuffer
readBuff
=
fileStore
.
readFully
(
start
,
length
);
ByteBuffer
readBuff
=
fileStore
.
readFully
(
start
,
length
);
Chunk
.
from
Header
(
readBuff
,
start
);
Chunk
.
readChunk
Header
(
readBuff
,
start
);
int
chunkHeaderLen
=
readBuff
.
position
();
int
chunkHeaderLen
=
readBuff
.
position
();
buff
.
position
(
chunkHeaderLen
);
buff
.
position
(
chunkHeaderLen
);
buff
.
put
(
readBuff
);
buff
.
put
(
readBuff
);
...
@@ -1247,10 +1300,11 @@ public class MVStore {
...
@@ -1247,10 +1300,11 @@ public class MVStore {
fileStore
.
markUsed
(
end
,
length
);
fileStore
.
markUsed
(
end
,
length
);
fileStore
.
free
(
start
,
length
);
fileStore
.
free
(
start
,
length
);
c
.
block
=
end
/
BLOCK_SIZE
;
c
.
block
=
end
/
BLOCK_SIZE
;
c
.
next
=
0
;
buff
.
position
(
0
);
buff
.
position
(
0
);
c
.
writeHeader
(
buff
,
chunkHeaderLen
);
c
.
write
Chunk
Header
(
buff
,
chunkHeaderLen
);
buff
.
position
(
length
-
CHUNK_FOOTER_LENGTH
);
buff
.
position
(
length
-
CHUNK_FOOTER_LENGTH
);
buff
.
put
(
getChunk
FooterBytes
());
buff
.
put
(
lastChunk
.
get
FooterBytes
());
buff
.
position
(
0
);
buff
.
position
(
0
);
write
(
end
,
buff
.
getBuffer
());
write
(
end
,
buff
.
getBuffer
());
releaseWriteBuffer
(
buff
);
releaseWriteBuffer
(
buff
);
...
@@ -1278,7 +1332,7 @@ public class MVStore {
...
@@ -1278,7 +1332,7 @@ public class MVStore {
int
length
=
c
.
len
*
BLOCK_SIZE
;
int
length
=
c
.
len
*
BLOCK_SIZE
;
buff
.
limit
(
length
);
buff
.
limit
(
length
);
ByteBuffer
readBuff
=
fileStore
.
readFully
(
start
,
length
);
ByteBuffer
readBuff
=
fileStore
.
readFully
(
start
,
length
);
Chunk
.
from
Header
(
readBuff
,
0
);
Chunk
.
readChunk
Header
(
readBuff
,
0
);
int
chunkHeaderLen
=
readBuff
.
position
();
int
chunkHeaderLen
=
readBuff
.
position
();
buff
.
position
(
chunkHeaderLen
);
buff
.
position
(
chunkHeaderLen
);
buff
.
put
(
readBuff
);
buff
.
put
(
readBuff
);
...
@@ -1286,9 +1340,9 @@ public class MVStore {
...
@@ -1286,9 +1340,9 @@ public class MVStore {
fileStore
.
free
(
start
,
length
);
fileStore
.
free
(
start
,
length
);
buff
.
position
(
0
);
buff
.
position
(
0
);
c
.
block
=
pos
/
BLOCK_SIZE
;
c
.
block
=
pos
/
BLOCK_SIZE
;
c
.
writeHeader
(
buff
,
chunkHeaderLen
);
c
.
write
Chunk
Header
(
buff
,
chunkHeaderLen
);
buff
.
position
(
length
-
CHUNK_FOOTER_LENGTH
);
buff
.
position
(
length
-
CHUNK_FOOTER_LENGTH
);
buff
.
put
(
getChunk
FooterBytes
());
buff
.
put
(
lastChunk
.
get
FooterBytes
());
buff
.
position
(
0
);
buff
.
position
(
0
);
write
(
pos
,
buff
.
getBuffer
());
write
(
pos
,
buff
.
getBuffer
());
releaseWriteBuffer
(
buff
);
releaseWriteBuffer
(
buff
);
...
@@ -1316,50 +1370,57 @@ public class MVStore {
...
@@ -1316,50 +1370,57 @@ public class MVStore {
}
}
/**
/**
* Try to reduce the file size by re-writing partially full chunks. Chunks
* Try to increase the fill rate by re-writing partially full chunks. Chunks
* with a low number of live items are re-written. If the current fill rate
* with a low number of live items are re-written.
* is higher than the target fill rate, nothing is done.
* <p>
* If the current fill rate is higher than the target fill rate, nothing is
* done. If not at least a minimum amount of space can be saved, nothing is
* done.
* <p>
* Please note this method will not necessarily reduce the file size, as
* empty chunks are not overwritten.
* <p>
* <p>
* Only data of open maps can be moved. For maps that are not open, the old
* Only data of open maps can be moved. For maps that are not open, the old
* chunk is still referenced. Therefore, it is recommended to open all maps
* chunk is still referenced. Therefore, it is recommended to open all maps
* before calling this method.
* before calling this method.
*
*
* @param fillRate the minimum percentage of live entries
* @param targetFillRate the minimum percentage of live entries
* @return if anything was written
* @param minSaving the minimum amount of saved space
* @return if a chunk was re-written
*/
*/
public
synchronized
boolean
compact
(
int
fillRate
)
{
public
synchronized
boolean
compact
(
int
targetFillRate
,
int
minSaving
)
{
checkOpen
();
checkOpen
();
if
(
chunks
.
size
()
==
0
)
{
if
(
lastChunk
==
null
)
{
// nothing to do
// nothing to do
return
false
;
return
false
;
}
}
long
maxLengthSum
=
0
,
maxLengthLiveSum
=
0
;
// calculate the fill rate
long
maxLengthSum
=
0
;
long
maxLengthLiveSum
=
0
;
for
(
Chunk
c
:
chunks
.
values
())
{
for
(
Chunk
c
:
chunks
.
values
())
{
maxLengthSum
+=
c
.
maxLength
;
maxLengthSum
+=
c
.
maxLength
;
maxLengthLiveSum
+=
c
.
maxLenLive
;
maxLengthLiveSum
+=
c
.
maxLenLive
;
}
}
// the fill rate of all chunks combined
if
(
maxLengthSum
<=
0
)
{
if
(
maxLengthSum
<=
0
)
{
// avoid division by 0
// avoid division by 0
maxLengthSum
=
1
;
maxLengthSum
=
1
;
}
}
// the fill rate of all chunks combined
int
fillRate
=
(
int
)
(
100
*
maxLengthLiveSum
/
maxLengthSum
);
int
totalChunkFillRate
=
(
int
)
(
100
*
maxLengthLiveSum
/
maxLengthSum
);
if
(
fillRate
>=
targetFillRate
)
{
if
(
totalChunkFillRate
>
fillRate
)
{
return
false
;
return
false
;
}
}
// calculate the average max length
int
averageMaxLength
=
(
int
)
(
maxLengthSum
/
chunks
.
size
());
long
time
=
getTime
();
long
time
=
getTime
();
// the 'old' list contains the chunks we want to free up
// the 'old' list contains the chunks we want to free up
ArrayList
<
Chunk
>
old
=
New
.
arrayList
();
ArrayList
<
Chunk
>
old
=
New
.
arrayList
();
Chunk
last
=
chunks
.
get
(
lastChunk
.
id
);
for
(
Chunk
c
:
chunks
.
values
())
{
for
(
Chunk
c
:
chunks
.
values
())
{
if
(
canOverwriteChunk
(
c
,
time
))
{
if
(
canOverwriteChunk
(
c
,
time
))
{
int
age
=
lastChunkId
-
c
.
id
+
1
;
long
age
=
last
.
version
-
c
.
version
+
1
;
c
.
collectPriority
=
c
.
getFillRate
()
/
age
;
c
.
collectPriority
=
(
int
)
(
c
.
getFillRate
()
/
age
)
;
old
.
add
(
c
);
old
.
add
(
c
);
}
}
}
}
...
@@ -1377,16 +1438,21 @@ public class MVStore {
...
@@ -1377,16 +1438,21 @@ public class MVStore {
});
});
// find out up to were in the old list we need to move
// find out up to were in the old list we need to move
// try to move one (average sized) chunk
long
saved
=
0
;
long
moved
=
0
;
Chunk
move
=
null
;
Chunk
move
=
null
;
for
(
Chunk
c
:
old
)
{
for
(
Chunk
c
:
old
)
{
if
(
move
!=
null
&&
moved
+
c
.
maxLenLive
>
averageMaxLength
)
{
long
save
=
c
.
maxLength
-
c
.
maxLenLive
;
break
;
if
(
move
!=
null
)
{
if
(
saved
>
minSaving
)
{
break
;
}
}
}
moved
+=
c
.
maxLenLi
ve
;
saved
+=
sa
ve
;
move
=
c
;
move
=
c
;
}
}
if
(
saved
<
minSaving
)
{
return
false
;
}
// remove the chunks we want to keep from this list
// remove the chunks we want to keep from this list
boolean
remove
=
false
;
boolean
remove
=
false
;
...
@@ -1412,7 +1478,7 @@ public class MVStore {
...
@@ -1412,7 +1478,7 @@ public class MVStore {
long
start
=
chunk
.
block
*
BLOCK_SIZE
;
long
start
=
chunk
.
block
*
BLOCK_SIZE
;
int
length
=
chunk
.
len
*
BLOCK_SIZE
;
int
length
=
chunk
.
len
*
BLOCK_SIZE
;
ByteBuffer
buff
=
fileStore
.
readFully
(
start
,
length
);
ByteBuffer
buff
=
fileStore
.
readFully
(
start
,
length
);
Chunk
.
from
Header
(
buff
,
start
);
Chunk
.
readChunk
Header
(
buff
,
start
);
int
pagesRemaining
=
chunk
.
pageCount
;
int
pagesRemaining
=
chunk
.
pageCount
;
markMetaChanged
();
markMetaChanged
();
while
(
pagesRemaining
--
>
0
)
{
while
(
pagesRemaining
--
>
0
)
{
...
@@ -1783,27 +1849,27 @@ public class MVStore {
...
@@ -1783,27 +1849,27 @@ public class MVStore {
boolean
loadFromFile
=
false
;
boolean
loadFromFile
=
false
;
// get the largest chunk with a version
// get the largest chunk with a version
// higher or equal the requested version
// higher or equal the requested version
int
removeChunksNewerThan
=
-
1
;
Chunk
removeChunksNewerThan
=
null
;
for
(
int
chunkId
=
lastChunkId
;;
chunkId
--)
{
Chunk
c
=
lastChunk
;
Chunk
x
=
chunks
.
get
(
chunkId
);
while
(
true
)
{
if
(
x
==
null
)
{
if
(
c
==
null
||
c
.
version
<
version
)
{
break
;
break
;
}
else
if
(
x
.
version
>=
version
)
{
removeChunksNewerThan
=
x
.
id
;
}
}
}
removeChunksNewerThan
=
c
;
if
(
removeChunksNewerThan
>=
0
&&
lastChunkId
>
removeChunksNewerThan
)
{
c
=
chunks
.
get
(
c
.
id
-
1
);
}
Chunk
last
=
lastChunk
;
if
(
removeChunksNewerThan
!=
null
&&
last
.
version
>
removeChunksNewerThan
.
version
)
{
revertTemp
(
version
);
revertTemp
(
version
);
loadFromFile
=
true
;
loadFromFile
=
true
;
Chunk
last
=
null
;
while
(
true
)
{
while
(
true
)
{
last
=
chunks
.
get
(
lastChunkId
)
;
last
=
lastChunk
;
if
(
last
==
null
)
{
if
(
last
==
null
)
{
break
;
break
;
}
else
if
(
last
.
id
<=
removeChunksNewerTha
n
)
{
}
else
if
(
last
.
version
<=
removeChunksNewerThan
.
versio
n
)
{
break
;
break
;
}
}
chunks
.
remove
(
lastChunk
I
d
);
chunks
.
remove
(
lastChunk
.
i
d
);
long
start
=
last
.
block
*
BLOCK_SIZE
;
long
start
=
last
.
block
*
BLOCK_SIZE
;
int
length
=
last
.
len
*
BLOCK_SIZE
;
int
length
=
last
.
len
*
BLOCK_SIZE
;
fileStore
.
free
(
start
,
length
);
fileStore
.
free
(
start
,
length
);
...
@@ -1815,12 +1881,10 @@ public class MVStore {
...
@@ -1815,12 +1881,10 @@ public class MVStore {
Arrays
.
fill
(
buff
.
getBuffer
().
array
(),
(
byte
)
0
);
Arrays
.
fill
(
buff
.
getBuffer
().
array
(),
(
byte
)
0
);
write
(
start
,
buff
.
getBuffer
());
write
(
start
,
buff
.
getBuffer
());
releaseWriteBuffer
(
buff
);
releaseWriteBuffer
(
buff
);
lastChunk
Id
--
;
lastChunk
=
chunks
.
get
(
lastChunk
.
id
-
1
)
;
}
}
lastChunkBlock
=
last
.
block
;
writeFileHeader
();
writeFileHeader
();
readFileHeader
();
readFileHeader
();
readMeta
();
}
}
for
(
MVMap
<?,
?>
m
:
New
.
arrayList
(
maps
.
values
()))
{
for
(
MVMap
<?,
?>
m
:
New
.
arrayList
(
maps
.
values
()))
{
int
id
=
m
.
getId
();
int
id
=
m
.
getId
();
...
@@ -1835,9 +1899,11 @@ public class MVStore {
...
@@ -1835,9 +1899,11 @@ public class MVStore {
}
}
// rollback might have rolled back the stored chunk metadata as well
// rollback might have rolled back the stored chunk metadata as well
Chunk
c
=
chunks
.
get
(
lastChunkId
-
1
);
if
(
lastChunk
!=
null
)
{
if
(
c
!=
null
)
{
c
=
chunks
.
get
(
lastChunk
.
id
-
1
);
meta
.
put
(
Chunk
.
getMetaKey
(
c
.
id
),
c
.
asString
());
if
(
c
!=
null
)
{
meta
.
put
(
Chunk
.
getMetaKey
(
c
.
id
),
c
.
asString
());
}
}
}
currentVersion
=
version
;
currentVersion
=
version
;
setWriteVersion
(
version
);
setWriteVersion
(
version
);
...
...
h2/src/main/org/h2/mvstore/MVStoreTool.java
浏览文件 @
2ed0a04b
...
@@ -85,7 +85,7 @@ public class MVStoreTool {
...
@@ -85,7 +85,7 @@ public class MVStoreTool {
continue
;
continue
;
}
}
block
.
position
(
0
);
block
.
position
(
0
);
Chunk
c
=
Chunk
.
from
Header
(
block
,
pos
);
Chunk
c
=
Chunk
.
readChunk
Header
(
block
,
pos
);
int
length
=
c
.
len
*
MVStore
.
BLOCK_SIZE
;
int
length
=
c
.
len
*
MVStore
.
BLOCK_SIZE
;
pw
.
println
(
" "
+
c
.
toString
());
pw
.
println
(
" "
+
c
.
toString
());
ByteBuffer
chunk
=
ByteBuffer
.
allocate
(
length
);
ByteBuffer
chunk
=
ByteBuffer
.
allocate
(
length
);
...
...
h2/src/main/org/h2/mvstore/OffHeapStore.java
浏览文件 @
2ed0a04b
...
@@ -50,8 +50,7 @@ public class OffHeapStore extends FileStore {
...
@@ -50,8 +50,7 @@ public class OffHeapStore extends FileStore {
freeSpace
.
free
(
pos
,
length
);
freeSpace
.
free
(
pos
,
length
);
ByteBuffer
buff
=
memory
.
remove
(
pos
);
ByteBuffer
buff
=
memory
.
remove
(
pos
);
if
(
buff
==
null
)
{
if
(
buff
==
null
)
{
throw
DataUtils
.
newIllegalStateException
(
DataUtils
.
ERROR_READING_FAILED
,
// nothing was written (just allocated)
"Could not find entry at position {0}"
,
pos
);
}
else
if
(
buff
.
remaining
()
!=
length
)
{
}
else
if
(
buff
.
remaining
()
!=
length
)
{
throw
DataUtils
.
newIllegalStateException
(
DataUtils
.
ERROR_READING_FAILED
,
throw
DataUtils
.
newIllegalStateException
(
DataUtils
.
ERROR_READING_FAILED
,
"Partial remove is not supported at position {0}"
,
pos
);
"Partial remove is not supported at position {0}"
,
pos
);
...
@@ -105,6 +104,7 @@ public class OffHeapStore extends FileStore {
...
@@ -105,6 +104,7 @@ public class OffHeapStore extends FileStore {
memory
.
clear
();
memory
.
clear
();
return
;
return
;
}
}
fileSize
=
size
;
for
(
Iterator
<
Long
>
it
=
memory
.
keySet
().
iterator
();
it
.
hasNext
();)
{
for
(
Iterator
<
Long
>
it
=
memory
.
keySet
().
iterator
();
it
.
hasNext
();)
{
long
pos
=
it
.
next
();
long
pos
=
it
.
next
();
if
(
pos
<
size
)
{
if
(
pos
<
size
)
{
...
...
h2/src/main/org/h2/mvstore/Page.java
浏览文件 @
2ed0a04b
...
@@ -187,7 +187,7 @@ public class Page {
...
@@ -187,7 +187,7 @@ public class Page {
int
length
=
maxLength
;
int
length
=
maxLength
;
if
(
length
<
0
)
{
if
(
length
<
0
)
{
throw
DataUtils
.
newIllegalStateException
(
DataUtils
.
ERROR_FILE_CORRUPT
,
throw
DataUtils
.
newIllegalStateException
(
DataUtils
.
ERROR_FILE_CORRUPT
,
"Illegal page length {0} reading at {1}; file size {
1
} "
,
length
,
filePos
,
fileSize
);
"Illegal page length {0} reading at {1}; file size {
2
} "
,
length
,
filePos
,
fileSize
);
}
}
buff
=
fileStore
.
readFully
(
filePos
,
length
);
buff
=
fileStore
.
readFully
(
filePos
,
length
);
Page
p
=
new
Page
(
map
,
0
);
Page
p
=
new
Page
(
map
,
0
);
...
...
h2/src/main/org/h2/mvstore/db/MVTableEngine.java
浏览文件 @
2ed0a04b
...
@@ -182,7 +182,7 @@ public class MVTableEngine implements TableEngine {
...
@@ -182,7 +182,7 @@ public class MVTableEngine implements TableEngine {
if
(
s
==
null
||
s
.
isReadOnly
())
{
if
(
s
==
null
||
s
.
isReadOnly
())
{
return
;
return
;
}
}
if
(!
store
.
compact
(
50
))
{
if
(!
store
.
compact
(
50
,
1024
*
1024
))
{
store
.
commit
();
store
.
commit
();
}
}
}
}
...
@@ -288,7 +288,7 @@ public class MVTableEngine implements TableEngine {
...
@@ -288,7 +288,7 @@ public class MVTableEngine implements TableEngine {
public
void
compactFile
(
long
maxCompactTime
)
{
public
void
compactFile
(
long
maxCompactTime
)
{
store
.
setRetentionTime
(
0
);
store
.
setRetentionTime
(
0
);
long
start
=
System
.
currentTimeMillis
();
long
start
=
System
.
currentTimeMillis
();
while
(
store
.
compact
(
99
))
{
while
(
store
.
compact
(
99
,
16
*
1024
))
{
store
.
sync
();
store
.
sync
();
long
time
=
System
.
currentTimeMillis
()
-
start
;
long
time
=
System
.
currentTimeMillis
()
-
start
;
if
(
time
>
maxCompactTime
)
{
if
(
time
>
maxCompactTime
)
{
...
@@ -311,7 +311,7 @@ public class MVTableEngine implements TableEngine {
...
@@ -311,7 +311,7 @@ public class MVTableEngine implements TableEngine {
if
(!
store
.
getFileStore
().
isReadOnly
())
{
if
(!
store
.
getFileStore
().
isReadOnly
())
{
transactionStore
.
close
();
transactionStore
.
close
();
long
start
=
System
.
currentTimeMillis
();
long
start
=
System
.
currentTimeMillis
();
while
(
store
.
compact
(
90
))
{
while
(
store
.
compact
(
90
,
32
*
1024
))
{
long
time
=
System
.
currentTimeMillis
()
-
start
;
long
time
=
System
.
currentTimeMillis
()
-
start
;
if
(
time
>
maxCompactTime
)
{
if
(
time
>
maxCompactTime
)
{
break
;
break
;
...
...
h2/src/test/org/h2/test/store/TestBenchmark.java
浏览文件 @
2ed0a04b
...
@@ -69,8 +69,8 @@ public class TestBenchmark extends TestBase {
...
@@ -69,8 +69,8 @@ public class TestBenchmark extends TestBase {
for
(
int
i
=
0
;
i
<
rowCount
;
i
++)
{
for
(
int
i
=
0
;
i
<
rowCount
;
i
++)
{
prep
.
setInt
(
1
,
i
);
prep
.
setInt
(
1
,
i
);
prep
.
setInt
(
2
,
i
);
//
prep.setInt(2, i);
//
prep.setInt(2, r.nextInt());
prep
.
setInt
(
2
,
r
.
nextInt
());
prep
.
execute
();
prep
.
execute
();
if
(
i
%
10000
==
0
)
{
if
(
i
%
10000
==
0
)
{
conn
.
commit
();
conn
.
commit
();
...
...
h2/src/test/org/h2/test/store/TestKillProcessWhileWriting.java
浏览文件 @
2ed0a04b
...
@@ -109,7 +109,7 @@ public class TestKillProcessWhileWriting extends TestBase {
...
@@ -109,7 +109,7 @@ public class TestKillProcessWhileWriting extends TestBase {
s
.
commit
();
s
.
commit
();
break
;
break
;
case
7
:
case
7
:
s
.
compact
(
80
);
s
.
compact
(
80
,
1024
);
break
;
break
;
case
8
:
case
8
:
m
.
clear
();
m
.
clear
();
...
...
h2/src/test/org/h2/test/store/TestMVStore.java
浏览文件 @
2ed0a04b
...
@@ -192,7 +192,7 @@ public class TestMVStore extends TestBase {
...
@@ -192,7 +192,7 @@ public class TestMVStore extends TestBase {
map
.
put
(
i
,
"Hello "
+
i
);
map
.
put
(
i
,
"Hello "
+
i
);
s
.
commit
();
s
.
commit
();
}
}
assertTrue
(
1000
<
offHeap
.
getWriteCount
()
);
assertTrue
(
offHeap
.
getWriteCount
()
>
1000
);
s
.
close
();
s
.
close
();
s
=
new
MVStore
.
Builder
().
s
=
new
MVStore
.
Builder
().
...
@@ -212,6 +212,7 @@ public class TestMVStore extends TestBase {
...
@@ -212,6 +212,7 @@ public class TestMVStore extends TestBase {
encryptionKey
(
"007"
.
toCharArray
()).
encryptionKey
(
"007"
.
toCharArray
()).
fileName
(
fileName
).
fileName
(
fileName
).
open
();
open
();
s
.
setRetentionTime
(
Integer
.
MAX_VALUE
);
Map
<
String
,
Object
>
header
=
s
.
getStoreHeader
();
Map
<
String
,
Object
>
header
=
s
.
getStoreHeader
();
assertEquals
(
"1"
,
header
.
get
(
"format"
).
toString
());
assertEquals
(
"1"
,
header
.
get
(
"format"
).
toString
());
header
.
put
(
"formatRead"
,
"1"
);
header
.
put
(
"formatRead"
,
"1"
);
...
@@ -528,19 +529,20 @@ public class TestMVStore extends TestBase {
...
@@ -528,19 +529,20 @@ public class TestMVStore extends TestBase {
MVStore
s
;
MVStore
s
;
MVMap
<
Integer
,
Integer
>
m
;
MVMap
<
Integer
,
Integer
>
m
;
s
=
openStore
(
fileName
);
s
=
openStore
(
fileName
);
s
.
setRetentionTime
(
Integer
.
MAX_VALUE
);
m
=
s
.
openMap
(
"test"
);
m
=
s
.
openMap
(
"test"
);
m
.
put
(
1
,
1
);
m
.
put
(
1
,
1
);
Map
<
String
,
Object
>
header
=
s
.
getStoreHeader
();
Map
<
String
,
Object
>
header
=
s
.
getStoreHeader
();
int
format
=
Integer
.
parseInt
(
header
.
get
(
"format"
).
toString
());
int
format
=
Integer
.
parseInt
(
header
.
get
(
"format"
).
toString
());
assertEquals
(
1
,
format
);
assertEquals
(
1
,
format
);
header
.
put
(
"format"
,
Integer
.
toString
(
format
+
1
));
header
.
put
(
"format"
,
Integer
.
toString
(
format
+
1
));
// ensure the file header is overwritten
for
(
int
i
=
0
;
i
<
10
;
i
++)
{
s
.
commit
();
if
(
i
>
5
)
{
m
.
put
(
1
,
1
0
);
s
.
setRetentionTime
(
0
);
s
.
commit
();
}
m
.
put
(
1
,
20
);
m
.
put
(
10
,
100
*
i
);
s
.
setRetentionTime
(
0
);
s
.
commit
(
);
s
.
commit
();
}
s
.
close
();
s
.
close
();
try
{
try
{
openStore
(
fileName
).
close
();
openStore
(
fileName
).
close
();
...
@@ -662,6 +664,7 @@ public class TestMVStore extends TestBase {
...
@@ -662,6 +664,7 @@ public class TestMVStore extends TestBase {
private
void
testFileHeader
()
{
private
void
testFileHeader
()
{
String
fileName
=
getBaseDir
()
+
"/testFileHeader.h3"
;
String
fileName
=
getBaseDir
()
+
"/testFileHeader.h3"
;
MVStore
s
=
openStore
(
fileName
);
MVStore
s
=
openStore
(
fileName
);
s
.
setRetentionTime
(
Integer
.
MAX_VALUE
);
long
time
=
System
.
currentTimeMillis
();
long
time
=
System
.
currentTimeMillis
();
Map
<
String
,
Object
>
m
=
s
.
getStoreHeader
();
Map
<
String
,
Object
>
m
=
s
.
getStoreHeader
();
assertEquals
(
"1"
,
m
.
get
(
"format"
).
toString
());
assertEquals
(
"1"
,
m
.
get
(
"format"
).
toString
());
...
@@ -671,12 +674,13 @@ public class TestMVStore extends TestBase {
...
@@ -671,12 +674,13 @@ public class TestMVStore extends TestBase {
MVMap
<
Integer
,
Integer
>
map
=
s
.
openMap
(
"test"
);
MVMap
<
Integer
,
Integer
>
map
=
s
.
openMap
(
"test"
);
map
.
put
(
10
,
100
);
map
.
put
(
10
,
100
);
// ensure the file header is overwritten
// ensure the file header is overwritten
s
.
commit
();
for
(
int
i
=
0
;
i
<
10
;
i
++)
{
map
.
put
(
10
,
110
);
if
(
i
>
5
)
{
s
.
commit
();
s
.
setRetentionTime
(
0
);
map
.
put
(
1
,
120
);
}
s
.
setRetentionTime
(
0
);
map
.
put
(
10
,
110
);
s
.
commit
();
s
.
commit
();
}
s
.
close
();
s
.
close
();
s
=
openStore
(
fileName
);
s
=
openStore
(
fileName
);
assertEquals
(
"123"
,
s
.
getStoreHeader
().
get
(
"test"
).
toString
());
assertEquals
(
"123"
,
s
.
getStoreHeader
().
get
(
"test"
).
toString
());
...
@@ -701,7 +705,7 @@ public class TestMVStore extends TestBase {
...
@@ -701,7 +705,7 @@ public class TestMVStore extends TestBase {
map
=
s
.
openMap
(
"test"
+
i
);
map
=
s
.
openMap
(
"test"
+
i
);
s
.
removeMap
(
map
);
s
.
removeMap
(
map
);
s
.
commit
();
s
.
commit
();
s
.
compact
(
100
);
s
.
compact
(
100
,
1
);
if
(
fs
.
getFile
().
size
()
<=
size
)
{
if
(
fs
.
getFile
().
size
()
<=
size
)
{
break
;
break
;
}
}
...
@@ -1058,6 +1062,7 @@ public class TestMVStore extends TestBase {
...
@@ -1058,6 +1062,7 @@ public class TestMVStore extends TestBase {
FileUtils
.
delete
(
fileName
);
FileUtils
.
delete
(
fileName
);
MVStore
s
;
MVStore
s
;
s
=
openStore
(
fileName
);
s
=
openStore
(
fileName
);
s
.
setRetentionTime
(
Integer
.
MAX_VALUE
);
MVMap
<
String
,
String
>
m
;
MVMap
<
String
,
String
>
m
;
m
=
s
.
openMap
(
"data"
);
m
=
s
.
openMap
(
"data"
);
long
first
=
s
.
getCurrentVersion
();
long
first
=
s
.
getCurrentVersion
();
...
@@ -1121,16 +1126,24 @@ public class TestMVStore extends TestBase {
...
@@ -1121,16 +1126,24 @@ public class TestMVStore extends TestBase {
MVMap
<
Integer
,
String
>
m
;
MVMap
<
Integer
,
String
>
m
;
s
=
openStore
(
fileName
);
s
=
openStore
(
fileName
);
m
=
s
.
openMap
(
"data"
);
m
=
s
.
openMap
(
"data"
);
for
(
int
i
=
0
;
i
<
1000
;
i
++)
{
String
data
=
new
String
(
new
char
[
10000
]).
replace
((
char
)
0
,
'x'
);
m
.
put
(
i
,
"Hello World"
);
for
(
int
i
=
1
;
i
<
10
;
i
++)
{
m
.
put
(
i
,
data
);
s
.
commit
();
}
}
s
.
close
();
s
.
close
();
long
len
=
FileUtils
.
size
(
fileName
);
long
len
=
FileUtils
.
size
(
fileName
);
s
=
openStore
(
fileName
);
s
=
openStore
(
fileName
);
s
.
setRetentionTime
(
0
);
s
.
setRetentionTime
(
0
);
// remove 50%
m
=
s
.
openMap
(
"data"
);
m
=
s
.
openMap
(
"data"
);
m
.
clear
();
for
(
int
i
=
0
;
i
<
10
;
i
+=
2
)
{
s
.
compact
(
100
);
m
.
remove
(
i
);
}
s
.
commit
();
assertTrue
(
s
.
compact
(
100
,
1
));
assertTrue
(
s
.
compact
(
100
,
1
));
assertTrue
(
s
.
compact
(
100
,
1
));
s
.
close
();
s
.
close
();
long
len2
=
FileUtils
.
size
(
fileName
);
long
len2
=
FileUtils
.
size
(
fileName
);
assertTrue
(
"len2: "
+
len2
+
" len: "
+
len
,
len2
<
len
);
assertTrue
(
"len2: "
+
len2
+
" len: "
+
len
,
len2
<
len
);
...
@@ -1160,7 +1173,7 @@ public class TestMVStore extends TestBase {
...
@@ -1160,7 +1173,7 @@ public class TestMVStore extends TestBase {
s
.
commit
();
s
.
commit
();
// ensure only nodes are read, but not leaves
// ensure only nodes are read, but not leaves
assertEquals
(
40
,
s
.
getFileStore
().
getReadCount
());
assertEquals
(
40
,
s
.
getFileStore
().
getReadCount
());
assert
Equals
(
1
,
s
.
getFileStore
().
getWriteCount
()
);
assert
True
(
s
.
getFileStore
().
getWriteCount
()
<
5
);
s
.
close
();
s
.
close
();
}
}
...
@@ -1209,6 +1222,7 @@ public class TestMVStore extends TestBase {
...
@@ -1209,6 +1222,7 @@ public class TestMVStore extends TestBase {
s
.
close
();
s
.
close
();
s
=
openStore
(
fileName
);
s
=
openStore
(
fileName
);
s
.
setRetentionTime
(
45000
);
assertEquals
(
2
,
s
.
getCurrentVersion
());
assertEquals
(
2
,
s
.
getCurrentVersion
());
meta
=
s
.
getMetaMap
();
meta
=
s
.
getMetaMap
();
m
=
s
.
openMap
(
"data"
);
m
=
s
.
openMap
(
"data"
);
...
@@ -1231,6 +1245,7 @@ public class TestMVStore extends TestBase {
...
@@ -1231,6 +1245,7 @@ public class TestMVStore extends TestBase {
s
.
close
();
s
.
close
();
s
=
openStore
(
fileName
);
s
=
openStore
(
fileName
);
s
.
setRetentionTime
(
45000
);
assertEquals
(
2
,
s
.
getCurrentVersion
());
assertEquals
(
2
,
s
.
getCurrentVersion
());
meta
=
s
.
getMetaMap
();
meta
=
s
.
getMetaMap
();
assertTrue
(
meta
.
get
(
"name.data"
)
!=
null
);
assertTrue
(
meta
.
get
(
"name.data"
)
!=
null
);
...
@@ -1248,12 +1263,14 @@ public class TestMVStore extends TestBase {
...
@@ -1248,12 +1263,14 @@ public class TestMVStore extends TestBase {
s
.
close
();
s
.
close
();
s
=
openStore
(
fileName
);
s
=
openStore
(
fileName
);
s
.
setRetentionTime
(
45000
);
assertEquals
(
3
,
s
.
getCurrentVersion
());
assertEquals
(
3
,
s
.
getCurrentVersion
());
m
=
s
.
openMap
(
"data"
);
m
=
s
.
openMap
(
"data"
);
m
.
put
(
"1"
,
"Hi"
);
m
.
put
(
"1"
,
"Hi"
);
s
.
close
();
s
.
close
();
s
=
openStore
(
fileName
);
s
=
openStore
(
fileName
);
s
.
setRetentionTime
(
45000
);
m
=
s
.
openMap
(
"data"
);
m
=
s
.
openMap
(
"data"
);
assertEquals
(
"Hi"
,
m
.
get
(
"1"
));
assertEquals
(
"Hi"
,
m
.
get
(
"1"
));
s
.
rollbackTo
(
v3
);
s
.
rollbackTo
(
v3
);
...
@@ -1261,6 +1278,7 @@ public class TestMVStore extends TestBase {
...
@@ -1261,6 +1278,7 @@ public class TestMVStore extends TestBase {
s
.
close
();
s
.
close
();
s
=
openStore
(
fileName
);
s
=
openStore
(
fileName
);
s
.
setRetentionTime
(
45000
);
m
=
s
.
openMap
(
"data"
);
m
=
s
.
openMap
(
"data"
);
assertEquals
(
"Hallo"
,
m
.
get
(
"1"
));
assertEquals
(
"Hallo"
,
m
.
get
(
"1"
));
s
.
close
();
s
.
close
();
...
@@ -1310,6 +1328,7 @@ public class TestMVStore extends TestBase {
...
@@ -1310,6 +1328,7 @@ public class TestMVStore extends TestBase {
String
fileName
=
getBaseDir
()
+
"/testMeta.h3"
;
String
fileName
=
getBaseDir
()
+
"/testMeta.h3"
;
FileUtils
.
delete
(
fileName
);
FileUtils
.
delete
(
fileName
);
MVStore
s
=
openStore
(
fileName
);
MVStore
s
=
openStore
(
fileName
);
s
.
setRetentionTime
(
Integer
.
MAX_VALUE
);
MVMap
<
String
,
String
>
m
=
s
.
getMetaMap
();
MVMap
<
String
,
String
>
m
=
s
.
getMetaMap
();
assertEquals
(
"[]"
,
s
.
getMapNames
().
toString
());
assertEquals
(
"[]"
,
s
.
getMapNames
().
toString
());
MVMap
<
String
,
String
>
data
=
s
.
openMap
(
"data"
);
MVMap
<
String
,
String
>
data
=
s
.
openMap
(
"data"
);
...
@@ -1480,9 +1499,8 @@ public class TestMVStore extends TestBase {
...
@@ -1480,9 +1499,8 @@ public class TestMVStore extends TestBase {
chunkCount1
++;
chunkCount1
++;
}
}
}
}
s
.
compact
(
80
,
1
);
assertTrue
(
s
.
compact
(
80
));
s
.
compact
(
80
,
1
);
assertTrue
(
s
.
compact
(
80
));
int
chunkCount2
=
0
;
int
chunkCount2
=
0
;
for
(
String
k
:
meta
.
keySet
())
{
for
(
String
k
:
meta
.
keySet
())
{
...
@@ -1493,8 +1511,8 @@ public class TestMVStore extends TestBase {
...
@@ -1493,8 +1511,8 @@ public class TestMVStore extends TestBase {
assertTrue
(
chunkCount2
>=
chunkCount1
);
assertTrue
(
chunkCount2
>=
chunkCount1
);
m
=
s
.
openMap
(
"data"
);
m
=
s
.
openMap
(
"data"
);
assertTrue
(
s
.
compact
(
80
));
assertTrue
(
s
.
compact
(
80
,
16
*
1024
));
assertTrue
(
s
.
compact
(
80
));
assertTrue
(
s
.
compact
(
80
,
1024
));
int
chunkCount3
=
0
;
int
chunkCount3
=
0
;
for
(
String
k
:
meta
.
keySet
())
{
for
(
String
k
:
meta
.
keySet
())
{
...
@@ -1522,7 +1540,7 @@ public class TestMVStore extends TestBase {
...
@@ -1522,7 +1540,7 @@ public class TestMVStore extends TestBase {
for
(
int
i
=
0
;
i
<
100
;
i
++)
{
for
(
int
i
=
0
;
i
<
100
;
i
++)
{
m
.
put
(
j
+
i
,
"Hello "
+
j
);
m
.
put
(
j
+
i
,
"Hello "
+
j
);
}
}
s
.
compact
(
80
);
s
.
compact
(
80
,
1024
);
s
.
close
();
s
.
close
();
long
len
=
FileUtils
.
size
(
fileName
);
long
len
=
FileUtils
.
size
(
fileName
);
// System.out.println(" len:" + len);
// System.out.println(" len:" + len);
...
@@ -1539,13 +1557,13 @@ public class TestMVStore extends TestBase {
...
@@ -1539,13 +1557,13 @@ public class TestMVStore extends TestBase {
for
(
int
i
=
0
;
i
<
100
;
i
++)
{
for
(
int
i
=
0
;
i
<
100
;
i
++)
{
m
.
remove
(
i
);
m
.
remove
(
i
);
}
}
s
.
compact
(
80
);
s
.
compact
(
80
,
1024
);
s
.
close
();
s
.
close
();
// len = FileUtils.size(fileName);
// len = FileUtils.size(fileName);
// System.out.println("len1: " + len);
// System.out.println("len1: " + len);
s
=
openStore
(
fileName
);
s
=
openStore
(
fileName
);
m
=
s
.
openMap
(
"data"
);
m
=
s
.
openMap
(
"data"
);
s
.
compact
(
80
);
s
.
compact
(
80
,
1024
);
s
.
close
();
s
.
close
();
// len = FileUtils.size(fileName);
// len = FileUtils.size(fileName);
// System.out.println("len2: " + len);
// System.out.println("len2: " + len);
...
...
h2/src/test/org/h2/test/store/TestRandomMapOps.java
浏览文件 @
2ed0a04b
...
@@ -45,16 +45,17 @@ public class TestRandomMapOps extends TestBase {
...
@@ -45,16 +45,17 @@ public class TestRandomMapOps extends TestBase {
testMap
(
"memFS:randomOps.h3"
);
testMap
(
"memFS:randomOps.h3"
);
}
}
private
void
testMap
(
String
fileName
)
{
private
void
testMap
(
String
fileName
)
throws
Exception
{
this
.
fileName
=
fileName
;
this
.
fileName
=
fileName
;
int
best
=
Integer
.
MAX_VALUE
;
int
best
=
Integer
.
MAX_VALUE
;
int
bestSeed
=
0
;
int
bestSeed
=
0
;
Throwable
failException
=
null
;
Throwable
failException
=
null
;
int
size
=
getSize
(
100
,
1000
);
for
(
seed
=
0
;
seed
<
100
;
seed
++)
{
for
(
seed
=
0
;
seed
<
100
;
seed
++)
{
FileUtils
.
delete
(
fileName
);
FileUtils
.
delete
(
fileName
);
Throwable
ex
=
null
;
Throwable
ex
=
null
;
try
{
try
{
test
Case
(
);
test
Ops
(
size
);
continue
;
continue
;
}
catch
(
Exception
e
)
{
}
catch
(
Exception
e
)
{
ex
=
e
;
ex
=
e
;
...
@@ -65,6 +66,7 @@ public class TestRandomMapOps extends TestBase {
...
@@ -65,6 +66,7 @@ public class TestRandomMapOps extends TestBase {
trace
(
seed
);
trace
(
seed
);
bestSeed
=
seed
;
bestSeed
=
seed
;
best
=
op
;
best
=
op
;
size
=
best
;
failException
=
ex
;
failException
=
ex
;
// System.out.println("seed:" + seed + " op:" + op);
// System.out.println("seed:" + seed + " op:" + op);
}
}
...
@@ -75,7 +77,7 @@ public class TestRandomMapOps extends TestBase {
...
@@ -75,7 +77,7 @@ public class TestRandomMapOps extends TestBase {
}
}
}
}
private
void
test
Case
(
)
throws
Exception
{
private
void
test
Ops
(
int
size
)
throws
Exception
{
FileUtils
.
delete
(
fileName
);
FileUtils
.
delete
(
fileName
);
MVStore
s
;
MVStore
s
;
s
=
openStore
(
fileName
);
s
=
openStore
(
fileName
);
...
@@ -87,7 +89,6 @@ public class TestRandomMapOps extends TestBase {
...
@@ -87,7 +89,6 @@ public class TestRandomMapOps extends TestBase {
}
}
Random
r
=
new
Random
(
seed
);
Random
r
=
new
Random
(
seed
);
op
=
0
;
op
=
0
;
int
size
=
getSize
(
100
,
1000
);
TreeMap
<
Integer
,
byte
[]>
map
=
new
TreeMap
<
Integer
,
byte
[]>();
TreeMap
<
Integer
,
byte
[]>
map
=
new
TreeMap
<
Integer
,
byte
[]>();
for
(;
op
<
size
;
op
++)
{
for
(;
op
<
size
;
op
++)
{
int
k
=
r
.
nextInt
(
100
);
int
k
=
r
.
nextInt
(
100
);
...
@@ -109,8 +110,8 @@ public class TestRandomMapOps extends TestBase {
...
@@ -109,8 +110,8 @@ public class TestRandomMapOps extends TestBase {
map
.
remove
(
k
);
map
.
remove
(
k
);
break
;
break
;
case
6
:
case
6
:
log
(
op
,
k
,
v
,
"s.compact(90)"
);
log
(
op
,
k
,
v
,
"s.compact(90
, 1024
)"
);
s
.
compact
(
90
);
s
.
compact
(
90
,
1024
);
break
;
break
;
case
7
:
case
7
:
log
(
op
,
k
,
v
,
"m.clear()"
);
log
(
op
,
k
,
v
,
"m.clear()"
);
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论