Skip to content
项目
群组
代码片段
帮助
正在加载...
帮助
为 GitLab 提交贡献
登录/注册
切换导航
H
h2database
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
分枝图
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
计划
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
分枝图
统计图
创建新议题
作业
提交
议题看板
打开侧边栏
Administrator
h2database
Commits
dc838843
提交
dc838843
authored
10月 17, 2014
作者:
Thomas Mueller
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Some minor speedup
上级
0338a8e2
隐藏空白字符变更
内嵌
并排
正在显示
1 个修改的文件
包含
31 行增加
和
7 行删除
+31
-7
MinimalPerfectHash.java
h2/src/tools/org/h2/dev/hash/MinimalPerfectHash.java
+31
-7
没有找到文件。
h2/src/tools/org/h2/dev/hash/MinimalPerfectHash.java
浏览文件 @
dc838843
...
...
@@ -76,18 +76,20 @@ public class MinimalPerfectHash<K> {
*/
private
static
final
int
DIVIDE
=
6
;
private
static
final
int
SPEEDUP
=
11
;
/**
* The maximum size of a small bucket (one that is not further split if
* possible).
*/
private
static
final
int
MAX_SIZE
=
1
2
;
private
static
final
int
MAX_SIZE
=
1
4
;
/**
* The maximum offset for hash functions of small buckets. At most that many
* hash functions are tried for the given size.
*/
private
static
final
int
[]
MAX_OFFSETS
=
{
0
,
0
,
8
,
18
,
47
,
123
,
319
,
831
,
2162
,
5622
,
14617
,
38006
,
38006
};
5622
,
14617
,
38006
,
98815
,
256920
,
667993
};
/**
* The output value to split the bucket into many (more than 2) smaller
...
...
@@ -106,6 +108,9 @@ public class MinimalPerfectHash<K> {
private
static
final
SecureRandom
RANDOM
=
new
SecureRandom
();
static
{
for
(
int
i
=
SPEEDUP
;
i
<
MAX_OFFSETS
.
length
;
i
++)
{
MAX_OFFSETS
[
i
]
=
(
int
)
(
MAX_OFFSETS
[
i
]
*
2.5
);
}
int
last
=
SPLIT_MANY
+
1
;
for
(
int
i
=
0
;
i
<
MAX_OFFSETS
.
length
;
i
++)
{
SIZE_OFFSETS
[
i
]
=
last
;
...
...
@@ -210,6 +215,15 @@ public class MinimalPerfectHash<K> {
}
else
if
(
n
>
SPLIT_MANY
)
{
int
size
=
getSize
(
n
);
int
offset
=
getOffset
(
n
,
size
);
if
(
size
>=
SPEEDUP
)
{
int
p
=
offset
%
(
size
+
1
);
offset
=
offset
/
(
size
+
1
);
int
result
=
hash
(
x
,
hash
,
level
,
seed
,
offset
,
size
+
1
);
if
(
result
>=
p
)
{
result
--;
}
return
result
;
}
return
hash
(
x
,
hash
,
level
,
seed
,
offset
,
size
);
}
pos
++;
...
...
@@ -344,7 +358,7 @@ public class MinimalPerfectHash<K> {
}
if
(
size
<=
MAX_SIZE
)
{
int
maxOffset
=
MAX_OFFSETS
[
size
];
// get the hash codes - we could stop early
// get the hash codes - we could stop early
// if we detect that two keys have the same hash
int
[]
hashes
=
new
int
[
size
];
for
(
int
i
=
0
;
i
<
size
;
i
++)
{
...
...
@@ -354,18 +368,28 @@ public class MinimalPerfectHash<K> {
// to make the hash code unique within this group -
// there might be a much faster way than that, by
// checking which bits of the hash code matter most
int
testSize
=
size
;
if
(
size
>=
SPEEDUP
)
{
testSize
++;
maxOffset
/=
testSize
;
}
nextOffset:
for
(
int
offset
=
0
;
offset
<
maxOffset
;
offset
++)
{
int
bits
=
0
;
for
(
int
i
=
0
;
i
<
size
;
i
++)
{
int
x
=
hashes
[
i
];
int
h
=
hash
(
x
,
level
,
offset
,
s
ize
);
int
h
=
hash
(
x
,
level
,
offset
,
testS
ize
);
if
((
bits
&
(
1
<<
h
))
!=
0
)
{
continue
nextOffset
;
}
bits
|=
1
<<
h
;
}
writeSizeOffset
(
out
,
size
,
offset
);
if
(
size
>=
SPEEDUP
)
{
int
pos
=
Integer
.
numberOfTrailingZeros
(~
bits
);
writeSizeOffset
(
out
,
size
,
offset
*
(
size
+
1
)
+
pos
);
}
else
{
writeSizeOffset
(
out
,
size
,
offset
);
}
return
;
}
}
...
...
@@ -491,7 +515,7 @@ public class MinimalPerfectHash<K> {
private
static
<
K
>
int
hash
(
K
o
,
UniversalHash
<
K
>
hash
,
int
level
,
int
seed
,
int
offset
,
int
size
)
{
int
x
=
hash
.
hashCode
(
o
,
level
,
seed
);
x
+=
level
+
offset
*
16
;
x
+=
level
+
offset
*
32
;
x
=
((
x
>>>
16
)
^
x
)
*
0x45d9f3b
;
x
=
((
x
>>>
16
)
^
x
)
*
0x45d9f3b
;
x
=
(
x
>>>
16
)
^
x
;
...
...
@@ -499,7 +523,7 @@ public class MinimalPerfectHash<K> {
}
private
static
<
K
>
int
hash
(
int
x
,
int
level
,
int
offset
,
int
size
)
{
x
+=
level
+
offset
*
16
;
x
+=
level
+
offset
*
32
;
x
=
((
x
>>>
16
)
^
x
)
*
0x45d9f3b
;
x
=
((
x
>>>
16
)
^
x
)
*
0x45d9f3b
;
x
=
(
x
>>>
16
)
^
x
;
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论