Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
OpenCV
opencv
提交
3935e136
O
opencv
项目概览
OpenCV
/
opencv
上一次同步 9 个月
通知
992
Star
71100
Fork
55581
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
O
opencv
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
前往新版Gitcode,体验更适合开发者的 AI 搜索 >>
提交
3935e136
编写于
10月 10, 2017
作者:
A
Alexander Alekhin
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
dnn(halide): don't compile Halide via parallel_for_()
To avoid problem with reduced stack size of inner threads.
上级
556768ea
变更
1
隐藏空白更改
内联
并排
Showing
1 changed file
with
28 addition
and
32 deletion
+28
-32
modules/dnn/src/dnn.cpp
modules/dnn/src/dnn.cpp
+28
-32
未找到文件。
modules/dnn/src/dnn.cpp
浏览文件 @
3935e136
...
...
@@ -589,33 +589,7 @@ struct Net::Impl
return
wrapper
;
}
class
HalideCompiler
:
public
ParallelLoopBody
{
public:
HalideCompiler
(
const
MapIdToLayerData
&
layers_
,
int
preferableTarget_
)
:
layers
(
&
layers_
),
preferableTarget
(
preferableTarget_
)
{}
void
operator
()(
const
Range
&
r
)
const
{
MapIdToLayerData
::
const_iterator
it
=
layers
->
begin
();
for
(
int
i
=
0
;
i
<
r
.
start
&&
it
!=
layers
->
end
();
++
i
,
++
it
)
{}
for
(
int
i
=
r
.
start
;
i
<
r
.
end
&&
it
!=
layers
->
end
();
++
i
,
++
it
)
{
const
LayerData
&
ld
=
it
->
second
;
Ptr
<
Layer
>
layer
=
ld
.
layerInstance
;
bool
skip
=
ld
.
skipFlags
.
find
(
DNN_BACKEND_HALIDE
)
->
second
;
if
(
layer
->
supportBackend
(
DNN_BACKEND_HALIDE
)
&&
!
skip
)
{
Ptr
<
BackendNode
>
node
=
ld
.
backendNodes
.
find
(
DNN_BACKEND_HALIDE
)
->
second
;
dnn
::
compileHalide
(
ld
.
outputBlobs
,
node
,
preferableTarget
);
}
}
}
private:
const
MapIdToLayerData
*
layers
;
int
preferableTarget
;
};
#ifdef HAVE_HALIDE
void
compileHalide
()
{
CV_TRACE_FUNCTION
();
...
...
@@ -623,8 +597,8 @@ struct Net::Impl
CV_Assert
(
preferableBackend
==
DNN_BACKEND_HALIDE
);
HalideScheduler
scheduler
(
halideConfigFile
);
MapIdToLayerData
::
iterator
it
;
for
(
it
=
layers
.
begin
();
it
!=
layers
.
end
();
++
it
)
std
::
vector
<
std
::
reference_wrapper
<
LayerData
>
>
compileList
;
compileList
.
reserve
(
64
)
;
for
(
MapIdToLayerData
::
iterator
it
=
layers
.
begin
();
it
!=
layers
.
end
();
++
it
)
{
LayerData
&
ld
=
it
->
second
;
Ptr
<
Layer
>
layer
=
ld
.
layerInstance
;
...
...
@@ -639,10 +613,30 @@ struct Net::Impl
ld
.
inputBlobs
,
ld
.
outputBlobs
,
preferableTarget
);
}
compileList
.
emplace_back
(
ld
);
}
}
parallel_for_
(
Range
(
0
,
layers
.
size
()),
HalideCompiler
(
layers
,
preferableTarget
));
std
::
atomic
<
int
>
progress
(
0
);
auto
fn
=
([
&
]
()
->
void
{
for
(;;)
{
int
id
=
progress
.
fetch_add
(
1
);
if
((
size_t
)
id
>=
compileList
.
size
())
return
;
const
LayerData
&
ld
=
compileList
[
id
].
get
();
Ptr
<
BackendNode
>
node
=
ld
.
backendNodes
.
find
(
DNN_BACKEND_HALIDE
)
->
second
;
dnn
::
compileHalide
(
ld
.
outputBlobs
,
node
,
preferableTarget
);
}
});
size_t
num_threads
=
std
::
min
(
compileList
.
size
(),
(
size_t
)
std
::
thread
::
hardware_concurrency
());
num_threads
=
std
::
max
((
size_t
)
1u
,
std
::
min
((
size_t
)
8u
,
num_threads
));
std
::
vector
<
std
::
thread
>
threads
(
num_threads
-
1
);
for
(
auto
&
t
:
threads
)
t
=
std
::
thread
(
fn
);
fn
();
// process own tasks
for
(
auto
&
t
:
threads
)
t
.
join
();
}
#endif
void
clear
()
{
...
...
@@ -692,10 +686,12 @@ struct Net::Impl
if
(
!
netWasAllocated
)
{
// If user didn't call compileHalide() between
// setPreferableBackend(DNN_BACKEND_HALIDE) and forward().
#ifdef HAVE_HALIDE
if
(
preferableBackend
==
DNN_BACKEND_HALIDE
)
compileHalide
();
#else
CV_Assert
(
preferableBackend
!=
DNN_BACKEND_HALIDE
);
#endif
}
netWasAllocated
=
true
;
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录