Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
OpenCV
opencv
提交
9d02d42a
O
opencv
项目概览
OpenCV
/
opencv
上一次同步 9 个月
通知
992
Star
71100
Fork
55581
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
O
opencv
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
前往新版Gitcode,体验更适合开发者的 AI 搜索 >>
提交
9d02d42a
编写于
10月 05, 2018
作者:
A
Alexander Alekhin
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
dnn(ocl4dnn): don't use getUMat()
especially in CPU only processing
上级
0101fa78
变更
7
隐藏空白更改
内联
并排
Showing
7 changed file
with
48 addition
and
20 deletion
+48
-20
modules/dnn/src/dnn.cpp
modules/dnn/src/dnn.cpp
+2
-2
modules/dnn/src/layers/batch_norm_layer.cpp
modules/dnn/src/layers/batch_norm_layer.cpp
+2
-2
modules/dnn/src/layers/convolution_layer.cpp
modules/dnn/src/layers/convolution_layer.cpp
+4
-2
modules/dnn/src/layers/elementwise_layers.cpp
modules/dnn/src/layers/elementwise_layers.cpp
+5
-1
modules/dnn/src/layers/fully_connected_layer.cpp
modules/dnn/src/layers/fully_connected_layer.cpp
+8
-8
modules/dnn/src/layers/mvn_layer.cpp
modules/dnn/src/layers/mvn_layer.cpp
+21
-4
modules/dnn/src/layers/region_layer.cpp
modules/dnn/src/layers/region_layer.cpp
+6
-1
未找到文件。
modules/dnn/src/dnn.cpp
浏览文件 @
9d02d42a
...
...
@@ -2518,7 +2518,7 @@ void Net::forward(OutputArrayOfArrays outputBlobs, const String& outputName)
if
(
outputBlobs
.
isUMat
())
{
outputBlobs
.
assign
(
impl
->
getBlob
(
layerName
).
getUMat
(
ACCESS_RW
)
);
impl
->
getBlob
(
layerName
).
copyTo
(
outputBlobs
);
}
else
if
(
outputBlobs
.
isMat
())
{
...
...
@@ -2566,7 +2566,7 @@ void Net::forward(OutputArrayOfArrays outputBlobs, const String& outputName)
{
outputvec
.
resize
(
ld
.
outputBlobs
.
size
());
for
(
int
i
=
0
;
i
<
outputvec
.
size
();
++
i
)
outputvec
[
i
]
=
ld
.
outputBlobs
[
i
].
getUMat
(
ACCESS_RW
);
ld
.
outputBlobs
[
i
].
copyTo
(
outputvec
[
i
]
);
}
}
}
...
...
modules/dnn/src/layers/batch_norm_layer.cpp
浏览文件 @
9d02d42a
...
...
@@ -172,8 +172,8 @@ public:
if
(
umat_weight
.
empty
())
{
umat_weight
=
weights_
.
getUMat
(
ACCESS_READ
);
umat_bias
=
bias_
.
getUMat
(
ACCESS_READ
);
weights_
.
copyTo
(
umat_weight
);
bias_
.
copyTo
(
umat_bias
);
}
UMat
&
inpBlob
=
inputs
[
0
];
...
...
modules/dnn/src/layers/convolution_layer.cpp
浏览文件 @
9d02d42a
...
...
@@ -1452,8 +1452,10 @@ public:
if
(
umat_weights
.
empty
())
{
transpose
(
blobs
[
0
].
reshape
(
1
,
inpCn
),
umat_weights
);
umat_biases
=
hasBias
()
?
blobs
[
1
].
reshape
(
1
,
outCn
).
getUMat
(
ACCESS_READ
)
:
UMat
::
zeros
(
outCn
,
1
,
CV_32F
);
if
(
hasBias
())
blobs
[
1
].
reshape
(
1
,
outCn
).
copyTo
(
umat_biases
);
else
umat_biases
=
UMat
::
zeros
(
outCn
,
1
,
CV_32F
);
}
String
buildopt
=
format
(
"-DT=%s "
,
ocl
::
typeToStr
(
inputs
[
0
].
type
()));
...
...
modules/dnn/src/layers/elementwise_layers.cpp
浏览文件 @
9d02d42a
...
...
@@ -969,11 +969,12 @@ struct ChannelsPReLUFunctor
{
typedef
ChannelsPReLULayer
Layer
;
Mat
scale
;
#ifdef HAVE_OPENCL
UMat
scale_umat
;
#endif
explicit
ChannelsPReLUFunctor
(
const
Mat
&
scale_
=
Mat
())
:
scale
(
scale_
)
{
scale_umat
=
scale
.
getUMat
(
ACCESS_READ
);
}
bool
supportBackend
(
int
backendId
,
int
)
...
...
@@ -1021,6 +1022,9 @@ struct ChannelsPReLUFunctor
#ifdef HAVE_OPENCL
bool
applyOCL
(
InputArrayOfArrays
inps
,
OutputArrayOfArrays
outs
,
OutputArrayOfArrays
internals
)
{
if
(
scale_umat
.
empty
())
scale
.
copyTo
(
scale_umat
);
std
::
vector
<
UMat
>
inputs
;
std
::
vector
<
UMat
>
outputs
;
...
...
modules/dnn/src/layers/fully_connected_layer.cpp
浏览文件 @
9d02d42a
...
...
@@ -96,12 +96,6 @@ public:
biasMat
=
blobs
[
1
]
=
blobs
[
1
].
reshape
(
1
,
1
);
else
biasMat
=
Mat
::
zeros
(
1
,
numOutput
,
weightsMat
.
type
());
#ifdef HAVE_OPENCL
size_t
n
=
blobs
.
size
();
umat_blobs
.
resize
(
n
);
for
(
int
i
=
0
;
i
<
n
;
i
++
)
umat_blobs
[
i
]
=
blobs
[
i
].
getUMat
(
ACCESS_READ
);
#endif
}
bool
getMemoryShapes
(
const
std
::
vector
<
MatShape
>
&
inputs
,
...
...
@@ -276,6 +270,8 @@ public:
virtual
void
finalize
(
InputArrayOfArrays
,
OutputArrayOfArrays
)
CV_OVERRIDE
{
innerProductOp
.
release
();
umat_blobs
.
clear
();
half_blobs
.
clear
();
}
bool
forward_ocl
(
InputArrayOfArrays
inps
,
OutputArrayOfArrays
outs
,
InputArrayOfArrays
internals
)
...
...
@@ -288,13 +284,17 @@ public:
outs
.
getUMatVector
(
outputs
);
int
axisCan
=
clamp
(
axis
,
inputs
[
0
].
dims
);
int
numOutput
=
umat_
blobs
[
0
].
size
[
0
];
int
innerSize
=
umat_
blobs
[
0
].
size
[
1
];
int
numOutput
=
blobs
[
0
].
size
[
0
];
int
innerSize
=
blobs
[
0
].
size
[
1
];
int
outerSize
=
total
(
shape
(
inputs
[
0
]),
0
,
axisCan
);
bool
ret
=
true
;
if
(
innerProductOp
.
empty
())
{
size_t
n
=
blobs
.
size
();
umat_blobs
.
resize
(
n
);
for
(
int
i
=
0
;
i
<
n
;
i
++
)
blobs
[
i
].
copyTo
(
umat_blobs
[
i
]);
OCL4DNNInnerProductConfig
config
;
config
.
num_output
=
numOutput
;
config
.
bias_term
=
bias
;
...
...
modules/dnn/src/layers/mvn_layer.cpp
浏览文件 @
9d02d42a
...
...
@@ -71,6 +71,9 @@ public:
}
Mat
scale
,
shift
;
#ifdef HAVE_OPENCL
UMat
umat_scale
,
umat_shift
;
#endif
bool
fuse_batch_norm
;
Ptr
<
ReLULayer
>
activ_relu
;
...
...
@@ -105,6 +108,10 @@ public:
for
(
i
=
0
;
i
<
splitDim
;
i
++
)
newRows
*=
inputs
[
0
].
size
[
i
];
zeroDev
=
inputs
[
0
].
total
()
==
newRows
;
#ifdef HAVE_OPENCL
umat_scale
.
release
();
umat_shift
.
release
();
#endif
}
virtual
bool
supportBackend
(
int
backendId
)
CV_OVERRIDE
...
...
@@ -118,8 +125,13 @@ public:
#ifdef HAVE_OPENCL
bool
fast_forward_ocl
(
std
::
vector
<
UMat
>
&
inputs
,
std
::
vector
<
UMat
>
&
outputs
)
{
UMat
bnorm_weight
=
scale
.
empty
()
?
UMat
()
:
scale
.
getUMat
(
ACCESS_READ
);
UMat
bnorm_bias
=
shift
.
empty
()
?
UMat
()
:
shift
.
getUMat
(
ACCESS_READ
);
if
(
umat_scale
.
empty
()
&&
!
scale
.
empty
())
scale
.
copyTo
(
umat_scale
);
if
(
umat_shift
.
empty
()
&&
!
shift
.
empty
())
shift
.
copyTo
(
umat_shift
);
UMat
&
bnorm_weight
=
umat_scale
;
UMat
&
bnorm_bias
=
umat_shift
;
bool
use_half
=
(
inputs
[
0
].
depth
()
==
CV_16S
);
String
opts
=
format
(
" -DT=%s -DT4=%s -Dconvert_T=%s"
,
use_half
?
"half"
:
"float"
,
use_half
?
"half4"
:
"float4"
,
use_half
?
"convert_half4"
:
"convert_float4"
);
...
...
@@ -177,6 +189,13 @@ public:
bool
forward_ocl
(
InputArrayOfArrays
inputs_
,
OutputArrayOfArrays
outputs_
,
OutputArrayOfArrays
internals_
)
{
if
(
umat_scale
.
empty
()
&&
!
scale
.
empty
())
scale
.
copyTo
(
umat_scale
);
if
(
umat_shift
.
empty
()
&&
!
shift
.
empty
())
shift
.
copyTo
(
umat_shift
);
UMat
&
bnorm_weight
=
umat_scale
;
UMat
&
bnorm_bias
=
umat_shift
;
std
::
vector
<
UMat
>
inputs
;
std
::
vector
<
UMat
>
outputs
;
...
...
@@ -192,8 +211,6 @@ public:
if
(
inputs
[
0
].
depth
()
==
CV_16S
)
return
false
;
UMat
bnorm_weight
=
scale
.
empty
()
?
UMat
()
:
scale
.
getUMat
(
ACCESS_READ
);
UMat
bnorm_bias
=
shift
.
empty
()
?
UMat
()
:
shift
.
getUMat
(
ACCESS_READ
);
String
opts
=
format
(
" -DT=float -DT4=float4 -Dconvert_T=convert_float4"
);
for
(
size_t
inpIdx
=
0
;
inpIdx
<
inputs
.
size
();
inpIdx
++
)
...
...
modules/dnn/src/layers/region_layer.cpp
浏览文件 @
9d02d42a
...
...
@@ -60,6 +60,9 @@ public:
int
coords
,
classes
,
anchors
,
classfix
;
float
thresh
,
nmsThreshold
;
bool
useSoftmax
,
useLogistic
;
#ifdef HAVE_OPENCL
UMat
blob_umat
;
#endif
RegionLayerImpl
(
const
LayerParams
&
params
)
{
...
...
@@ -123,6 +126,9 @@ public:
#ifdef HAVE_OPENCL
bool
forward_ocl
(
InputArrayOfArrays
inps
,
OutputArrayOfArrays
outs
,
OutputArrayOfArrays
internals
)
{
if
(
blob_umat
.
empty
())
blobs
[
0
].
copyTo
(
blob_umat
);
std
::
vector
<
UMat
>
inputs
;
std
::
vector
<
UMat
>
outputs
;
...
...
@@ -135,7 +141,6 @@ public:
CV_Assert
(
inputs
.
size
()
>=
1
);
int
const
cell_size
=
classes
+
coords
+
1
;
UMat
blob_umat
=
blobs
[
0
].
getUMat
(
ACCESS_READ
);
for
(
size_t
ii
=
0
;
ii
<
outputs
.
size
();
ii
++
)
{
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录