mirror of
https://github.com/opencv/opencv.git
synced 2025-01-19 06:53:50 +08:00
237 lines
5.6 KiB
C++
237 lines
5.6 KiB
C++
// This file is part of OpenCV project.
|
|
// It is subject to the license terms in the LICENSE file found in the top-level directory
|
|
// of this distribution and at http://opencv.org/license.html.
|
|
//
|
|
// Copyright (C) 2018, Intel Corporation, all rights reserved.
|
|
// Third party copyrights are property of their respective owners.
|
|
|
|
#include "precomp.hpp"
|
|
#include <opencv2/dnn/shape_utils.hpp>
|
|
#include "op_vkcom.hpp"
|
|
#include "net_impl.hpp"
|
|
|
|
namespace cv
|
|
{
|
|
namespace dnn
|
|
{
|
|
#ifdef HAVE_VULKAN
|
|
|
|
CV__DNN_INLINE_NS_BEGIN
|
|
|
|
void Net::Impl::initVkComBackend()
|
|
{
|
|
CV_TRACE_FUNCTION();
|
|
CV_Assert(preferableBackend == DNN_BACKEND_VKCOM);
|
|
|
|
context = vkcom::Context::create();
|
|
|
|
for (MapIdToLayerData::iterator it = layers.begin(); it != layers.end(); it++)
|
|
{
|
|
LayerData &ld = it->second;
|
|
Ptr<Layer> layer = ld.layerInstance;
|
|
if (!layer->supportBackend(preferableBackend))
|
|
{
|
|
continue;
|
|
}
|
|
|
|
try
|
|
{
|
|
ld.backendNodes[DNN_BACKEND_VKCOM] = layer->initVkCom(ld.inputBlobsWrappers, ld.outputBlobsWrappers);
|
|
}
|
|
catch (const cv::Exception& e)
|
|
{
|
|
CV_LOG_ERROR(NULL, "initVkCom failed, fallback to CPU implementation. " << e.what());
|
|
ld.backendNodes[DNN_BACKEND_VKCOM] = Ptr<BackendNode>();
|
|
}
|
|
}
|
|
}
|
|
|
|
CV__DNN_INLINE_NS_END
|
|
|
|
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
int transFusedActivType(Ptr<ActivationLayer> &actLayer)
|
|
{
|
|
if (actLayer)
|
|
{
|
|
Ptr<ReLULayer> activ_relu = actLayer.dynamicCast<ReLULayer>();
|
|
Ptr<ReLU6Layer> activ_relu6 = actLayer.dynamicCast<ReLU6Layer>();
|
|
|
|
if (!activ_relu.empty())
|
|
{
|
|
if (activ_relu->negativeSlope == 0.0f)
|
|
{
|
|
return 1; // kFusedActivRelu
|
|
}
|
|
else // Leaky ReLU
|
|
{
|
|
return -1; // kFusedActivNone
|
|
}
|
|
}
|
|
else if (!activ_relu6.empty())
|
|
{
|
|
return 2; // kFusedActivRelu6
|
|
}
|
|
else
|
|
return -1; // kFusedActivUnsupport
|
|
}
|
|
else
|
|
return 0; // kFusedActivNone
|
|
}
|
|
|
|
void copyToTensor(vkcom::Tensor &dst, const Mat &src)
|
|
{
|
|
CV_Assert(src.isContinuous() && src.type() == CV_32F);
|
|
|
|
std::vector<int> mat_shape = shape(src);
|
|
|
|
// The following code will copy the src data from CPU Mat to GPU VkBuffer.
|
|
dst.reshape((const char*)src.data, mat_shape);
|
|
}
|
|
|
|
void copyToMat(Mat &dst, vkcom::Tensor &src)
|
|
{
|
|
CV_Assert(dst.type() == CV_32F);
|
|
|
|
std::vector<int> shape = src.getShape();
|
|
void *data = src.map();
|
|
Mat tmp(shape, CV_32F, data);
|
|
tmp.copyTo(dst);
|
|
src.unMap();
|
|
}
|
|
|
|
vkcom::Tensor VkComTensor(const Ptr<BackendWrapper>& ptr)
|
|
{
|
|
CV_Assert(!ptr.empty());
|
|
return ptr.dynamicCast<VkComBackendWrapper>()->getTensor();
|
|
}
|
|
|
|
void setDirty(std::vector<Ptr<BackendWrapper> >& ptrs)
|
|
{
|
|
for (const Ptr<BackendWrapper>& ptr : ptrs)
|
|
{
|
|
ptr.dynamicCast<VkComBackendWrapper>()->setDeviceDirty();
|
|
}
|
|
}
|
|
|
|
std::vector<vkcom::Tensor> VkComTensors(const std::vector<Ptr<BackendWrapper> >& ptrs)
|
|
{
|
|
std::vector<vkcom::Tensor> vec;
|
|
vec.reserve(ptrs.size());
|
|
for (const Ptr<BackendWrapper>& ptr : ptrs)
|
|
{
|
|
vec.push_back(VkComTensor(ptr));
|
|
}
|
|
return vec;
|
|
}
|
|
|
|
VkComBackendNode::VkComBackendNode(const std::vector<Ptr<BackendWrapper> >& inputsWrapper,
|
|
const Ptr<vkcom::OpBase>& op,
|
|
const std::vector<Ptr<BackendWrapper> >& outputsWrapper)
|
|
: BackendNode(DNN_BACKEND_VKCOM)
|
|
{
|
|
operation = op;
|
|
|
|
inputsWrapper_ = inputsWrapper;
|
|
ins = VkComTensors(inputsWrapper_);
|
|
|
|
outputsWrapper_ = outputsWrapper;
|
|
outs = VkComTensors(outputsWrapper_);
|
|
}
|
|
|
|
bool VkComBackendNode::forward()
|
|
{
|
|
for (int i = 0, n = inputsWrapper_.size(); i < n; ++i)
|
|
{
|
|
inputsWrapper_[i].dynamicCast<VkComBackendWrapper>()->copyToDevice();
|
|
}
|
|
|
|
return operation->forward(ins, outs);
|
|
}
|
|
|
|
VkComBackendWrapper::VkComBackendWrapper(Mat& m) : BackendWrapper(DNN_BACKEND_VKCOM, DNN_TARGET_VULKAN)
|
|
{
|
|
CV_Assert(m.isContinuous());
|
|
copyToTensor(tensor, m);
|
|
host = &m;
|
|
hostDirty = false;
|
|
deviceDirty = false;
|
|
}
|
|
|
|
// Other constructor, need change the logical. The purpose is to decline the data copy.
|
|
VkComBackendWrapper::VkComBackendWrapper(const Ptr<BackendWrapper>& baseBuffer, Mat& m)
|
|
: BackendWrapper(DNN_BACKEND_VKCOM, DNN_TARGET_VULKAN)
|
|
{
|
|
Ptr<VkComBackendWrapper> base = baseBuffer.dynamicCast<VkComBackendWrapper>();
|
|
CV_Assert(!base.empty());
|
|
|
|
host = &m;
|
|
tensor = base->tensor;
|
|
CV_Assert(tensor.count() >= m.total());
|
|
tensor.reshape(0, shape(m));
|
|
hostDirty = false;
|
|
deviceDirty = false;
|
|
}
|
|
|
|
void VkComBackendWrapper::copyToHost()
|
|
{
|
|
if (deviceDirty)
|
|
copyToMat(*host, tensor);
|
|
}
|
|
|
|
void VkComBackendWrapper::setHostDirty()
|
|
{
|
|
hostDirty = true;
|
|
};
|
|
|
|
void VkComBackendWrapper::setDeviceDirty()
|
|
{
|
|
deviceDirty = true;
|
|
};
|
|
|
|
void VkComBackendWrapper::copyToDevice()
|
|
{
|
|
if (hostDirty)
|
|
{
|
|
copyToTensor(tensor, *host);
|
|
hostDirty = false;
|
|
}
|
|
}
|
|
|
|
vkcom::Tensor VkComBackendWrapper::getTensor()
|
|
{
|
|
return tensor;
|
|
}
|
|
|
|
Mat* VkComBackendWrapper::getMat()
|
|
{
|
|
return host;
|
|
}
|
|
|
|
#endif
|
|
|
|
void forwardVkCom(std::vector<Ptr<BackendWrapper> > &outputs,
|
|
const Ptr<BackendNode>& node)
|
|
{
|
|
#ifdef HAVE_VULKAN
|
|
CV_Assert(!node.empty());
|
|
|
|
Ptr<VkComBackendNode> node_ = node.dynamicCast<VkComBackendNode>();
|
|
|
|
CV_Assert(node_->forward());
|
|
setDirty(outputs);
|
|
#endif
|
|
}
|
|
|
|
bool haveVulkan()
|
|
{
|
|
#ifdef HAVE_VULKAN
|
|
return vkcom::isAvailable();
|
|
#else
|
|
return false;
|
|
#endif // HAVE_VULKAN
|
|
}
|
|
|
|
} // namespace dnn
|
|
} // namespace cv
|