* Add functionality to filter homography decompositions
* documentation + small refactor
* fix comparing int to size_t (compiler warning)
* fix whitespace issues
* clarification of function return values in documentation
* refactor of function parameters and change in loop nesting
* cleanup useless .h, fix size_t to int compare, small refactor
* fix documentation and whitespace
* change output from return value to outputarray parameter
* update function docs to reflect changes in parameters
* whitespace
* failing test
* fixed warnings related to extended initialisers and improper types
* initialize vectors from arrays
* initialize vectors from arrays part 2
* fix whitespace
* fix trailing whitespace
* Include <inttypes.h> in test_filter_homography_decomp.cpp, should fix 'uint8_t' : undeclared identifier error
* updated the test (made it shorter and providing better diagnostic) and significantly improved implementation (get rid of heavy repeated and/or unnecessary operations)
* fixed compile warning; removed trailing whitespace
* Pass RANSAC parameters as function input
* Clean up unnecessary code
* Keep the original function signature
* Clean up based on PR comments
Replace array with vector.
Correct naming convention for input variables.
Add checks on input variables.
* Use vector instead of array for dynamic size
* Revert change.
* Use dynamic array
* Fix wrong syntax in array allocation
* Undo change
* Fix variable name
* Use vector and not array
* fixed compile warning on Windows
* temporarily disabled OpenCL use in DNN module on Mac, since some of the tests fail
* disable OpenCL in DNN on Mac at CMake level, not source level (thanks to alalek for the advice)
* optimize ocl kernel enqueue in fc layer
Signed-off-by: Li Peng <peng.li@intel.com>
* use CV_LOG_INFO in convolution auto tuning
Signed-off-by: Li Peng <peng.li@intel.com>
* update convolution IDLF kernel
extend parameter tuning range, also cleanup
ocl kernel implementation
Signed-off-by: Li Peng <peng.li@intel.com>
* update in-memory convolution cache config
fp16 and fp32 cache config are stored separately
Signed-off-by: Li Peng <peng.li@intel.com>
dnn: Fix output mismatch when forward dnn model contain [depthwise conv(group=1) + bn + prelu] (#11649)
* this can make sure [depthwise conv(group=1) + bn + prelu] output not shift
* add TEST to show the output mismatch in [DWconv+Prelu]
* fix typo
* change loading image to init cvMat directly
* build runtime model, without loading external model
* remove whitespace
* change way to create a cvmat
* add bias_term, add target output
* fix [dwconv + prelu] value mismatch when no optimizations
* fix Test error when change output channels
* add parametric test
* change num_output to group value
* change conv code and change test back
* Added ResizeBilinear op for tf
Combined ResizeNearestNeighbor and ResizeBilinear layers into Resize (with an interpolation param).
Minor changes to tf_importer and resize layer to save some code lines
Minor changes in init.cpp
Minor changes in tf_importer.cpp
* Replaced implementation of a custom ResizeBilinear layer to all layers
* Use Mat::ptr. Replace interpolation flags