opencv/modules/imgproc/src/thresh.cpp
Pierre Chatelier 0db6a496ba
Merge pull request #26842 from chacha21:threshold_with_mask
Added optional mask to cv::threshold #26842
 
Proposal for #26777

To avoid code duplication, and keep performance when no mask is used, inner implementation always propagate the const cv::Mat& mask, but they use a template<bool useMask> parameter that let the compiler optimize out unnecessary tests when the mask is not to be used.

See details at https://github.com/opencv/opencv/wiki/How_to_contribute#making-a-good-pull-request

- [x] I agree to contribute to the project under Apache 2 License.
- [x] To the best of my knowledge, the proposed patch is not based on a code under GPL or another license that is incompatible with OpenCV
- [X] The PR is proposed to the proper branch
- [X] There is a reference to the original bug report and related work
- [X] There is accuracy test, performance test and test data in opencv_extra repository, if applicable
      Patch to opencv_extra has the same branch name.
- [ ] The feature is well documented and sample code can be built with the project CMake
2025-03-12 17:55:07 +03:00

1999 lines
66 KiB
C++

/*M///////////////////////////////////////////////////////////////////////////////////////
//
// IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING.
//
// By downloading, copying, installing or using the software you agree to this license.
// If you do not agree to this license, do not download, install,
// copy or use the software.
//
//
// License Agreement
// For Open Source Computer Vision Library
//
// Copyright (C) 2000-2008, Intel Corporation, all rights reserved.
// Copyright (C) 2009, Willow Garage Inc., all rights reserved.
// Third party copyrights are property of their respective owners.
//
// Redistribution and use in source and binary forms, with or without modification,
// are permitted provided that the following conditions are met:
//
// * Redistribution's of source code must retain the above copyright notice,
// this list of conditions and the following disclaimer.
//
// * Redistribution's in binary form must reproduce the above copyright notice,
// this list of conditions and the following disclaimer in the documentation
// and/or other materials provided with the distribution.
//
// * The name of the copyright holders may not be used to endorse or promote products
// derived from this software without specific prior written permission.
//
// This software is provided by the copyright holders and contributors "as is" and
// any express or implied warranties, including, but not limited to, the implied
// warranties of merchantability and fitness for a particular purpose are disclaimed.
// In no event shall the Intel Corporation or contributors be liable for any direct,
// indirect, incidental, special, exemplary, or consequential damages
// (including, but not limited to, procurement of substitute goods or services;
// loss of use, data, or profits; or business interruption) however caused
// and on any theory of liability, whether in contract, strict liability,
// or tort (including negligence or otherwise) arising in any way out of
// the use of this software, even if advised of the possibility of such damage.
//
//M*/
#include "precomp.hpp"
#include "opencl_kernels_imgproc.hpp"
#include "opencv2/core/hal/intrin.hpp"
namespace cv
{
template <typename T>
static inline T threshBinary(const T& src, const T& thresh, const T& maxval)
{
return src > thresh ? maxval : 0;
}
template <typename T>
static inline T threshBinaryInv(const T& src, const T& thresh, const T& maxval)
{
return src <= thresh ? maxval : 0;
}
template <typename T>
static inline T threshTrunc(const T& src, const T& thresh)
{
return std::min(src, thresh);
}
template <typename T>
static inline T threshToZero(const T& src, const T& thresh)
{
return src > thresh ? src : 0;
}
template <typename T>
static inline T threshToZeroInv(const T& src, const T& thresh)
{
return src <= thresh ? src : 0;
}
template <typename T>
static void threshGeneric(Size roi, const T* src, size_t src_step, T* dst,
size_t dst_step, T thresh, T maxval, int type)
{
int i = 0, j;
switch (type)
{
case THRESH_BINARY:
for (; i < roi.height; i++, src += src_step, dst += dst_step)
for (j = 0; j < roi.width; j++)
dst[j] = threshBinary<T>(src[j], thresh, maxval);
return;
case THRESH_BINARY_INV:
for (; i < roi.height; i++, src += src_step, dst += dst_step)
for (j = 0; j < roi.width; j++)
dst[j] = threshBinaryInv<T>(src[j], thresh, maxval);
return;
case THRESH_TRUNC:
for (; i < roi.height; i++, src += src_step, dst += dst_step)
for (j = 0; j < roi.width; j++)
dst[j] = threshTrunc<T>(src[j], thresh);
return;
case THRESH_TOZERO:
for (; i < roi.height; i++, src += src_step, dst += dst_step)
for (j = 0; j < roi.width; j++)
dst[j] = threshToZero<T>(src[j], thresh);
return;
case THRESH_TOZERO_INV:
for (; i < roi.height; i++, src += src_step, dst += dst_step)
for (j = 0; j < roi.width; j++)
dst[j] = threshToZeroInv<T>(src[j], thresh);
return;
default:
CV_Error( cv::Error::StsBadArg, "" ); return;
}
}
template <typename T>
static void threshGenericWithMask(const Mat& _src, Mat& _dst, const Mat& _mask,
T thresh, T maxval, int type)
{
Size roi = _src.size();
const int cn = _src.channels();
roi.width *= cn;
size_t src_step = _src.step/_src.elemSize1();
size_t dst_step = _dst.step/_src.elemSize1();
const T* src = _src.ptr<T>(0);
T* dst = _dst.ptr<T>(0);
const unsigned char* mask = _mask.ptr<unsigned char>(0);
size_t mask_step = _mask.step;
int i = 0, j;
switch (type)
{
case THRESH_BINARY:
for (; i < roi.height; i++, src += src_step, dst += dst_step, mask += mask_step)
for (j = 0; j < roi.width; j++)
if (mask[j/cn] != 0)
dst[j] = threshBinary<T>(src[j], thresh, maxval);
return;
case THRESH_BINARY_INV:
for (; i < roi.height; i++, src += src_step, dst += dst_step, mask += mask_step)
for (j = 0; j < roi.width; j++)
if (mask[j/cn] != 0)
dst[j] = threshBinaryInv<T>(src[j], thresh, maxval);
return;
case THRESH_TRUNC:
for (; i < roi.height; i++, src += src_step, dst += dst_step, mask += mask_step)
for (j = 0; j < roi.width; j++)
if (mask[j/cn] != 0)
dst[j] = threshTrunc<T>(src[j], thresh);
return;
case THRESH_TOZERO:
for (; i < roi.height; i++, src += src_step, dst += dst_step, mask += mask_step)
for (j = 0; j < roi.width; j++)
if (mask[j/cn] != 0)
dst[j] = threshToZero<T>(src[j], thresh);
return;
case THRESH_TOZERO_INV:
for (; i < roi.height; i++, src += src_step, dst += dst_step, mask += mask_step)
for (j = 0; j < roi.width; j++)
if (mask[j/cn] != 0)
dst[j] = threshToZeroInv<T>(src[j], thresh);
return;
default:
CV_Error( cv::Error::StsBadArg, "" ); return;
}
}
static void
thresh_8u( const Mat& _src, Mat& _dst, uchar thresh, uchar maxval, int type )
{
Size roi = _src.size();
roi.width *= _src.channels();
size_t src_step = _src.step;
size_t dst_step = _dst.step;
if( _src.isContinuous() && _dst.isContinuous() )
{
roi.width *= roi.height;
roi.height = 1;
src_step = dst_step = roi.width;
}
#if defined(HAVE_IPP)
CV_IPP_CHECK()
{
IppiSize sz = { roi.width, roi.height };
CV_SUPPRESS_DEPRECATED_START
switch( type )
{
case THRESH_TRUNC:
if (_src.data == _dst.data && CV_INSTRUMENT_FUN_IPP(ippiThreshold_GT_8u_C1IR, _dst.ptr(), (int)dst_step, sz, thresh) >= 0)
{
CV_IMPL_ADD(CV_IMPL_IPP);
return;
}
if (CV_INSTRUMENT_FUN_IPP(ippiThreshold_GT_8u_C1R, _src.ptr(), (int)src_step, _dst.ptr(), (int)dst_step, sz, thresh) >= 0)
{
CV_IMPL_ADD(CV_IMPL_IPP);
return;
}
setIppErrorStatus();
break;
case THRESH_TOZERO:
if (_src.data == _dst.data && CV_INSTRUMENT_FUN_IPP(ippiThreshold_LTVal_8u_C1IR, _dst.ptr(), (int)dst_step, sz, thresh+1, 0) >= 0)
{
CV_IMPL_ADD(CV_IMPL_IPP);
return;
}
if (CV_INSTRUMENT_FUN_IPP(ippiThreshold_LTVal_8u_C1R, _src.ptr(), (int)src_step, _dst.ptr(), (int)dst_step, sz, thresh + 1, 0) >= 0)
{
CV_IMPL_ADD(CV_IMPL_IPP);
return;
}
setIppErrorStatus();
break;
case THRESH_TOZERO_INV:
if (_src.data == _dst.data && CV_INSTRUMENT_FUN_IPP(ippiThreshold_GTVal_8u_C1IR, _dst.ptr(), (int)dst_step, sz, thresh, 0) >= 0)
{
CV_IMPL_ADD(CV_IMPL_IPP);
return;
}
if (CV_INSTRUMENT_FUN_IPP(ippiThreshold_GTVal_8u_C1R, _src.ptr(), (int)src_step, _dst.ptr(), (int)dst_step, sz, thresh, 0) >= 0)
{
CV_IMPL_ADD(CV_IMPL_IPP);
return;
}
setIppErrorStatus();
break;
}
CV_SUPPRESS_DEPRECATED_END
}
#endif
int j = 0;
const uchar* src = _src.ptr();
uchar* dst = _dst.ptr();
#if (CV_SIMD || CV_SIMD_SCALABLE)
v_uint8 thresh_u = vx_setall_u8( thresh );
v_uint8 maxval16 = vx_setall_u8( maxval );
switch( type )
{
case THRESH_BINARY:
for( int i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
for( j = 0; j <= roi.width - VTraits<v_uint8>::vlanes(); j += VTraits<v_uint8>::vlanes())
{
v_uint8 v0;
v0 = vx_load( src + j );
v0 = v_lt(thresh_u, v0);
v0 = v_and(v0, maxval16);
v_store( dst + j, v0 );
}
}
break;
case THRESH_BINARY_INV:
for( int i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
for( j = 0; j <= roi.width - VTraits<v_uint8>::vlanes(); j += VTraits<v_uint8>::vlanes())
{
v_uint8 v0;
v0 = vx_load( src + j );
v0 = v_le(v0, thresh_u);
v0 = v_and(v0, maxval16);
v_store( dst + j, v0 );
}
}
break;
case THRESH_TRUNC:
for( int i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
for( j = 0; j <= roi.width - VTraits<v_uint8>::vlanes(); j += VTraits<v_uint8>::vlanes())
{
v_uint8 v0;
v0 = vx_load( src + j );
v0 = v_sub(v0, v_sub(v0, thresh_u));
v_store( dst + j, v0 );
}
}
break;
case THRESH_TOZERO:
for( int i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
for( j = 0; j <= roi.width - VTraits<v_uint8>::vlanes(); j += VTraits<v_uint8>::vlanes())
{
v_uint8 v0;
v0 = vx_load( src + j );
v0 = v_and(v_lt(thresh_u, v0), v0);
v_store( dst + j, v0 );
}
}
break;
case THRESH_TOZERO_INV:
for( int i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
for( j = 0; j <= roi.width - VTraits<v_uint8>::vlanes(); j += VTraits<v_uint8>::vlanes())
{
v_uint8 v0;
v0 = vx_load( src + j );
v0 = v_and(v_le(v0, thresh_u), v0);
v_store( dst + j, v0 );
}
}
break;
}
#endif
int j_scalar = j;
if( j_scalar < roi.width )
{
const int thresh_pivot = thresh + 1;
uchar tab[256] = {0};
switch( type )
{
case THRESH_BINARY:
memset(tab, 0, thresh_pivot);
if (thresh_pivot < 256) {
memset(tab + thresh_pivot, maxval, 256 - thresh_pivot);
}
break;
case THRESH_BINARY_INV:
memset(tab, maxval, thresh_pivot);
if (thresh_pivot < 256) {
memset(tab + thresh_pivot, 0, 256 - thresh_pivot);
}
break;
case THRESH_TRUNC:
for( int i = 0; i <= thresh; i++ )
tab[i] = (uchar)i;
if (thresh_pivot < 256) {
memset(tab + thresh_pivot, thresh, 256 - thresh_pivot);
}
break;
case THRESH_TOZERO:
memset(tab, 0, thresh_pivot);
for( int i = thresh_pivot; i < 256; i++ )
tab[i] = (uchar)i;
break;
case THRESH_TOZERO_INV:
for( int i = 0; i <= thresh; i++ )
tab[i] = (uchar)i;
if (thresh_pivot < 256) {
memset(tab + thresh_pivot, 0, 256 - thresh_pivot);
}
break;
}
src = _src.ptr();
dst = _dst.ptr();
for( int i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
j = j_scalar;
#if CV_ENABLE_UNROLLED
for( ; j <= roi.width - 4; j += 4 )
{
uchar t0 = tab[src[j]];
uchar t1 = tab[src[j+1]];
dst[j] = t0;
dst[j+1] = t1;
t0 = tab[src[j+2]];
t1 = tab[src[j+3]];
dst[j+2] = t0;
dst[j+3] = t1;
}
#endif
for( ; j < roi.width; j++ )
dst[j] = tab[src[j]];
}
}
}
static void
thresh_16u(const Mat& _src, Mat& _dst, ushort thresh, ushort maxval, int type)
{
Size roi = _src.size();
roi.width *= _src.channels();
size_t src_step = _src.step / _src.elemSize1();
size_t dst_step = _dst.step / _dst.elemSize1();
if (_src.isContinuous() && _dst.isContinuous())
{
roi.width *= roi.height;
roi.height = 1;
src_step = dst_step = roi.width;
}
// HAVE_IPP not supported
const ushort* src = _src.ptr<ushort>();
ushort* dst = _dst.ptr<ushort>();
#if (CV_SIMD || CV_SIMD_SCALABLE)
int i, j;
v_uint16 thresh_u = vx_setall_u16(thresh);
v_uint16 maxval16 = vx_setall_u16(maxval);
switch (type)
{
case THRESH_BINARY:
for (i = 0; i < roi.height; i++, src += src_step, dst += dst_step)
{
for (j = 0; j <= roi.width - 2*VTraits<v_uint16>::vlanes(); j += 2*VTraits<v_uint16>::vlanes())
{
v_uint16 v0, v1;
v0 = vx_load(src + j);
v1 = vx_load(src + j + VTraits<v_uint16>::vlanes());
v0 = v_lt(thresh_u, v0);
v1 = v_lt(thresh_u, v1);
v0 = v_and(v0, maxval16);
v1 = v_and(v1, maxval16);
v_store(dst + j, v0);
v_store(dst + j + VTraits<v_uint16>::vlanes(), v1);
}
if (j <= roi.width - VTraits<v_uint16>::vlanes())
{
v_uint16 v0 = vx_load(src + j);
v0 = v_lt(thresh_u, v0);
v0 = v_and(v0, maxval16);
v_store(dst + j, v0);
j += VTraits<v_uint16>::vlanes();
}
for (; j < roi.width; j++)
dst[j] = threshBinary<ushort>(src[j], thresh, maxval);
}
break;
case THRESH_BINARY_INV:
for (i = 0; i < roi.height; i++, src += src_step, dst += dst_step)
{
j = 0;
for (; j <= roi.width - 2*VTraits<v_uint16>::vlanes(); j += 2*VTraits<v_uint16>::vlanes())
{
v_uint16 v0, v1;
v0 = vx_load(src + j);
v1 = vx_load(src + j + VTraits<v_uint16>::vlanes());
v0 = v_le(v0, thresh_u);
v1 = v_le(v1, thresh_u);
v0 = v_and(v0, maxval16);
v1 = v_and(v1, maxval16);
v_store(dst + j, v0);
v_store(dst + j + VTraits<v_uint16>::vlanes(), v1);
}
if (j <= roi.width - VTraits<v_uint16>::vlanes())
{
v_uint16 v0 = vx_load(src + j);
v0 = v_le(v0, thresh_u);
v0 = v_and(v0, maxval16);
v_store(dst + j, v0);
j += VTraits<v_uint16>::vlanes();
}
for (; j < roi.width; j++)
dst[j] = threshBinaryInv<ushort>(src[j], thresh, maxval);
}
break;
case THRESH_TRUNC:
for (i = 0; i < roi.height; i++, src += src_step, dst += dst_step)
{
j = 0;
for (; j <= roi.width - 2*VTraits<v_uint16>::vlanes(); j += 2*VTraits<v_uint16>::vlanes())
{
v_uint16 v0, v1;
v0 = vx_load(src + j);
v1 = vx_load(src + j + VTraits<v_uint16>::vlanes());
v0 = v_min(v0, thresh_u);
v1 = v_min(v1, thresh_u);
v_store(dst + j, v0);
v_store(dst + j + VTraits<v_uint16>::vlanes(), v1);
}
if (j <= roi.width - VTraits<v_uint16>::vlanes())
{
v_uint16 v0 = vx_load(src + j);
v0 = v_min(v0, thresh_u);
v_store(dst + j, v0);
j += VTraits<v_uint16>::vlanes();
}
for (; j < roi.width; j++)
dst[j] = threshTrunc<ushort>(src[j], thresh);
}
break;
case THRESH_TOZERO:
for (i = 0; i < roi.height; i++, src += src_step, dst += dst_step)
{
j = 0;
for (; j <= roi.width - 2*VTraits<v_uint16>::vlanes(); j += 2*VTraits<v_uint16>::vlanes())
{
v_uint16 v0, v1;
v0 = vx_load(src + j);
v1 = vx_load(src + j + VTraits<v_uint16>::vlanes());
v0 = v_and(v_lt(thresh_u, v0), v0);
v1 = v_and(v_lt(thresh_u, v1), v1);
v_store(dst + j, v0);
v_store(dst + j + VTraits<v_uint16>::vlanes(), v1);
}
if (j <= roi.width - VTraits<v_uint16>::vlanes())
{
v_uint16 v0 = vx_load(src + j);
v0 = v_and(v_lt(thresh_u, v0), v0);
v_store(dst + j, v0);
j += VTraits<v_uint16>::vlanes();
}
for (; j < roi.width; j++)
dst[j] = threshToZero<ushort>(src[j], thresh);
}
break;
case THRESH_TOZERO_INV:
for (i = 0; i < roi.height; i++, src += src_step, dst += dst_step)
{
j = 0;
for (; j <= roi.width - 2*VTraits<v_uint16>::vlanes(); j += 2*VTraits<v_uint16>::vlanes())
{
v_uint16 v0, v1;
v0 = vx_load(src + j);
v1 = vx_load(src + j + VTraits<v_uint16>::vlanes());
v0 = v_and(v_le(v0, thresh_u), v0);
v1 = v_and(v_le(v1, thresh_u), v1);
v_store(dst + j, v0);
v_store(dst + j + VTraits<v_uint16>::vlanes(), v1);
}
if (j <= roi.width - VTraits<v_uint16>::vlanes())
{
v_uint16 v0 = vx_load(src + j);
v0 = v_and(v_le(v0, thresh_u), v0);
v_store(dst + j, v0);
j += VTraits<v_uint16>::vlanes();
}
for (; j < roi.width; j++)
dst[j] = threshToZeroInv<ushort>(src[j], thresh);
}
break;
}
#else
threshGeneric<ushort>(roi, src, src_step, dst, dst_step, thresh, maxval, type);
#endif
}
static void
thresh_16s( const Mat& _src, Mat& _dst, short thresh, short maxval, int type )
{
Size roi = _src.size();
roi.width *= _src.channels();
const short* src = _src.ptr<short>();
short* dst = _dst.ptr<short>();
size_t src_step = _src.step/sizeof(src[0]);
size_t dst_step = _dst.step/sizeof(dst[0]);
if( _src.isContinuous() && _dst.isContinuous() )
{
roi.width *= roi.height;
roi.height = 1;
src_step = dst_step = roi.width;
}
#if defined(HAVE_IPP)
CV_IPP_CHECK()
{
IppiSize sz = { roi.width, roi.height };
CV_SUPPRESS_DEPRECATED_START
switch( type )
{
case THRESH_TRUNC:
if (_src.data == _dst.data && CV_INSTRUMENT_FUN_IPP(ippiThreshold_GT_16s_C1IR, dst, (int)dst_step*sizeof(dst[0]), sz, thresh) >= 0)
{
CV_IMPL_ADD(CV_IMPL_IPP);
return;
}
if (CV_INSTRUMENT_FUN_IPP(ippiThreshold_GT_16s_C1R, src, (int)src_step*sizeof(src[0]), dst, (int)dst_step*sizeof(dst[0]), sz, thresh) >= 0)
{
CV_IMPL_ADD(CV_IMPL_IPP);
return;
}
setIppErrorStatus();
break;
case THRESH_TOZERO:
if (_src.data == _dst.data && CV_INSTRUMENT_FUN_IPP(ippiThreshold_LTVal_16s_C1IR, dst, (int)dst_step*sizeof(dst[0]), sz, thresh + 1, 0) >= 0)
{
CV_IMPL_ADD(CV_IMPL_IPP);
return;
}
if (CV_INSTRUMENT_FUN_IPP(ippiThreshold_LTVal_16s_C1R, src, (int)src_step*sizeof(src[0]), dst, (int)dst_step*sizeof(dst[0]), sz, thresh + 1, 0) >= 0)
{
CV_IMPL_ADD(CV_IMPL_IPP);
return;
}
setIppErrorStatus();
break;
case THRESH_TOZERO_INV:
if (_src.data == _dst.data && CV_INSTRUMENT_FUN_IPP(ippiThreshold_GTVal_16s_C1IR, dst, (int)dst_step*sizeof(dst[0]), sz, thresh, 0) >= 0)
{
CV_IMPL_ADD(CV_IMPL_IPP);
return;
}
if (CV_INSTRUMENT_FUN_IPP(ippiThreshold_GTVal_16s_C1R, src, (int)src_step*sizeof(src[0]), dst, (int)dst_step*sizeof(dst[0]), sz, thresh, 0) >= 0)
{
CV_IMPL_ADD(CV_IMPL_IPP);
return;
}
setIppErrorStatus();
break;
}
CV_SUPPRESS_DEPRECATED_END
}
#endif
#if (CV_SIMD || CV_SIMD_SCALABLE)
int i, j;
v_int16 thresh8 = vx_setall_s16( thresh );
v_int16 maxval8 = vx_setall_s16( maxval );
switch( type )
{
case THRESH_BINARY:
for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
j = 0;
for( ; j <= roi.width - 2*VTraits<v_int16>::vlanes(); j += 2*VTraits<v_int16>::vlanes() )
{
v_int16 v0, v1;
v0 = vx_load( src + j );
v1 = vx_load( src + j + VTraits<v_int16>::vlanes() );
v0 = v_lt(thresh8, v0);
v1 = v_lt(thresh8, v1);
v0 = v_and(v0, maxval8);
v1 = v_and(v1, maxval8);
v_store( dst + j, v0 );
v_store( dst + j + VTraits<v_int16>::vlanes(), v1 );
}
if( j <= roi.width - VTraits<v_int16>::vlanes() )
{
v_int16 v0 = vx_load( src + j );
v0 = v_lt(thresh8, v0);
v0 = v_and(v0, maxval8);
v_store( dst + j, v0 );
j += VTraits<v_int16>::vlanes();
}
for( ; j < roi.width; j++ )
dst[j] = threshBinary<short>(src[j], thresh, maxval);
}
break;
case THRESH_BINARY_INV:
for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
j = 0;
for( ; j <= roi.width - 2*VTraits<v_int16>::vlanes(); j += 2*VTraits<v_int16>::vlanes() )
{
v_int16 v0, v1;
v0 = vx_load( src + j );
v1 = vx_load( src + j + VTraits<v_int16>::vlanes() );
v0 = v_le(v0, thresh8);
v1 = v_le(v1, thresh8);
v0 = v_and(v0, maxval8);
v1 = v_and(v1, maxval8);
v_store( dst + j, v0 );
v_store( dst + j + VTraits<v_int16>::vlanes(), v1 );
}
if( j <= roi.width - VTraits<v_int16>::vlanes() )
{
v_int16 v0 = vx_load( src + j );
v0 = v_le(v0, thresh8);
v0 = v_and(v0, maxval8);
v_store( dst + j, v0 );
j += VTraits<v_int16>::vlanes();
}
for( ; j < roi.width; j++ )
dst[j] = threshBinaryInv<short>(src[j], thresh, maxval);
}
break;
case THRESH_TRUNC:
for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
j = 0;
for( ; j <= roi.width - 2*VTraits<v_int16>::vlanes(); j += 2*VTraits<v_int16>::vlanes() )
{
v_int16 v0, v1;
v0 = vx_load( src + j );
v1 = vx_load( src + j + VTraits<v_int16>::vlanes() );
v0 = v_min( v0, thresh8 );
v1 = v_min( v1, thresh8 );
v_store( dst + j, v0 );
v_store( dst + j + VTraits<v_int16>::vlanes(), v1 );
}
if( j <= roi.width - VTraits<v_int16>::vlanes() )
{
v_int16 v0 = vx_load( src + j );
v0 = v_min( v0, thresh8 );
v_store( dst + j, v0 );
j += VTraits<v_int16>::vlanes();
}
for( ; j < roi.width; j++ )
dst[j] = threshTrunc<short>( src[j], thresh );
}
break;
case THRESH_TOZERO:
for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
j = 0;
for( ; j <= roi.width - 2*VTraits<v_int16>::vlanes(); j += 2*VTraits<v_int16>::vlanes() )
{
v_int16 v0, v1;
v0 = vx_load( src + j );
v1 = vx_load( src + j + VTraits<v_int16>::vlanes() );
v0 = v_and(v_lt(thresh8, v0), v0);
v1 = v_and(v_lt(thresh8, v1), v1);
v_store( dst + j, v0 );
v_store( dst + j + VTraits<v_int16>::vlanes(), v1 );
}
if( j <= roi.width - VTraits<v_int16>::vlanes() )
{
v_int16 v0 = vx_load( src + j );
v0 = v_and(v_lt(thresh8, v0), v0);
v_store( dst + j, v0 );
j += VTraits<v_int16>::vlanes();
}
for( ; j < roi.width; j++ )
dst[j] = threshToZero<short>(src[j], thresh);
}
break;
case THRESH_TOZERO_INV:
for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
j = 0;
for( ; j <= roi.width - 2*VTraits<v_int16>::vlanes(); j += 2*VTraits<v_int16>::vlanes() )
{
v_int16 v0, v1;
v0 = vx_load( src + j );
v1 = vx_load( src + j + VTraits<v_int16>::vlanes() );
v0 = v_and(v_le(v0, thresh8), v0);
v1 = v_and(v_le(v1, thresh8), v1);
v_store( dst + j, v0 );
v_store( dst + j + VTraits<v_int16>::vlanes(), v1 );
}
if( j <= roi.width - VTraits<v_int16>::vlanes() )
{
v_int16 v0 = vx_load( src + j );
v0 = v_and(v_le(v0, thresh8), v0);
v_store( dst + j, v0 );
j += VTraits<v_int16>::vlanes();
}
for( ; j < roi.width; j++ )
dst[j] = threshToZeroInv<short>(src[j], thresh);
}
break;
default:
CV_Error( cv::Error::StsBadArg, "" ); return;
}
#else
threshGeneric<short>(roi, src, src_step, dst, dst_step, thresh, maxval, type);
#endif
}
static void
thresh_32f( const Mat& _src, Mat& _dst, float thresh, float maxval, int type )
{
Size roi = _src.size();
roi.width *= _src.channels();
const float* src = _src.ptr<float>();
float* dst = _dst.ptr<float>();
size_t src_step = _src.step/sizeof(src[0]);
size_t dst_step = _dst.step/sizeof(dst[0]);
if( _src.isContinuous() && _dst.isContinuous() )
{
roi.width *= roi.height;
roi.height = 1;
}
#if defined(HAVE_IPP)
CV_IPP_CHECK()
{
IppiSize sz = { roi.width, roi.height };
switch( type )
{
case THRESH_TRUNC:
if (0 <= CV_INSTRUMENT_FUN_IPP(ippiThreshold_GT_32f_C1R, src, (int)src_step*sizeof(src[0]), dst, (int)dst_step*sizeof(dst[0]), sz, thresh))
{
CV_IMPL_ADD(CV_IMPL_IPP);
return;
}
setIppErrorStatus();
break;
case THRESH_TOZERO:
if (0 <= CV_INSTRUMENT_FUN_IPP(ippiThreshold_LTVal_32f_C1R, src, (int)src_step*sizeof(src[0]), dst, (int)dst_step*sizeof(dst[0]), sz, nextafterf(thresh, std::numeric_limits<float>::infinity()), 0))
{
CV_IMPL_ADD(CV_IMPL_IPP);
return;
}
setIppErrorStatus();
break;
case THRESH_TOZERO_INV:
if (0 <= CV_INSTRUMENT_FUN_IPP(ippiThreshold_GTVal_32f_C1R, src, (int)src_step*sizeof(src[0]), dst, (int)dst_step*sizeof(dst[0]), sz, thresh, 0))
{
CV_IMPL_ADD(CV_IMPL_IPP);
return;
}
setIppErrorStatus();
break;
}
}
#endif
#if (CV_SIMD || CV_SIMD_SCALABLE)
int i, j;
v_float32 thresh4 = vx_setall_f32( thresh );
v_float32 maxval4 = vx_setall_f32( maxval );
switch( type )
{
case THRESH_BINARY:
for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
j = 0;
for( ; j <= roi.width - 2*VTraits<v_float32>::vlanes(); j += 2*VTraits<v_float32>::vlanes() )
{
v_float32 v0, v1;
v0 = vx_load( src + j );
v1 = vx_load( src + j + VTraits<v_float32>::vlanes() );
v0 = v_lt(thresh4, v0);
v1 = v_lt(thresh4, v1);
v0 = v_and(v0, maxval4);
v1 = v_and(v1, maxval4);
v_store( dst + j, v0 );
v_store( dst + j + VTraits<v_float32>::vlanes(), v1 );
}
if( j <= roi.width - VTraits<v_float32>::vlanes() )
{
v_float32 v0 = vx_load( src + j );
v0 = v_lt(thresh4, v0);
v0 = v_and(v0, maxval4);
v_store( dst + j, v0 );
j += VTraits<v_float32>::vlanes();
}
for( ; j < roi.width; j++ )
dst[j] = threshBinary<float>(src[j], thresh, maxval);
}
break;
case THRESH_BINARY_INV:
for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
j = 0;
for( ; j <= roi.width - 2*VTraits<v_float32>::vlanes(); j += 2*VTraits<v_float32>::vlanes() )
{
v_float32 v0, v1;
v0 = vx_load( src + j );
v1 = vx_load( src + j + VTraits<v_float32>::vlanes() );
v0 = v_le(v0, thresh4);
v1 = v_le(v1, thresh4);
v0 = v_and(v0, maxval4);
v1 = v_and(v1, maxval4);
v_store( dst + j, v0 );
v_store( dst + j + VTraits<v_float32>::vlanes(), v1 );
}
if( j <= roi.width - VTraits<v_float32>::vlanes() )
{
v_float32 v0 = vx_load( src + j );
v0 = v_le(v0, thresh4);
v0 = v_and(v0, maxval4);
v_store( dst + j, v0 );
j += VTraits<v_float32>::vlanes();
}
for( ; j < roi.width; j++ )
dst[j] = threshBinaryInv<float>(src[j], thresh, maxval);
}
break;
case THRESH_TRUNC:
for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
j = 0;
for( ; j <= roi.width - 2*VTraits<v_float32>::vlanes(); j += 2*VTraits<v_float32>::vlanes() )
{
v_float32 v0, v1;
v0 = vx_load( src + j );
v1 = vx_load( src + j + VTraits<v_float32>::vlanes() );
v0 = v_min( v0, thresh4 );
v1 = v_min( v1, thresh4 );
v_store( dst + j, v0 );
v_store( dst + j + VTraits<v_float32>::vlanes(), v1 );
}
if( j <= roi.width - VTraits<v_float32>::vlanes() )
{
v_float32 v0 = vx_load( src + j );
v0 = v_min( v0, thresh4 );
v_store( dst + j, v0 );
j += VTraits<v_float32>::vlanes();
}
for( ; j < roi.width; j++ )
dst[j] = threshTrunc<float>(src[j], thresh);
}
break;
case THRESH_TOZERO:
for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
j = 0;
for( ; j <= roi.width - 2*VTraits<v_float32>::vlanes(); j += 2*VTraits<v_float32>::vlanes() )
{
v_float32 v0, v1;
v0 = vx_load( src + j );
v1 = vx_load( src + j + VTraits<v_float32>::vlanes() );
v0 = v_and(v_lt(thresh4, v0), v0);
v1 = v_and(v_lt(thresh4, v1), v1);
v_store( dst + j, v0 );
v_store( dst + j + VTraits<v_float32>::vlanes(), v1 );
}
if( j <= roi.width - VTraits<v_float32>::vlanes() )
{
v_float32 v0 = vx_load( src + j );
v0 = v_and(v_lt(thresh4, v0), v0);
v_store( dst + j, v0 );
j += VTraits<v_float32>::vlanes();
}
for( ; j < roi.width; j++ )
dst[j] = threshToZero<float>(src[j], thresh);
}
break;
case THRESH_TOZERO_INV:
for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
j = 0;
for( ; j <= roi.width - 2*VTraits<v_float32>::vlanes(); j += 2*VTraits<v_float32>::vlanes() )
{
v_float32 v0, v1;
v0 = vx_load( src + j );
v1 = vx_load( src + j + VTraits<v_float32>::vlanes() );
v0 = v_and(v_le(v0, thresh4), v0);
v1 = v_and(v_le(v1, thresh4), v1);
v_store( dst + j, v0 );
v_store( dst + j + VTraits<v_float32>::vlanes(), v1 );
}
if( j <= roi.width - VTraits<v_float32>::vlanes() )
{
v_float32 v0 = vx_load( src + j );
v0 = v_and(v_le(v0, thresh4), v0);
v_store( dst + j, v0 );
j += VTraits<v_float32>::vlanes();
}
for( ; j < roi.width; j++ )
dst[j] = threshToZeroInv<float>(src[j], thresh);
}
break;
default:
CV_Error( cv::Error::StsBadArg, "" ); return;
}
#else
threshGeneric<float>(roi, src, src_step, dst, dst_step, thresh, maxval, type);
#endif
}
static void
thresh_64f(const Mat& _src, Mat& _dst, double thresh, double maxval, int type)
{
Size roi = _src.size();
roi.width *= _src.channels();
const double* src = _src.ptr<double>();
double* dst = _dst.ptr<double>();
size_t src_step = _src.step / sizeof(src[0]);
size_t dst_step = _dst.step / sizeof(dst[0]);
if (_src.isContinuous() && _dst.isContinuous())
{
roi.width *= roi.height;
roi.height = 1;
}
#if (CV_SIMD_64F || CV_SIMD_SCALABLE_64F)
int i, j;
v_float64 thresh2 = vx_setall_f64( thresh );
v_float64 maxval2 = vx_setall_f64( maxval );
switch( type )
{
case THRESH_BINARY:
for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
j = 0;
for( ; j <= roi.width - 2*VTraits<v_float64>::vlanes(); j += 2*VTraits<v_float64>::vlanes() )
{
v_float64 v0, v1;
v0 = vx_load( src + j );
v1 = vx_load( src + j + VTraits<v_float64>::vlanes() );
v0 = v_lt(thresh2, v0);
v1 = v_lt(thresh2, v1);
v0 = v_and(v0, maxval2);
v1 = v_and(v1, maxval2);
v_store( dst + j, v0 );
v_store( dst + j + VTraits<v_float64>::vlanes(), v1 );
}
if( j <= roi.width - VTraits<v_float64>::vlanes() )
{
v_float64 v0 = vx_load( src + j );
v0 = v_lt(thresh2, v0);
v0 = v_and(v0, maxval2);
v_store( dst + j, v0 );
j += VTraits<v_float64>::vlanes();
}
for( ; j < roi.width; j++ )
dst[j] = threshBinary<double>(src[j], thresh, maxval);
}
break;
case THRESH_BINARY_INV:
for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
j = 0;
for( ; j <= roi.width - 2*VTraits<v_float64>::vlanes(); j += 2*VTraits<v_float64>::vlanes() )
{
v_float64 v0, v1;
v0 = vx_load( src + j );
v1 = vx_load( src + j + VTraits<v_float64>::vlanes() );
v0 = v_le(v0, thresh2);
v1 = v_le(v1, thresh2);
v0 = v_and(v0, maxval2);
v1 = v_and(v1, maxval2);
v_store( dst + j, v0 );
v_store( dst + j + VTraits<v_float64>::vlanes(), v1 );
}
if( j <= roi.width - VTraits<v_float64>::vlanes() )
{
v_float64 v0 = vx_load( src + j );
v0 = v_le(v0, thresh2);
v0 = v_and(v0, maxval2);
v_store( dst + j, v0 );
j += VTraits<v_float64>::vlanes();
}
for( ; j < roi.width; j++ )
dst[j] = threshBinaryInv<double>(src[j], thresh, maxval);
}
break;
case THRESH_TRUNC:
for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
j = 0;
for( ; j <= roi.width - 2*VTraits<v_float64>::vlanes(); j += 2*VTraits<v_float64>::vlanes() )
{
v_float64 v0, v1;
v0 = vx_load( src + j );
v1 = vx_load( src + j + VTraits<v_float64>::vlanes() );
v0 = v_min( v0, thresh2 );
v1 = v_min( v1, thresh2 );
v_store( dst + j, v0 );
v_store( dst + j + VTraits<v_float64>::vlanes(), v1 );
}
if( j <= roi.width - VTraits<v_float64>::vlanes() )
{
v_float64 v0 = vx_load( src + j );
v0 = v_min( v0, thresh2 );
v_store( dst + j, v0 );
j += VTraits<v_float64>::vlanes();
}
for( ; j < roi.width; j++ )
dst[j] = threshTrunc<double>(src[j], thresh);
}
break;
case THRESH_TOZERO:
for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
j = 0;
for( ; j <= roi.width - 2*VTraits<v_float64>::vlanes(); j += 2*VTraits<v_float64>::vlanes() )
{
v_float64 v0, v1;
v0 = vx_load( src + j );
v1 = vx_load( src + j + VTraits<v_float64>::vlanes() );
v0 = v_and(v_lt(thresh2, v0), v0);
v1 = v_and(v_lt(thresh2, v1), v1);
v_store( dst + j, v0 );
v_store( dst + j + VTraits<v_float64>::vlanes(), v1 );
}
if( j <= roi.width - VTraits<v_float64>::vlanes() )
{
v_float64 v0 = vx_load( src + j );
v0 = v_and(v_lt(thresh2, v0), v0);
v_store( dst + j, v0 );
j += VTraits<v_float64>::vlanes();
}
for( ; j < roi.width; j++ )
dst[j] = threshToZero<double>(src[j], thresh);
}
break;
case THRESH_TOZERO_INV:
for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
{
j = 0;
for( ; j <= roi.width - 2*VTraits<v_float64>::vlanes(); j += 2*VTraits<v_float64>::vlanes() )
{
v_float64 v0, v1;
v0 = vx_load( src + j );
v1 = vx_load( src + j + VTraits<v_float64>::vlanes() );
v0 = v_and(v_le(v0, thresh2), v0);
v1 = v_and(v_le(v1, thresh2), v1);
v_store( dst + j, v0 );
v_store( dst + j + VTraits<v_float64>::vlanes(), v1 );
}
if( j <= roi.width - VTraits<v_float64>::vlanes() )
{
v_float64 v0 = vx_load( src + j );
v0 = v_and(v_le(v0, thresh2), v0);
v_store( dst + j, v0 );
j += VTraits<v_float64>::vlanes();
}
for( ; j < roi.width; j++ )
dst[j] = threshToZeroInv<double>(src[j], thresh);
}
break;
default:
CV_Error(cv::Error::StsBadArg, ""); return;
}
#else
threshGeneric<double>(roi, src, src_step, dst, dst_step, thresh, maxval, type);
#endif
}
#ifdef HAVE_IPP
static bool ipp_getThreshVal_Otsu_8u( const unsigned char* _src, int step, Size size, unsigned char &thresh)
{
CV_INSTRUMENT_REGION_IPP();
// Performance degradations
#if IPP_VERSION_X100 >= 201800
IppiSize srcSize = { size.width, size.height };
if(CV_INSTRUMENT_FUN_IPP(ippiComputeThreshold_Otsu_8u_C1R, _src, step, srcSize, &thresh) < 0)
return false;
return true;
#else
CV_UNUSED(_src); CV_UNUSED(step); CV_UNUSED(size); CV_UNUSED(thresh);
return false;
#endif
}
#endif
template<typename T, size_t BinsOnStack = 0u, bool useMask = false>
static double getThreshVal_Otsu( const Mat& _src, const Mat& _mask, const Size& size )
{
const int N = std::numeric_limits<T>::max() + 1;
int i, j;
#if CV_ENABLE_UNROLLED
AutoBuffer<int, 4 * BinsOnStack> hBuf(4 * N);
#else
AutoBuffer<int, BinsOnStack> hBuf(N);
#endif
memset(hBuf.data(), 0, hBuf.size() * sizeof(int));
int* h = hBuf.data();
#if CV_ENABLE_UNROLLED
int* h_unrolled[3] = {h + N, h + 2 * N, h + 3 * N };
#endif
int maskCount = 0;
for( i = 0; i < size.height; i++ )
{
const T* src = _src.ptr<T>(i, 0);
const unsigned char* pMask = nullptr;
if ( useMask )
pMask = _mask.ptr<unsigned char>(i, 0);
j = 0;
#if CV_ENABLE_UNROLLED
for( ; j <= size.width - 4; j += 4 )
{
int v0 = src[j], v1 = src[j+1];
if ( useMask )
{
h[v0] += (pMask[j] != 0) ? ++maskCount,1 : 0;
h_unrolled[0][v1] += (pMask[j+1] != 0) ? ++maskCount,1 : 0;
}
else
{
h[v0]++;
h_unrolled[0][v1]++;
}
v0 = src[j+2]; v1 = src[j+3];
if ( useMask )
{
h_unrolled[1][v0] += (pMask[j+2] != 0) ? ++maskCount,1 : 0;
h_unrolled[2][v1] += (pMask[j+3] != 0) ? ++maskCount,1 : 0;
}
else
{
h_unrolled[1][v0]++;
h_unrolled[2][v1]++;
}
}
#endif
for( ; j < size.width; j++ )
{
if ( useMask )
h[src[j]] += (pMask[j] != 0) ? ++maskCount,1 : 0;
else
h[src[j]]++;
}
}
double mu = 0, scale = 1./( useMask ? maskCount : ( size.width*size.height ) );
for( i = 0; i < N; i++ )
{
#if CV_ENABLE_UNROLLED
h[i] += h_unrolled[0][i] + h_unrolled[1][i] + h_unrolled[2][i];
#endif
mu += i*(double)h[i];
}
mu *= scale;
double mu1 = 0, q1 = 0;
double max_sigma = 0, max_val = 0;
for(i = 0; i < N; i++ )
{
double p_i, q2, mu2, sigma;
p_i = h[i]*scale;
mu1 *= q1;
q1 += p_i;
q2 = 1. - q1;
if( std::min(q1,q2) < FLT_EPSILON || std::max(q1,q2) > 1. - FLT_EPSILON )
continue;
mu1 = (mu1 + i*p_i)/q1;
mu2 = (mu - q1*mu1)/q2;
sigma = q1*q2*(mu1 - mu2)*(mu1 - mu2);
if( sigma > max_sigma )
{
max_sigma = sigma;
max_val = i;
}
}
return max_val;
}
static double
getThreshVal_Otsu_8u( const Mat& _src, const Mat& _mask = cv::Mat())
{
Size size = _src.size();
int step = (int) _src.step;
if( _src.isContinuous() && ( _mask.empty() || _mask.isContinuous() ) )
{
size.width *= size.height;
size.height = 1;
step = size.width;
}
if (_mask.empty())
{
#ifdef HAVE_IPP
unsigned char thresh = 0;
CV_IPP_RUN_FAST(ipp_getThreshVal_Otsu_8u(_src.ptr(), step, size, thresh), thresh);
#else
CV_UNUSED(step);
#endif
}
if (!_mask.empty())
return getThreshVal_Otsu<uchar, 256u, true>(_src, _mask, size);
else
return getThreshVal_Otsu<uchar, 256u, false>(_src, _mask, size);
}
static double
getThreshVal_Otsu_16u( const Mat& _src, const Mat& _mask = cv::Mat() )
{
Size size = _src.size();
if( _src.isContinuous() && ( _mask.empty() || _mask.isContinuous() ) )
{
size.width *= size.height;
size.height = 1;
}
if (!_mask.empty())
return getThreshVal_Otsu<ushort, 0u, true>(_src, _mask, size);
else
return getThreshVal_Otsu<ushort, 0u, false>(_src, _mask, size);
}
template<bool useMask>
static double
getThreshVal_Triangle_8u( const Mat& _src, const Mat& _mask = cv::Mat() )
{
Size size = _src.size();
int step = (int) _src.step;
if( _src.isContinuous() && ( _mask.empty() || _mask.isContinuous() ) )
{
size.width *= size.height;
size.height = 1;
step = size.width;
}
const int N = 256;
int i, j, h[N] = {0};
#if CV_ENABLE_UNROLLED
int h_unrolled[3][N] = {};
#endif
for( i = 0; i < size.height; i++ )
{
const uchar* src = _src.ptr() + step*i;
const uchar* pMask = nullptr;
if ( useMask )
pMask = _mask.ptr<unsigned char>(i);
j = 0;
#if CV_ENABLE_UNROLLED
for( ; j <= size.width - 4; j += 4 )
{
int v0 = src[j], v1 = src[j+1];
if ( useMask )
{
h[v0] += (pMask[j] != 0) ? 1 : 0;
h_unrolled[0][v1] += (pMask[j+1] != 0) ? 1 : 0;
}
else
{
h[v0]++;
h_unrolled[0][v1]++;
}
v0 = src[j+2]; v1 = src[j+3];
if ( useMask )
{
h_unrolled[1][v0] += (pMask[j+2] != 0) ? 1 : 0;
h_unrolled[2][v1] += (pMask[j+3] != 0) ? 1 : 0;
}
else
{
h_unrolled[1][v0]++;
h_unrolled[2][v1]++;
}
}
#endif
for( ; j < size.width; j++ )
{
if ( useMask )
h[src[j]] += (pMask[j] != 0) ? 1 : 0;
else
h[src[j]]++;
}
}
int left_bound = 0, right_bound = 0, max_ind = 0, max = 0;
int temp;
bool isflipped = false;
#if CV_ENABLE_UNROLLED
for( i = 0; i < N; i++ )
{
h[i] += h_unrolled[0][i] + h_unrolled[1][i] + h_unrolled[2][i];
}
#endif
for( i = 0; i < N; i++ )
{
if( h[i] > 0 )
{
left_bound = i;
break;
}
}
if( left_bound > 0 )
left_bound--;
for( i = N-1; i > 0; i-- )
{
if( h[i] > 0 )
{
right_bound = i;
break;
}
}
if( right_bound < N-1 )
right_bound++;
for( i = 0; i < N; i++ )
{
if( h[i] > max)
{
max = h[i];
max_ind = i;
}
}
if( max_ind-left_bound < right_bound-max_ind)
{
isflipped = true;
i = 0, j = N-1;
while( i < j )
{
temp = h[i]; h[i] = h[j]; h[j] = temp;
i++; j--;
}
left_bound = N-1-right_bound;
max_ind = N-1-max_ind;
}
double thresh = left_bound;
double a, b, dist = 0, tempdist;
/*
* We do not need to compute precise distance here. Distance is maximized, so some constants can
* be omitted. This speeds up a computation a bit.
*/
a = max; b = left_bound-max_ind;
for( i = left_bound+1; i <= max_ind; i++ )
{
tempdist = a*i + b*h[i];
if( tempdist > dist)
{
dist = tempdist;
thresh = i;
}
}
thresh--;
if( isflipped )
thresh = N-1-thresh;
return thresh;
}
class ThresholdRunner : public ParallelLoopBody
{
public:
ThresholdRunner(Mat _src, Mat _dst, const Mat& _mask, double _thresh, double _maxval, int _thresholdType)
{
src = _src;
dst = _dst;
mask = _mask;
thresh = _thresh;
maxval = _maxval;
thresholdType = _thresholdType;
}
void operator () (const Range& range) const CV_OVERRIDE
{
int row0 = range.start;
int row1 = range.end;
Mat srcStripe = src.rowRange(row0, row1);
Mat dstStripe = dst.rowRange(row0, row1);
const bool useMask = !mask.empty();
if ( !useMask )
{
CALL_HAL(threshold, cv_hal_threshold, srcStripe.data, srcStripe.step, dstStripe.data, dstStripe.step,
srcStripe.cols, srcStripe.rows, srcStripe.depth(), srcStripe.channels(),
thresh, maxval, thresholdType);
}
if (srcStripe.depth() == CV_8U)
{
if ( useMask )
threshGenericWithMask<uchar>( srcStripe, dstStripe, mask.rowRange(row0, row1), (uchar)thresh, (uchar)maxval, thresholdType );
else
thresh_8u( srcStripe, dstStripe, (uchar)thresh, (uchar)maxval, thresholdType );
}
else if( srcStripe.depth() == CV_16S )
{
if ( useMask )
threshGenericWithMask<short>( srcStripe, dstStripe, mask.rowRange(row0, row1), (short)thresh, (short)maxval, thresholdType );
else
thresh_16s( srcStripe, dstStripe, (short)thresh, (short)maxval, thresholdType );
}
else if( srcStripe.depth() == CV_16U )
{
if ( useMask )
threshGenericWithMask<ushort>( srcStripe, dstStripe, mask.rowRange(row0, row1), (ushort)thresh, (ushort)maxval, thresholdType );
else
thresh_16u( srcStripe, dstStripe, (ushort)thresh, (ushort)maxval, thresholdType );
}
else if( srcStripe.depth() == CV_32F )
{
if ( useMask )
threshGenericWithMask<float>( srcStripe, dstStripe, mask.rowRange(row0, row1), (float)thresh, (float)maxval, thresholdType );
else
thresh_32f( srcStripe, dstStripe, (float)thresh, (float)maxval, thresholdType );
}
else if( srcStripe.depth() == CV_64F )
{
if ( useMask )
threshGenericWithMask<double>( srcStripe, dstStripe, mask.rowRange(row0, row1), thresh, maxval, thresholdType );
else
thresh_64f(srcStripe, dstStripe, thresh, maxval, thresholdType);
}
}
private:
Mat src;
Mat dst;
Mat mask;
double thresh;
double maxval;
int thresholdType;
};
#ifdef HAVE_OPENCL
static bool ocl_threshold( InputArray _src, OutputArray _dst, InputArray _mask, double & thresh, double maxval, int thresh_type )
{
int type = _src.type(), depth = CV_MAT_DEPTH(type), cn = CV_MAT_CN(type),
kercn = ocl::predictOptimalVectorWidth(_src, _dst), ktype = CV_MAKE_TYPE(depth, kercn);
bool doubleSupport = ocl::Device::getDefault().doubleFPConfig() > 0;
const bool isDisabled = ((thresh_type & THRESH_DRYRUN) != 0);
thresh_type &= ~THRESH_DRYRUN;
if ( isDisabled ||
!(thresh_type == THRESH_BINARY || thresh_type == THRESH_BINARY_INV || thresh_type == THRESH_TRUNC ||
thresh_type == THRESH_TOZERO || thresh_type == THRESH_TOZERO_INV) ||
(!doubleSupport && depth == CV_64F))
return false;
const char * const thresholdMap[] = { "THRESH_BINARY", "THRESH_BINARY_INV", "THRESH_TRUNC",
"THRESH_TOZERO", "THRESH_TOZERO_INV" };
ocl::Device dev = ocl::Device::getDefault();
int stride_size = dev.isIntel() && (dev.type() & ocl::Device::TYPE_GPU) ? 4 : 1;
const bool useMask = !_mask.empty();
ocl::Kernel k =
!useMask ?
ocl::Kernel("threshold", ocl::imgproc::threshold_oclsrc,
format("-D %s -D T=%s -D T1=%s -D STRIDE_SIZE=%d%s", thresholdMap[thresh_type],
ocl::typeToStr(ktype), ocl::typeToStr(depth), stride_size,
doubleSupport ? " -D DOUBLE_SUPPORT" : "")) :
ocl::Kernel("threshold_mask", ocl::imgproc::threshold_oclsrc,
format("-D %s -D T=%s -D T1=%s -D CN=%d -D STRIDE_SIZE=%d%s", thresholdMap[thresh_type],
ocl::typeToStr(ktype), ocl::typeToStr(depth), cn, stride_size,
doubleSupport ? " -D DOUBLE_SUPPORT" : ""));
if (k.empty())
return false;
UMat src = _src.getUMat();
_dst.create(src.size(), type);
UMat dst = _dst.getUMat();
UMat mask = !useMask ? cv::UMat() : _mask.getUMat();
if (depth <= CV_32S)
thresh = cvFloor(thresh);
const double min_vals[] = { 0, CHAR_MIN, 0, SHRT_MIN, INT_MIN, -FLT_MAX, -DBL_MAX, 0 };
double min_val = min_vals[depth];
if (!useMask)
k.args(ocl::KernelArg::ReadOnlyNoSize(src), ocl::KernelArg::WriteOnly(dst, cn, kercn),
ocl::KernelArg::Constant(Mat(1, 1, depth, Scalar::all(thresh))),
ocl::KernelArg::Constant(Mat(1, 1, depth, Scalar::all(maxval))),
ocl::KernelArg::Constant(Mat(1, 1, depth, Scalar::all(min_val))));
else
k.args(ocl::KernelArg::ReadOnlyNoSize(src), ocl::KernelArg::WriteOnly(dst, cn, kercn),
ocl::KernelArg::ReadOnlyNoSize(mask),
ocl::KernelArg::Constant(Mat(1, 1, depth, Scalar::all(thresh))),
ocl::KernelArg::Constant(Mat(1, 1, depth, Scalar::all(maxval))),
ocl::KernelArg::Constant(Mat(1, 1, depth, Scalar::all(min_val))));
size_t globalsize[2] = { (size_t)dst.cols * cn / kercn, (size_t)dst.rows };
globalsize[1] = (globalsize[1] + stride_size - 1) / stride_size;
return k.run(2, globalsize, NULL, false);
}
#endif
}
double cv::threshold( InputArray _src, OutputArray _dst, double thresh, double maxval, int type )
{
CV_INSTRUMENT_REGION();
CV_OCL_RUN_(_src.dims() <= 2 && _dst.isUMat(),
ocl_threshold(_src, _dst, cv::noArray(), thresh, maxval, type), thresh)
const bool isDisabled = ((type & THRESH_DRYRUN) != 0);
type &= ~THRESH_DRYRUN;
Mat src = _src.getMat();
if (!isDisabled)
_dst.create( src.size(), src.type() );
Mat dst = isDisabled ? cv::Mat() : _dst.getMat();
int automatic_thresh = (type & ~cv::THRESH_MASK);
type &= THRESH_MASK;
CV_Assert( automatic_thresh != (cv::THRESH_OTSU | cv::THRESH_TRIANGLE) );
if( automatic_thresh == cv::THRESH_OTSU )
{
int src_type = src.type();
CV_CheckType(src_type, src_type == CV_8UC1 || src_type == CV_16UC1, "THRESH_OTSU mode");
CALL_HAL_RET(thresholdOtsu, cv_hal_threshold_otsu, thresh, src.data, src.step, dst.data, dst.step,
src.cols, src.rows, src_type, maxval, type);
thresh = src.type() == CV_8UC1 ? getThreshVal_Otsu_8u( src )
: getThreshVal_Otsu_16u( src );
}
else if( automatic_thresh == cv::THRESH_TRIANGLE )
{
CV_Assert( src.type() == CV_8UC1 );
thresh = getThreshVal_Triangle_8u<false>( src );
}
if( src.depth() == CV_8U )
{
int ithresh = cvFloor(thresh);
thresh = ithresh;
if (isDisabled)
return thresh;
int imaxval = cvRound(maxval);
if( type == THRESH_TRUNC )
imaxval = ithresh;
imaxval = saturate_cast<uchar>(imaxval);
if( ithresh < 0 || ithresh >= 255 )
{
if( type == THRESH_BINARY || type == THRESH_BINARY_INV ||
((type == THRESH_TRUNC || type == THRESH_TOZERO_INV) && ithresh < 0) ||
(type == THRESH_TOZERO && ithresh >= 255) )
{
int v = type == THRESH_BINARY ? (ithresh >= 255 ? 0 : imaxval) :
type == THRESH_BINARY_INV ? (ithresh >= 255 ? imaxval : 0) :
/*type == THRESH_TRUNC ? imaxval :*/ 0;
dst.setTo(v);
}
else
src.copyTo(dst);
return thresh;
}
thresh = ithresh;
maxval = imaxval;
}
else if( src.depth() == CV_16S )
{
int ithresh = cvFloor(thresh);
thresh = ithresh;
if (isDisabled)
return thresh;
int imaxval = cvRound(maxval);
if( type == THRESH_TRUNC )
imaxval = ithresh;
imaxval = saturate_cast<short>(imaxval);
if( ithresh < SHRT_MIN || ithresh >= SHRT_MAX )
{
if( type == THRESH_BINARY || type == THRESH_BINARY_INV ||
((type == THRESH_TRUNC || type == THRESH_TOZERO_INV) && ithresh < SHRT_MIN) ||
(type == THRESH_TOZERO && ithresh >= SHRT_MAX) )
{
int v = type == THRESH_BINARY ? (ithresh >= SHRT_MAX ? 0 : imaxval) :
type == THRESH_BINARY_INV ? (ithresh >= SHRT_MAX ? imaxval : 0) :
/*type == THRESH_TRUNC ? imaxval :*/ 0;
dst.setTo(v);
}
else
src.copyTo(dst);
return thresh;
}
thresh = ithresh;
maxval = imaxval;
}
else if (src.depth() == CV_16U )
{
int ithresh = cvFloor(thresh);
thresh = ithresh;
if (isDisabled)
return thresh;
int imaxval = cvRound(maxval);
if (type == THRESH_TRUNC)
imaxval = ithresh;
imaxval = saturate_cast<ushort>(imaxval);
int ushrt_min = 0;
if (ithresh < ushrt_min || ithresh >= (int)USHRT_MAX)
{
if (type == THRESH_BINARY || type == THRESH_BINARY_INV ||
((type == THRESH_TRUNC || type == THRESH_TOZERO_INV) && ithresh < ushrt_min) ||
(type == THRESH_TOZERO && ithresh >= (int)USHRT_MAX))
{
int v = type == THRESH_BINARY ? (ithresh >= (int)USHRT_MAX ? 0 : imaxval) :
type == THRESH_BINARY_INV ? (ithresh >= (int)USHRT_MAX ? imaxval : 0) :
/*type == THRESH_TRUNC ? imaxval :*/ 0;
dst.setTo(v);
}
else
src.copyTo(dst);
return thresh;
}
thresh = ithresh;
maxval = imaxval;
}
else if( src.depth() == CV_32F )
;
else if( src.depth() == CV_64F )
;
else
CV_Error( cv::Error::StsUnsupportedFormat, "" );
if (isDisabled)
return thresh;
parallel_for_(Range(0, dst.rows),
ThresholdRunner(src, dst, cv::Mat(), thresh, maxval, type),
dst.total()/(double)(1<<16));
return thresh;
}
double cv::thresholdWithMask( InputArray _src, InputOutputArray _dst, InputArray _mask, double thresh, double maxval, int type )
{
CV_INSTRUMENT_REGION();
CV_Assert( _mask.empty() || ( ( _dst.size() == _src.size() ) && ( _dst.type() == _src.type() ) ) );
if ( _mask.empty() )
return cv::threshold(_src, _dst, thresh, maxval, type);
CV_OCL_RUN_(_src.dims() <= 2 && _dst.isUMat(),
ocl_threshold(_src, _dst, _mask, thresh, maxval, type), thresh)
const bool isDisabled = ((type & THRESH_DRYRUN) != 0);
type &= ~THRESH_DRYRUN;
Mat src = _src.getMat();
Mat mask = _mask.getMat();
if (!isDisabled)
_dst.create( src.size(), src.type() );
Mat dst = isDisabled ? cv::Mat() : _dst.getMat();
int automatic_thresh = (type & ~cv::THRESH_MASK);
type &= THRESH_MASK;
CV_Assert( automatic_thresh != (cv::THRESH_OTSU | cv::THRESH_TRIANGLE) );
if( automatic_thresh == cv::THRESH_OTSU )
{
int src_type = src.type();
CV_CheckType(src_type, src_type == CV_8UC1 || src_type == CV_16UC1, "THRESH_OTSU mode");
thresh = src.type() == CV_8UC1 ? getThreshVal_Otsu_8u( src, mask )
: getThreshVal_Otsu_16u( src, mask );
}
else if( automatic_thresh == cv::THRESH_TRIANGLE )
{
CV_Assert( src.type() == CV_8UC1 );
thresh = getThreshVal_Triangle_8u<true>( src, mask );
}
if( src.depth() == CV_8U )
{
int ithresh = cvFloor(thresh);
thresh = ithresh;
if (isDisabled)
return thresh;
int imaxval = cvRound(maxval);
if( type == THRESH_TRUNC )
imaxval = ithresh;
imaxval = saturate_cast<uchar>(imaxval);
if( ithresh < 0 || ithresh >= 255 )
{
if( type == THRESH_BINARY || type == THRESH_BINARY_INV ||
((type == THRESH_TRUNC || type == THRESH_TOZERO_INV) && ithresh < 0) ||
(type == THRESH_TOZERO && ithresh >= 255) )
{
int v = type == THRESH_BINARY ? (ithresh >= 255 ? 0 : imaxval) :
type == THRESH_BINARY_INV ? (ithresh >= 255 ? imaxval : 0) :
/*type == THRESH_TRUNC ? imaxval :*/ 0;
dst.setTo(v);
}
else
src.copyTo(dst);
return thresh;
}
thresh = ithresh;
maxval = imaxval;
}
else if( src.depth() == CV_16S )
{
int ithresh = cvFloor(thresh);
thresh = ithresh;
if (isDisabled)
return thresh;
int imaxval = cvRound(maxval);
if( type == THRESH_TRUNC )
imaxval = ithresh;
imaxval = saturate_cast<short>(imaxval);
if( ithresh < SHRT_MIN || ithresh >= SHRT_MAX )
{
if( type == THRESH_BINARY || type == THRESH_BINARY_INV ||
((type == THRESH_TRUNC || type == THRESH_TOZERO_INV) && ithresh < SHRT_MIN) ||
(type == THRESH_TOZERO && ithresh >= SHRT_MAX) )
{
int v = type == THRESH_BINARY ? (ithresh >= SHRT_MAX ? 0 : imaxval) :
type == THRESH_BINARY_INV ? (ithresh >= SHRT_MAX ? imaxval : 0) :
/*type == THRESH_TRUNC ? imaxval :*/ 0;
dst.setTo(v);
}
else
src.copyTo(dst);
return thresh;
}
thresh = ithresh;
maxval = imaxval;
}
else if (src.depth() == CV_16U )
{
int ithresh = cvFloor(thresh);
thresh = ithresh;
if (isDisabled)
return thresh;
int imaxval = cvRound(maxval);
if (type == THRESH_TRUNC)
imaxval = ithresh;
imaxval = saturate_cast<ushort>(imaxval);
int ushrt_min = 0;
if (ithresh < ushrt_min || ithresh >= (int)USHRT_MAX)
{
if (type == THRESH_BINARY || type == THRESH_BINARY_INV ||
((type == THRESH_TRUNC || type == THRESH_TOZERO_INV) && ithresh < ushrt_min) ||
(type == THRESH_TOZERO && ithresh >= (int)USHRT_MAX))
{
int v = type == THRESH_BINARY ? (ithresh >= (int)USHRT_MAX ? 0 : imaxval) :
type == THRESH_BINARY_INV ? (ithresh >= (int)USHRT_MAX ? imaxval : 0) :
/*type == THRESH_TRUNC ? imaxval :*/ 0;
dst.setTo(v);
}
else
src.copyTo(dst);
return thresh;
}
thresh = ithresh;
maxval = imaxval;
}
else if( src.depth() == CV_32F )
;
else if( src.depth() == CV_64F )
;
else
CV_Error( cv::Error::StsUnsupportedFormat, "" );
if (isDisabled)
return thresh;
parallel_for_(Range(0, dst.rows),
ThresholdRunner(src, dst, mask, thresh, maxval, type),
dst.total()/(double)(1<<16));
return thresh;
}
void cv::adaptiveThreshold( InputArray _src, OutputArray _dst, double maxValue,
int method, int type, int blockSize, double delta )
{
CV_INSTRUMENT_REGION();
Mat src = _src.getMat();
CV_Assert( src.type() == CV_8UC1 );
CV_Assert( blockSize % 2 == 1 && blockSize > 1 );
Size size = src.size();
_dst.create( size, src.type() );
Mat dst = _dst.getMat();
if( maxValue < 0 )
{
dst = Scalar(0);
return;
}
CALL_HAL(adaptiveThreshold, cv_hal_adaptiveThreshold, src.data, src.step, dst.data, dst.step, src.cols, src.rows,
maxValue, method, type, blockSize, delta);
Mat mean;
if( src.data != dst.data )
mean = dst;
if (method == ADAPTIVE_THRESH_MEAN_C)
boxFilter( src, mean, src.type(), Size(blockSize, blockSize),
Point(-1,-1), true, BORDER_REPLICATE|BORDER_ISOLATED );
else if (method == ADAPTIVE_THRESH_GAUSSIAN_C)
{
Mat srcfloat,meanfloat;
src.convertTo(srcfloat,CV_32F);
meanfloat=srcfloat;
GaussianBlur(srcfloat, meanfloat, Size(blockSize, blockSize), 0, 0, BORDER_REPLICATE|BORDER_ISOLATED);
meanfloat.convertTo(mean, src.type());
}
else
CV_Error( cv::Error::StsBadFlag, "Unknown/unsupported adaptive threshold method" );
int i, j;
uchar imaxval = saturate_cast<uchar>(maxValue);
int idelta = type == THRESH_BINARY ? cvCeil(delta) : cvFloor(delta);
uchar tab[768];
if( type == cv::THRESH_BINARY )
for( i = 0; i < 768; i++ )
tab[i] = (uchar)(i - 255 > -idelta ? imaxval : 0);
else if( type == cv::THRESH_BINARY_INV )
for( i = 0; i < 768; i++ )
tab[i] = (uchar)(i - 255 <= -idelta ? imaxval : 0);
else
CV_Error( cv::Error::StsBadFlag, "Unknown/unsupported threshold type" );
if( src.isContinuous() && mean.isContinuous() && dst.isContinuous() )
{
size.width *= size.height;
size.height = 1;
}
for( i = 0; i < size.height; i++ )
{
const uchar* sdata = src.ptr(i);
const uchar* mdata = mean.ptr(i);
uchar* ddata = dst.ptr(i);
for( j = 0; j < size.width; j++ )
ddata[j] = tab[sdata[j] - mdata[j] + 255];
}
}
CV_IMPL double
cvThreshold( const void* srcarr, void* dstarr, double thresh, double maxval, int type )
{
cv::Mat src = cv::cvarrToMat(srcarr), dst = cv::cvarrToMat(dstarr), dst0 = dst;
CV_Assert( src.size == dst.size && src.channels() == dst.channels() &&
(src.depth() == dst.depth() || dst.depth() == CV_8U));
thresh = cv::threshold( src, dst, thresh, maxval, type );
if( dst0.data != dst.data )
dst.convertTo( dst0, dst0.depth() );
return thresh;
}
CV_IMPL void
cvAdaptiveThreshold( const void *srcIm, void *dstIm, double maxValue,
int method, int type, int blockSize, double delta )
{
cv::Mat src = cv::cvarrToMat(srcIm), dst = cv::cvarrToMat(dstIm);
CV_Assert( src.size == dst.size && src.type() == dst.type() );
cv::adaptiveThreshold( src, dst, maxValue, method, type, blockSize, delta );
}
/* End of file. */