Merge pull request #8058 from tomoaki0705:universalIntrinsicFast

This commit is contained in:
Alexander Alekhin 2017-01-23 13:13:14 +00:00
commit 533d399dab

View File

@ -44,6 +44,7 @@ The references are:
#include "precomp.hpp" #include "precomp.hpp"
#include "fast_score.hpp" #include "fast_score.hpp"
#include "opencl_kernels_features2d.hpp" #include "opencl_kernels_features2d.hpp"
#include "opencv2/core/hal/intrin.hpp"
#include "opencv2/core/openvx/ovx_defs.hpp" #include "opencv2/core/openvx/ovx_defs.hpp"
#if defined _MSC_VER #if defined _MSC_VER
@ -58,9 +59,10 @@ void FAST_t(InputArray _img, std::vector<KeyPoint>& keypoints, int threshold, bo
{ {
Mat img = _img.getMat(); Mat img = _img.getMat();
const int K = patternSize/2, N = patternSize + K + 1; const int K = patternSize/2, N = patternSize + K + 1;
#if CV_SSE2 #if CV_SIMD128
const int quarterPatternSize = patternSize/4; const int quarterPatternSize = patternSize/4;
(void)quarterPatternSize; v_uint8x16 delta = v_setall_u8(0x80), t = v_setall_u8((char)threshold), K16 = v_setall_u8((char)K);
bool hasSimd = hasSIMD128();
#endif #endif
int i, j, k, pixel[25]; int i, j, k, pixel[25];
makeOffsets(pixel, (int)img.step, patternSize); makeOffsets(pixel, (int)img.step, patternSize);
@ -69,12 +71,6 @@ void FAST_t(InputArray _img, std::vector<KeyPoint>& keypoints, int threshold, bo
threshold = std::min(std::max(threshold, 0), 255); threshold = std::min(std::max(threshold, 0), 255);
#if CV_SSE2
__m128i delta = _mm_set1_epi8(-128), t = _mm_set1_epi8((char)threshold), K16 = _mm_set1_epi8((char)K);
(void)K16;
(void)delta;
(void)t;
#endif
uchar threshold_tab[512]; uchar threshold_tab[512];
for( i = -255; i <= 255; i++ ) for( i = -255; i <= 255; i++ )
threshold_tab[i+255] = (uchar)(i < -threshold ? 1 : i > threshold ? 2 : 0); threshold_tab[i+255] = (uchar)(i < -threshold ? 1 : i > threshold ? 2 : 0);
@ -99,30 +95,34 @@ void FAST_t(InputArray _img, std::vector<KeyPoint>& keypoints, int threshold, bo
if( i < img.rows - 3 ) if( i < img.rows - 3 )
{ {
j = 3; j = 3;
#if CV_SSE2 #if CV_SIMD128
if( hasSimd )
{
if( patternSize == 16 ) if( patternSize == 16 )
{ {
for(; j < img.cols - 16 - 3; j += 16, ptr += 16) for(; j < img.cols - 16 - 3; j += 16, ptr += 16)
{ {
__m128i m0, m1; v_uint8x16 v = v_load(ptr);
__m128i v0 = _mm_loadu_si128((const __m128i*)ptr); v_int8x16 v0 = v_reinterpret_as_s8((v + t) ^ delta);
__m128i v1 = _mm_xor_si128(_mm_subs_epu8(v0, t), delta); v_int8x16 v1 = v_reinterpret_as_s8((v - t) ^ delta);
v0 = _mm_xor_si128(_mm_adds_epu8(v0, t), delta);
__m128i x0 = _mm_sub_epi8(_mm_loadu_si128((const __m128i*)(ptr + pixel[0])), delta); v_int8x16 x0 = v_reinterpret_as_s8(v_sub_wrap(v_load(ptr + pixel[0]), delta));
__m128i x1 = _mm_sub_epi8(_mm_loadu_si128((const __m128i*)(ptr + pixel[quarterPatternSize])), delta); v_int8x16 x1 = v_reinterpret_as_s8(v_sub_wrap(v_load(ptr + pixel[quarterPatternSize]), delta));
__m128i x2 = _mm_sub_epi8(_mm_loadu_si128((const __m128i*)(ptr + pixel[2*quarterPatternSize])), delta); v_int8x16 x2 = v_reinterpret_as_s8(v_sub_wrap(v_load(ptr + pixel[2*quarterPatternSize]), delta));
__m128i x3 = _mm_sub_epi8(_mm_loadu_si128((const __m128i*)(ptr + pixel[3*quarterPatternSize])), delta); v_int8x16 x3 = v_reinterpret_as_s8(v_sub_wrap(v_load(ptr + pixel[3*quarterPatternSize]), delta));
m0 = _mm_and_si128(_mm_cmpgt_epi8(x0, v0), _mm_cmpgt_epi8(x1, v0));
m1 = _mm_and_si128(_mm_cmpgt_epi8(v1, x0), _mm_cmpgt_epi8(v1, x1)); v_int8x16 m0, m1;
m0 = _mm_or_si128(m0, _mm_and_si128(_mm_cmpgt_epi8(x1, v0), _mm_cmpgt_epi8(x2, v0))); m0 = (v0 < x0) & (v0 < x1);
m1 = _mm_or_si128(m1, _mm_and_si128(_mm_cmpgt_epi8(v1, x1), _mm_cmpgt_epi8(v1, x2))); m1 = (x0 < v1) & (x1 < v1);
m0 = _mm_or_si128(m0, _mm_and_si128(_mm_cmpgt_epi8(x2, v0), _mm_cmpgt_epi8(x3, v0))); m0 = m0 | ((v0 < x1) & (v0 < x2));
m1 = _mm_or_si128(m1, _mm_and_si128(_mm_cmpgt_epi8(v1, x2), _mm_cmpgt_epi8(v1, x3))); m1 = m1 | ((x1 < v1) & (x2 < v1));
m0 = _mm_or_si128(m0, _mm_and_si128(_mm_cmpgt_epi8(x3, v0), _mm_cmpgt_epi8(x0, v0))); m0 = m0 | ((v0 < x2) & (v0 < x3));
m1 = _mm_or_si128(m1, _mm_and_si128(_mm_cmpgt_epi8(v1, x3), _mm_cmpgt_epi8(v1, x0))); m1 = m1 | ((x2 < v1) & (x3 < v1));
m0 = _mm_or_si128(m0, m1); m0 = m0 | ((v0 < x3) & (v0 < x0));
int mask = _mm_movemask_epi8(m0); m1 = m1 | ((x3 < v1) & (x0 < v1));
m0 = m0 | m1;
int mask = v_signmask(m0);
if( mask == 0 ) if( mask == 0 )
continue; continue;
if( (mask & 255) == 0 ) if( (mask & 255) == 0 )
@ -132,24 +132,28 @@ void FAST_t(InputArray _img, std::vector<KeyPoint>& keypoints, int threshold, bo
continue; continue;
} }
__m128i c0 = _mm_setzero_si128(), c1 = c0, max0 = c0, max1 = c0; v_int8x16 c0 = v_setzero_s8();
v_int8x16 c1 = v_setzero_s8();
v_uint8x16 max0 = v_setzero_u8();
v_uint8x16 max1 = v_setzero_u8();
for( k = 0; k < N; k++ ) for( k = 0; k < N; k++ )
{ {
__m128i x = _mm_xor_si128(_mm_loadu_si128((const __m128i*)(ptr + pixel[k])), delta); v_int8x16 x = v_reinterpret_as_s8(v_load((ptr + pixel[k])) ^ delta);
m0 = _mm_cmpgt_epi8(x, v0); m0 = v0 < x;
m1 = _mm_cmpgt_epi8(v1, x); m1 = x < v1;
c0 = _mm_and_si128(_mm_sub_epi8(c0, m0), m0); c0 = v_sub_wrap(c0, m0) & m0;
c1 = _mm_and_si128(_mm_sub_epi8(c1, m1), m1); c1 = v_sub_wrap(c1, m1) & m1;
max0 = _mm_max_epu8(max0, c0); max0 = v_max(max0, v_reinterpret_as_u8(c0));
max1 = _mm_max_epu8(max1, c1); max1 = v_max(max1, v_reinterpret_as_u8(c1));
} }
max0 = _mm_max_epu8(max0, max1); max0 = v_max(max0, max1);
int m = _mm_movemask_epi8(_mm_cmpgt_epi8(max0, K16)); int m = v_signmask(K16 < max0);
for( k = 0; m > 0 && k < 16; k++, m >>= 1 ) for( k = 0; m > 0 && k < 16; k++, m >>= 1 )
{
if(m & 1) if(m & 1)
{ {
cornerpos[ncorners++] = j+k; cornerpos[ncorners++] = j+k;
@ -158,6 +162,8 @@ void FAST_t(InputArray _img, std::vector<KeyPoint>& keypoints, int threshold, bo
} }
} }
} }
}
}
#endif #endif
for( ; j < img.cols - 3; j++, ptr++ ) for( ; j < img.cols - 3; j++, ptr++ )
{ {