/*M/////////////////////////////////////////////////////////////////////////////////////// // // IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING. // // By downloading, copying, installing or using the software you agree to this license. // If you do not agree to this license, do not download, install, // copy or use the software. // // // License Agreement // For Open Source Computer Vision Library // // Copyright (C) 2010-2012, Institute Of Software Chinese Academy Of Science, all rights reserved. // Copyright (C) 2010-2012, Advanced Micro Devices, Inc., all rights reserved. // Third party copyrights are property of their respective owners. // // @Authors // Jia Haipeng, jiahaipeng95@gmail.com // // // Redistribution and use in source and binary forms, with or without modification, // are permitted provided that the following conditions are met: // // * Redistribution's of source code must retain the above copyright notice, // this list of conditions and the following disclaimer. // // * Redistribution's in binary form must reproduce the above copyright notice, // this list of conditions and the following disclaimer in the documentation // and/or other oclMaterials provided with the distribution. // // * The name of the copyright holders may not be used to endorse or promote products // derived from this software without specific prior written permission. // // This software is provided by the copyright holders and contributors as is and // any express or implied warranties, including, but not limited to, the implied // warranties of merchantability and fitness for a particular purpose are disclaimed. // In no event shall the Intel Corporation or contributors be liable for any direct, // indirect, incidental, special, exemplary, or consequential damages // (including, but not limited to, procurement of substitute goods or services; // loss of use, data, or profits; or business interruption) however caused // and on any theory of liability, whether in contract, strict liability, // or tort (including negligence or otherwise) arising in any way out of // the use of this software, even if advised of the possibility of such damage. // //M*/ #if defined (DOUBLE_SUPPORT) #ifdef cl_khr_fp64 #pragma OPENCL EXTENSION cl_khr_fp64:enable #elif defined (cl_amd_fp64) #pragma OPENCL EXTENSION cl_amd_fp64:enable #endif #endif #ifdef ARITHM_ADD #define ARITHM_OP(A,B) ((A)+(B)) #elif defined ARITHM_SUB #define ARITHM_OP(A,B) ((A)-(B)) #endif ////////////////////////////////////////////////////////////////////////////////////////////////////// /////////////////////////////////////////////ADD//////////////////////////////////////////////////// /////////////////////////////////////////////////////////////////////////////////////////////////////// /**************************************add without mask**************************************/ __kernel void arithm_add_D0 (__global uchar *src1, int src1_step, int src1_offset, __global uchar *src2, int src2_step, int src2_offset, __global uchar *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { x = x << 2; #ifdef dst_align #undef dst_align #endif #define dst_align (dst_offset & 3) int src1_index = mad24(y, src1_step, x + src1_offset - dst_align); int src2_index = mad24(y, src2_step, x + src2_offset - dst_align); int dst_start = mad24(y, dst_step, dst_offset); int dst_end = mad24(y, dst_step, dst_offset + dst_step1); int dst_index = mad24(y, dst_step, dst_offset + x & (int)0xfffffffc); int src1_index_fix = src1_index < 0 ? 0 : src1_index; int src2_index_fix = src2_index < 0 ? 0 : src2_index; uchar4 src1_data = vload4(0, src1 + src1_index_fix); uchar4 src2_data = vload4(0, src2 + src2_index_fix); if(src1_index < 0) { uchar4 tmp; tmp.xyzw = (src1_index == -2) ? src1_data.zwxy:src1_data.yzwx; src1_data.xyzw = (src1_index == -1) ? src1_data.wxyz:tmp.xyzw; } if(src2_index < 0) { uchar4 tmp; tmp.xyzw = (src2_index == -2) ? src2_data.zwxy:src2_data.yzwx; src2_data.xyzw = (src2_index == -1) ? src2_data.wxyz:tmp.xyzw; } uchar4 dst_data = *((__global uchar4 *)(dst + dst_index)); short4 tmp = ARITHM_OP(convert_short4_sat(src1_data), convert_short4_sat(src2_data)); uchar4 tmp_data = convert_uchar4_sat(tmp); dst_data.x = ((dst_index + 0 >= dst_start) && (dst_index + 0 < dst_end)) ? tmp_data.x : dst_data.x; dst_data.y = ((dst_index + 1 >= dst_start) && (dst_index + 1 < dst_end)) ? tmp_data.y : dst_data.y; dst_data.z = ((dst_index + 2 >= dst_start) && (dst_index + 2 < dst_end)) ? tmp_data.z : dst_data.z; dst_data.w = ((dst_index + 3 >= dst_start) && (dst_index + 3 < dst_end)) ? tmp_data.w : dst_data.w; *((__global uchar4 *)(dst + dst_index)) = dst_data; } } __kernel void arithm_add_D2 (__global ushort *src1, int src1_step, int src1_offset, __global ushort *src2, int src2_step, int src2_offset, __global ushort *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { x = x << 2; #ifdef dst_align #undef dst_align #endif #define dst_align ((dst_offset >> 1) & 3) int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1)); int src2_index = mad24(y, src2_step, (x << 1) + src2_offset - (dst_align << 1)); int dst_start = mad24(y, dst_step, dst_offset); int dst_end = mad24(y, dst_step, dst_offset + dst_step1); int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffff8); ushort4 src1_data = vload4(0, (__global ushort *)((__global char *)src1 + src1_index)); ushort4 src2_data = vload4(0, (__global ushort *)((__global char *)src2 + src2_index)); ushort4 dst_data = *((__global ushort4 *)((__global char *)dst + dst_index)); int4 tmp = ARITHM_OP(convert_int4_sat(src1_data), convert_int4_sat(src2_data)); ushort4 tmp_data = convert_ushort4_sat(tmp); dst_data.x = ((dst_index + 0 >= dst_start) && (dst_index + 0 < dst_end)) ? tmp_data.x : dst_data.x; dst_data.y = ((dst_index + 2 >= dst_start) && (dst_index + 2 < dst_end)) ? tmp_data.y : dst_data.y; dst_data.z = ((dst_index + 4 >= dst_start) && (dst_index + 4 < dst_end)) ? tmp_data.z : dst_data.z; dst_data.w = ((dst_index + 6 >= dst_start) && (dst_index + 6 < dst_end)) ? tmp_data.w : dst_data.w; *((__global ushort4 *)((__global char *)dst + dst_index)) = dst_data; } } __kernel void arithm_add_D3 (__global short *src1, int src1_step, int src1_offset, __global short *src2, int src2_step, int src2_offset, __global short *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { x = x << 2; #ifdef dst_align #undef dst_align #endif #define dst_align ((dst_offset >> 1) & 3) int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1)); int src2_index = mad24(y, src2_step, (x << 1) + src2_offset - (dst_align << 1)); int dst_start = mad24(y, dst_step, dst_offset); int dst_end = mad24(y, dst_step, dst_offset + dst_step1); int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffff8); short4 src1_data = vload4(0, (__global short *)((__global char *)src1 + src1_index)); short4 src2_data = vload4(0, (__global short *)((__global char *)src2 + src2_index)); short4 dst_data = *((__global short4 *)((__global char *)dst + dst_index)); int4 tmp = ARITHM_OP(convert_int4_sat(src1_data), convert_int4_sat(src2_data)); short4 tmp_data = convert_short4_sat(tmp); dst_data.x = ((dst_index + 0 >= dst_start) && (dst_index + 0 < dst_end)) ? tmp_data.x : dst_data.x; dst_data.y = ((dst_index + 2 >= dst_start) && (dst_index + 2 < dst_end)) ? tmp_data.y : dst_data.y; dst_data.z = ((dst_index + 4 >= dst_start) && (dst_index + 4 < dst_end)) ? tmp_data.z : dst_data.z; dst_data.w = ((dst_index + 6 >= dst_start) && (dst_index + 6 < dst_end)) ? tmp_data.w : dst_data.w; *((__global short4 *)((__global char *)dst + dst_index)) = dst_data; } } __kernel void arithm_add_D4 (__global int *src1, int src1_step, int src1_offset, __global int *src2, int src2_step, int src2_offset, __global int *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { int src1_index = mad24(y, src1_step, (x << 2) + src1_offset); int src2_index = mad24(y, src2_step, (x << 2) + src2_offset); int dst_index = mad24(y, dst_step, (x << 2) + dst_offset); int data1 = *((__global int *)((__global char *)src1 + src1_index)); int data2 = *((__global int *)((__global char *)src2 + src2_index)); long tmp = ARITHM_OP((long)(data1), (long)(data2)); *((__global int *)((__global char *)dst + dst_index)) = convert_int_sat(tmp); } } __kernel void arithm_add_D5 (__global float *src1, int src1_step, int src1_offset, __global float *src2, int src2_step, int src2_offset, __global float *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { int src1_index = mad24(y, src1_step, (x << 2) + src1_offset); int src2_index = mad24(y, src2_step, (x << 2) + src2_offset); int dst_index = mad24(y, dst_step, (x << 2) + dst_offset); float data1 = *((__global float *)((__global char *)src1 + src1_index)); float data2 = *((__global float *)((__global char *)src2 + src2_index)); float tmp = ARITHM_OP(data1, data2); *((__global float *)((__global char *)dst + dst_index)) = tmp; } } #if defined (DOUBLE_SUPPORT) __kernel void arithm_add_D6 (__global double *src1, int src1_step, int src1_offset, __global double *src2, int src2_step, int src2_offset, __global double *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { int src1_index = mad24(y, src1_step, (x << 3) + src1_offset); int src2_index = mad24(y, src2_step, (x << 3) + src2_offset); int dst_index = mad24(y, dst_step, (x << 3) + dst_offset); double data1 = *((__global double *)((__global char *)src1 + src1_index)); double data2 = *((__global double *)((__global char *)src2 + src2_index)); *((__global double *)((__global char *)dst + dst_index)) = ARITHM_OP(data1, data2); } } #endif /**************************************add with mask**************************************/ __kernel void arithm_add_with_mask_C1_D0 (__global uchar *src1, int src1_step, int src1_offset, __global uchar *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global uchar *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { x = x << 2; #ifdef dst_align #undef dst_align #endif #define dst_align (dst_offset & 3) int src1_index = mad24(y, src1_step, x + src1_offset - dst_align); int src2_index = mad24(y, src2_step, x + src2_offset - dst_align); int mask_index = mad24(y, mask_step, x + mask_offset - dst_align); int dst_start = mad24(y, dst_step, dst_offset); int dst_end = mad24(y, dst_step, dst_offset + dst_step1); int dst_index = mad24(y, dst_step, dst_offset + x & (int)0xfffffffc); int src1_index_fix = src1_index < 0 ? 0 : src1_index; int src2_index_fix = src2_index < 0 ? 0 : src2_index; int mask_index_fix = mask_index < 0 ? 0 : mask_index; uchar4 src1_data = vload4(0, src1 + src1_index_fix); uchar4 src2_data = vload4(0, src2 + src2_index_fix); uchar4 mask_data = vload4(0, mask + mask_index_fix); if(src1_index < 0) { uchar4 tmp; tmp.xyzw = (src1_index == -2) ? src1_data.zwxy:src1_data.yzwx; src1_data.xyzw = (src1_index == -1) ? src1_data.wxyz:tmp.xyzw; } if(src2_index < 0) { uchar4 tmp; tmp.xyzw = (src2_index == -2) ? src2_data.zwxy:src2_data.yzwx; src2_data.xyzw = (src2_index == -1) ? src2_data.wxyz:tmp.xyzw; } if(mask_index < 0) { uchar4 tmp; tmp.xyzw = (mask_index == -2) ? mask_data.zwxy:mask_data.yzwx; mask_data.xyzw = (mask_index == -1) ? mask_data.wxyz:tmp.xyzw; } uchar4 data = *((__global uchar4 *)(dst + dst_index)); short4 tmp = ARITHM_OP(convert_short4_sat(src1_data), convert_short4_sat(src2_data)); uchar4 tmp_data = convert_uchar4_sat(tmp); data.x = ((mask_data.x) && (dst_index + 0 >= dst_start) && (dst_index + 0 < dst_end)) ? tmp_data.x : data.x; data.y = ((mask_data.y) && (dst_index + 1 >= dst_start) && (dst_index + 1 < dst_end)) ? tmp_data.y : data.y; data.z = ((mask_data.z) && (dst_index + 2 >= dst_start) && (dst_index + 2 < dst_end)) ? tmp_data.z : data.z; data.w = ((mask_data.w) && (dst_index + 3 >= dst_start) && (dst_index + 3 < dst_end)) ? tmp_data.w : data.w; *((__global uchar4 *)(dst + dst_index)) = data; } } __kernel void arithm_add_with_mask_C1_D2 (__global ushort *src1, int src1_step, int src1_offset, __global ushort *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global ushort *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { x = x << 1; #ifdef dst_align #undef dst_align #endif #define dst_align ((dst_offset >> 1) & 1) int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1)); int src2_index = mad24(y, src2_step, (x << 1) + src2_offset - (dst_align << 1)); int mask_index = mad24(y, mask_step, x + mask_offset - dst_align); int dst_start = mad24(y, dst_step, dst_offset); int dst_end = mad24(y, dst_step, dst_offset + dst_step1); int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffffc); ushort2 src1_data = vload2(0, (__global ushort *)((__global char *)src1 + src1_index)); ushort2 src2_data = vload2(0, (__global ushort *)((__global char *)src2 + src2_index)); uchar2 mask_data = vload2(0, mask + mask_index); ushort2 data = *((__global ushort2 *)((__global uchar *)dst + dst_index)); int2 tmp = ARITHM_OP(convert_int2_sat(src1_data), convert_int2_sat(src2_data)); ushort2 tmp_data = convert_ushort2_sat(tmp); data.x = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.x : data.x; data.y = ((mask_data.y) && (dst_index + 2 < dst_end )) ? tmp_data.y : data.y; *((__global ushort2 *)((__global uchar *)dst + dst_index)) = data; } } __kernel void arithm_add_with_mask_C1_D3 (__global short *src1, int src1_step, int src1_offset, __global short *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global short *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { x = x << 1; #ifdef dst_align #undef dst_align #endif #define dst_align ((dst_offset >> 1) & 1) int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1)); int src2_index = mad24(y, src2_step, (x << 1) + src2_offset - (dst_align << 1)); int mask_index = mad24(y, mask_step, x + mask_offset - dst_align); int dst_start = mad24(y, dst_step, dst_offset); int dst_end = mad24(y, dst_step, dst_offset + dst_step1); int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffffc); short2 src1_data = vload2(0, (__global short *)((__global char *)src1 + src1_index)); short2 src2_data = vload2(0, (__global short *)((__global char *)src2 + src2_index)); uchar2 mask_data = vload2(0, mask + mask_index); short2 data = *((__global short2 *)((__global uchar *)dst + dst_index)); int2 tmp = ARITHM_OP(convert_int2_sat(src1_data), convert_int2_sat(src2_data)); short2 tmp_data = convert_short2_sat(tmp); data.x = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.x : data.x; data.y = ((mask_data.y) && (dst_index + 2 < dst_end )) ? tmp_data.y : data.y; *((__global short2 *)((__global uchar *)dst + dst_index)) = data; } } __kernel void arithm_add_with_mask_C1_D4 (__global int *src1, int src1_step, int src1_offset, __global int *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global int *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { int src1_index = mad24(y, src1_step, (x << 2) + src1_offset); int src2_index = mad24(y, src2_step, (x << 2) + src2_offset); int mask_index = mad24(y, mask_step, x + mask_offset); int dst_index = mad24(y, dst_step, (x << 2) + dst_offset); uchar mask_data = *(mask + mask_index); int src_data1 = *((__global int *)((__global char *)src1 + src1_index)); int src_data2 = *((__global int *)((__global char *)src2 + src2_index)); int dst_data = *((__global int *)((__global char *)dst + dst_index)); int data = convert_int_sat(ARITHM_OP((long)src_data1, (long)src_data2)); data = mask_data ? data : dst_data; *((__global int *)((__global char *)dst + dst_index)) = data; } } __kernel void arithm_add_with_mask_C1_D5 (__global float *src1, int src1_step, int src1_offset, __global float *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global float *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { int src1_index = mad24(y, src1_step, (x << 2) + src1_offset); int src2_index = mad24(y, src2_step, (x << 2) + src2_offset); int mask_index = mad24(y, mask_step, x + mask_offset); int dst_index = mad24(y, dst_step, (x << 2) + dst_offset); uchar mask_data = *(mask + mask_index); float src_data1 = *((__global float *)((__global char *)src1 + src1_index)); float src_data2 = *((__global float *)((__global char *)src2 + src2_index)); float dst_data = *((__global float *)((__global char *)dst + dst_index)); float data = ARITHM_OP(src_data1, src_data2); data = mask_data ? data : dst_data; *((__global float *)((__global char *)dst + dst_index)) = data; } } #if defined (DOUBLE_SUPPORT) __kernel void arithm_add_with_mask_C1_D6 (__global double *src1, int src1_step, int src1_offset, __global double *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global double *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { int src1_index = mad24(y, src1_step, (x << 3) + src1_offset); int src2_index = mad24(y, src2_step, (x << 3) + src2_offset); int mask_index = mad24(y, mask_step, x + mask_offset); int dst_index = mad24(y, dst_step, (x << 3) + dst_offset); uchar mask_data = *(mask + mask_index); double src_data1 = *((__global double *)((__global char *)src1 + src1_index)); double src_data2 = *((__global double *)((__global char *)src2 + src2_index)); double dst_data = *((__global double *)((__global char *)dst + dst_index)); double data = ARITHM_OP(src_data1, src_data2); data = mask_data ? data : dst_data; *((__global double *)((__global char *)dst + dst_index)) = data; } } #endif __kernel void arithm_add_with_mask_C2_D0 (__global uchar *src1, int src1_step, int src1_offset, __global uchar *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global uchar *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { x = x << 1; #ifdef dst_align #undef dst_align #endif #define dst_align ((dst_offset >> 1) & 1) int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1)); int src2_index = mad24(y, src2_step, (x << 1) + src2_offset - (dst_align << 1)); int mask_index = mad24(y, mask_step, x + mask_offset - dst_align); int dst_start = mad24(y, dst_step, dst_offset); int dst_end = mad24(y, dst_step, dst_offset + dst_step1); int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffffc); uchar4 src1_data = vload4(0, src1 + src1_index); uchar4 src2_data = vload4(0, src2 + src2_index); uchar2 mask_data = vload2(0, mask + mask_index); uchar4 data = *((__global uchar4 *)(dst + dst_index)); short4 tmp = ARITHM_OP(convert_short4_sat(src1_data), convert_short4_sat(src2_data)); uchar4 tmp_data = convert_uchar4_sat(tmp); data.xy = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.xy : data.xy; data.zw = ((mask_data.y) && (dst_index + 2 < dst_end )) ? tmp_data.zw : data.zw; *((__global uchar4 *)(dst + dst_index)) = data; } } __kernel void arithm_add_with_mask_C2_D2 (__global ushort *src1, int src1_step, int src1_offset, __global ushort *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global ushort *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { int src1_index = mad24(y, src1_step, (x << 2) + src1_offset); int src2_index = mad24(y, src2_step, (x << 2) + src2_offset); int mask_index = mad24(y, mask_step, x + mask_offset); int dst_index = mad24(y, dst_step, (x << 2) + dst_offset); uchar mask_data = *(mask + mask_index); ushort2 src_data1 = *((__global ushort2 *)((__global char *)src1 + src1_index)); ushort2 src_data2 = *((__global ushort2 *)((__global char *)src2 + src2_index)); ushort2 dst_data = *((__global ushort2 *)((__global char *)dst + dst_index)); int2 tmp = ARITHM_OP(convert_int2_sat(src_data1), convert_int2_sat(src_data2)); ushort2 data = convert_ushort2_sat(tmp); data = mask_data ? data : dst_data; *((__global ushort2 *)((__global char *)dst + dst_index)) = data; } } __kernel void arithm_add_with_mask_C2_D3 (__global short *src1, int src1_step, int src1_offset, __global short *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global short *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { int src1_index = mad24(y, src1_step, (x << 2) + src1_offset); int src2_index = mad24(y, src2_step, (x << 2) + src2_offset); int mask_index = mad24(y, mask_step, x + mask_offset); int dst_index = mad24(y, dst_step, (x << 2) + dst_offset); uchar mask_data = *(mask + mask_index); short2 src_data1 = *((__global short2 *)((__global char *)src1 + src1_index)); short2 src_data2 = *((__global short2 *)((__global char *)src2 + src2_index)); short2 dst_data = *((__global short2 *)((__global char *)dst + dst_index)); int2 tmp = ARITHM_OP(convert_int2_sat(src_data1), convert_int2_sat(src_data2)); short2 data = convert_short2_sat(tmp); data = mask_data ? data : dst_data; *((__global short2 *)((__global char *)dst + dst_index)) = data; } } __kernel void arithm_add_with_mask_C2_D4 (__global int *src1, int src1_step, int src1_offset, __global int *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global int *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { int src1_index = mad24(y, src1_step, (x << 3) + src1_offset); int src2_index = mad24(y, src2_step, (x << 3) + src2_offset); int mask_index = mad24(y, mask_step, x + mask_offset); int dst_index = mad24(y, dst_step, (x << 3) + dst_offset); uchar mask_data = *(mask + mask_index); int2 src_data1 = *((__global int2 *)((__global char *)src1 + src1_index)); int2 src_data2 = *((__global int2 *)((__global char *)src2 + src2_index)); int2 dst_data = *((__global int2 *)((__global char *)dst + dst_index)); int2 data = convert_int2_sat(ARITHM_OP(convert_long2_sat(src_data1), convert_long2_sat(src_data2))); data = mask_data ? data : dst_data; *((__global int2 *)((__global char *)dst + dst_index)) = data; } } __kernel void arithm_add_with_mask_C2_D5 (__global float *src1, int src1_step, int src1_offset, __global float *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global float *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { int src1_index = mad24(y, src1_step, (x << 3) + src1_offset); int src2_index = mad24(y, src2_step, (x << 3) + src2_offset); int mask_index = mad24(y, mask_step, x + mask_offset); int dst_index = mad24(y, dst_step, (x << 3) + dst_offset); uchar mask_data = *(mask + mask_index); float2 src_data1 = *((__global float2 *)((__global char *)src1 + src1_index)); float2 src_data2 = *((__global float2 *)((__global char *)src2 + src2_index)); float2 dst_data = *((__global float2 *)((__global char *)dst + dst_index)); float2 data = ARITHM_OP(src_data1, src_data2); data = mask_data ? data : dst_data; *((__global float2 *)((__global char *)dst + dst_index)) = data; } } #if defined (DOUBLE_SUPPORT) __kernel void arithm_add_with_mask_C2_D6 (__global double *src1, int src1_step, int src1_offset, __global double *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global double *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { int src1_index = mad24(y, src1_step, (x << 4) + src1_offset); int src2_index = mad24(y, src2_step, (x << 4) + src2_offset); int mask_index = mad24(y, mask_step, x + mask_offset); int dst_index = mad24(y, dst_step, (x << 4) + dst_offset); uchar mask_data = *(mask + mask_index); double2 src_data1 = *((__global double2 *)((__global char *)src1 + src1_index)); double2 src_data2 = *((__global double2 *)((__global char *)src2 + src2_index)); double2 dst_data = *((__global double2 *)((__global char *)dst + dst_index)); double2 data = ARITHM_OP(src_data1, src_data2); data = mask_data ? data : dst_data; *((__global double2 *)((__global char *)dst + dst_index)) = data; } } #endif __kernel void arithm_add_with_mask_C4_D0 (__global uchar *src1, int src1_step, int src1_offset, __global uchar *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global uchar *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { int src1_index = mad24(y, src1_step, (x << 2) + src1_offset); int src2_index = mad24(y, src2_step, (x << 2) + src2_offset); int mask_index = mad24(y, mask_step, x + mask_offset); int dst_index = mad24(y, dst_step, (x << 2) + dst_offset); uchar mask_data = *(mask + mask_index); uchar4 src_data1 = *((__global uchar4 *)(src1 + src1_index)); uchar4 src_data2 = *((__global uchar4 *)(src2 + src2_index)); uchar4 dst_data = *((__global uchar4 *)(dst + dst_index)); uchar4 data = convert_uchar4_sat(ARITHM_OP(convert_short4_sat(src_data1), convert_short4_sat(src_data2))); data = mask_data ? data : dst_data; *((__global uchar4 *)(dst + dst_index)) = data; } } __kernel void arithm_add_with_mask_C4_D2 (__global ushort *src1, int src1_step, int src1_offset, __global ushort *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global ushort *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { int src1_index = mad24(y, src1_step, (x << 3) + src1_offset); int src2_index = mad24(y, src2_step, (x << 3) + src2_offset); int mask_index = mad24(y, mask_step, x + mask_offset); int dst_index = mad24(y, dst_step, (x << 3) + dst_offset); uchar mask_data = *(mask + mask_index); ushort4 src_data1 = *((__global ushort4 *)((__global char *)src1 + src1_index)); ushort4 src_data2 = *((__global ushort4 *)((__global char *)src2 + src2_index)); ushort4 dst_data = *((__global ushort4 *)((__global char *)dst + dst_index)); ushort4 data = convert_ushort4_sat(ARITHM_OP(convert_int4_sat(src_data1), convert_int4_sat(src_data2))); data = mask_data ? data : dst_data; *((__global ushort4 *)((__global char *)dst + dst_index)) = data; } } __kernel void arithm_add_with_mask_C4_D3 (__global short *src1, int src1_step, int src1_offset, __global short *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global short *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { int src1_index = mad24(y, src1_step, (x << 3) + src1_offset); int src2_index = mad24(y, src2_step, (x << 3) + src2_offset); int mask_index = mad24(y, mask_step, x + mask_offset); int dst_index = mad24(y, dst_step, (x << 3) + dst_offset); uchar mask_data = *(mask + mask_index); short4 src_data1 = *((__global short4 *)((__global char *)src1 + src1_index)); short4 src_data2 = *((__global short4 *)((__global char *)src2 + src2_index)); short4 dst_data = *((__global short4 *)((__global char *)dst + dst_index)); short4 data = convert_short4_sat(ARITHM_OP(convert_int4_sat(src_data1), convert_int4_sat(src_data2))); data = mask_data ? data : dst_data; *((__global short4 *)((__global char *)dst + dst_index)) = data; } } __kernel void arithm_add_with_mask_C4_D4 (__global int *src1, int src1_step, int src1_offset, __global int *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global int *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { int src1_index = mad24(y, src1_step, (x << 4) + src1_offset); int src2_index = mad24(y, src2_step, (x << 4) + src2_offset); int mask_index = mad24(y, mask_step, x + mask_offset); int dst_index = mad24(y, dst_step, (x << 4) + dst_offset); uchar mask_data = *(mask + mask_index); int4 src_data1 = *((__global int4 *)((__global char *)src1 + src1_index)); int4 src_data2 = *((__global int4 *)((__global char *)src2 + src2_index)); int4 dst_data = *((__global int4 *)((__global char *)dst + dst_index)); int4 data = convert_int4_sat(ARITHM_OP(convert_long4_sat(src_data1), convert_long4_sat(src_data2))); data = mask_data ? data : dst_data; *((__global int4 *)((__global char *)dst + dst_index)) = data; } } __kernel void arithm_add_with_mask_C4_D5 (__global float *src1, int src1_step, int src1_offset, __global float *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global float *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { int src1_index = mad24(y, src1_step, (x << 4) + src1_offset); int src2_index = mad24(y, src2_step, (x << 4) + src2_offset); int mask_index = mad24(y, mask_step, x + mask_offset); int dst_index = mad24(y, dst_step, (x << 4) + dst_offset); uchar mask_data = *(mask + mask_index); float4 src_data1 = *((__global float4 *)((__global char *)src1 + src1_index)); float4 src_data2 = *((__global float4 *)((__global char *)src2 + src2_index)); float4 dst_data = *((__global float4 *)((__global char *)dst + dst_index)); float4 data = ARITHM_OP(src_data1, src_data2); data = mask_data ? data : dst_data; *((__global float4 *)((__global char *)dst + dst_index)) = data; } } #if defined (DOUBLE_SUPPORT) __kernel void arithm_add_with_mask_C4_D6 (__global double *src1, int src1_step, int src1_offset, __global double *src2, int src2_step, int src2_offset, __global uchar *mask, int mask_step, int mask_offset, __global double *dst, int dst_step, int dst_offset, int rows, int cols, int dst_step1) { int x = get_global_id(0); int y = get_global_id(1); if (x < cols && y < rows) { int src1_index = mad24(y, src1_step, (x << 5) + src1_offset); int src2_index = mad24(y, src2_step, (x << 5) + src2_offset); int mask_index = mad24(y, mask_step, x + mask_offset); int dst_index = mad24(y, dst_step, (x << 5) + dst_offset); uchar mask_data = *(mask + mask_index); double4 src_data1 = *((__global double4 *)((__global char *)src1 + src1_index)); double4 src_data2 = *((__global double4 *)((__global char *)src2 + src2_index)); double4 dst_data = *((__global double4 *)((__global char *)dst + dst_index)); double4 data = ARITHM_OP(src_data1, src_data2); data = mask_data ? data : dst_data; *((__global double4 *)((__global char *)dst + dst_index)) = data; } } #endif