2012-07-17 01:08:14 +08:00
|
|
|
/*M///////////////////////////////////////////////////////////////////////////////////////
|
|
|
|
//
|
|
|
|
// IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING.
|
|
|
|
//
|
|
|
|
// By downloading, copying, installing or using the software you agree to this license.
|
|
|
|
// If you do not agree to this license, do not download, install,
|
|
|
|
// copy or use the software.
|
|
|
|
//
|
|
|
|
//
|
|
|
|
// License Agreement
|
|
|
|
// For Open Source Computer Vision Library
|
|
|
|
//
|
|
|
|
// Copyright (C) 2010-2012, Institute Of Software Chinese Academy Of Science, all rights reserved.
|
|
|
|
// Copyright (C) 2010-2012, Advanced Micro Devices, Inc., all rights reserved.
|
|
|
|
// Third party copyrights are property of their respective owners.
|
|
|
|
//
|
|
|
|
// @Authors
|
|
|
|
// Jia Haipeng, jiahaipeng95@gmail.com
|
|
|
|
//
|
|
|
|
// Redistribution and use in source and binary forms, with or without modification,
|
|
|
|
// are permitted provided that the following conditions are met:
|
|
|
|
//
|
|
|
|
// * Redistribution's of source code must retain the above copyright notice,
|
|
|
|
// this list of conditions and the following disclaimer.
|
|
|
|
//
|
|
|
|
// * Redistribution's in binary form must reproduce the above copyright notice,
|
|
|
|
// this list of conditions and the following disclaimer in the documentation
|
|
|
|
// and/or other oclMaterials provided with the distribution.
|
|
|
|
//
|
|
|
|
// * The name of the copyright holders may not be used to endorse or promote products
|
|
|
|
// derived from this software without specific prior written permission.
|
|
|
|
//
|
|
|
|
// This software is provided by the copyright holders and contributors as is and
|
|
|
|
// any express or implied warranties, including, but not limited to, the implied
|
|
|
|
// warranties of merchantability and fitness for a particular purpose are disclaimed.
|
|
|
|
// In no event shall the Intel Corporation or contributors be liable for any direct,
|
|
|
|
// indirect, incidental, special, exemplary, or consequential damages
|
|
|
|
// (including, but not limited to, procurement of substitute goods or services;
|
|
|
|
// loss of use, data, or profits; or business interruption) however caused
|
|
|
|
// and on any theory of liability, whether in contract, strict liability,
|
|
|
|
// or tort (including negligence or otherwise) arising in any way out of
|
|
|
|
// the use of this software, even if advised of the possibility of such damage.
|
|
|
|
//
|
|
|
|
//M*/
|
|
|
|
|
|
|
|
#if defined (DOUBLE_SUPPORT)
|
2013-04-05 08:15:05 +08:00
|
|
|
#ifdef cl_khr_fp64
|
2012-07-17 01:08:14 +08:00
|
|
|
#pragma OPENCL EXTENSION cl_khr_fp64:enable
|
2013-04-05 08:15:05 +08:00
|
|
|
#elif defined (cl_amd_fp64)
|
|
|
|
#pragma OPENCL EXTENSION cl_amd_fp64:enable
|
|
|
|
#endif
|
2012-07-17 01:08:14 +08:00
|
|
|
#endif
|
|
|
|
|
2013-04-06 13:37:36 +08:00
|
|
|
#ifdef ARITHM_ADD
|
|
|
|
#define ARITHM_OP(A,B) ((A)+(B))
|
|
|
|
#elif defined ARITHM_SUB
|
|
|
|
#define ARITHM_OP(A,B) ((A)-(B))
|
|
|
|
#endif
|
2012-07-17 01:08:14 +08:00
|
|
|
/**************************************add with scalar with mask**************************************/
|
|
|
|
__kernel void arithm_s_add_with_mask_C1_D0 (__global uchar *src1, int src1_step, int src1_offset,
|
|
|
|
__global uchar *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
int4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
x = x << 2;
|
2013-08-21 20:44:09 +08:00
|
|
|
|
2013-04-05 08:15:05 +08:00
|
|
|
#ifdef dst_align
|
|
|
|
#undef dst_align
|
|
|
|
#endif
|
|
|
|
#define dst_align (dst_offset & 3)
|
2012-10-17 07:18:30 +08:00
|
|
|
int src1_index = mad24(y, src1_step, x + src1_offset - dst_align);
|
2012-07-17 01:08:14 +08:00
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset - dst_align);
|
|
|
|
|
|
|
|
int dst_start = mad24(y, dst_step, dst_offset);
|
|
|
|
int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
|
|
|
|
int dst_index = mad24(y, dst_step, dst_offset + x & (int)0xfffffffc);
|
2012-10-17 07:18:30 +08:00
|
|
|
int src1_index_fix = src1_index < 0 ? 0 : src1_index;
|
|
|
|
int mask_index_fix = mask_index < 0 ? 0 : mask_index;
|
2012-10-11 16:22:47 +08:00
|
|
|
uchar4 src1_data = vload4(0, src1 + src1_index_fix);
|
2012-07-17 01:08:14 +08:00
|
|
|
int4 src2_data = (int4)(src2.x, src2.x, src2.x, src2.x);
|
2012-10-17 07:18:30 +08:00
|
|
|
uchar4 mask_data = vload4(0, mask + mask_index_fix);
|
|
|
|
if(src1_index < 0)
|
|
|
|
{
|
|
|
|
uchar4 tmp;
|
|
|
|
tmp.xyzw = (src1_index == -2) ? src1_data.zwxy:src1_data.yzwx;
|
|
|
|
src1_data.xyzw = (src1_index == -1) ? src1_data.wxyz:tmp.xyzw;
|
|
|
|
}
|
|
|
|
if(mask_index < 0)
|
|
|
|
{
|
|
|
|
uchar4 tmp;
|
|
|
|
tmp.xyzw = (mask_index == -2) ? mask_data.zwxy:mask_data.yzwx;
|
|
|
|
mask_data.xyzw = (mask_index == -1) ? mask_data.wxyz:tmp.xyzw;
|
|
|
|
}
|
2012-07-17 01:08:14 +08:00
|
|
|
|
|
|
|
uchar4 data = *((__global uchar4 *)(dst + dst_index));
|
2013-04-06 13:37:36 +08:00
|
|
|
int4 tmp = ARITHM_OP(convert_int4_sat(src1_data), src2_data);
|
2012-07-17 01:08:14 +08:00
|
|
|
uchar4 tmp_data = convert_uchar4_sat(tmp);
|
|
|
|
|
|
|
|
data.x = ((mask_data.x) && (dst_index + 0 >= dst_start) && (dst_index + 0 < dst_end)) ? tmp_data.x : data.x;
|
|
|
|
data.y = ((mask_data.y) && (dst_index + 1 >= dst_start) && (dst_index + 1 < dst_end)) ? tmp_data.y : data.y;
|
|
|
|
data.z = ((mask_data.z) && (dst_index + 2 >= dst_start) && (dst_index + 2 < dst_end)) ? tmp_data.z : data.z;
|
|
|
|
data.w = ((mask_data.w) && (dst_index + 3 >= dst_start) && (dst_index + 3 < dst_end)) ? tmp_data.w : data.w;
|
|
|
|
|
|
|
|
*((__global uchar4 *)(dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
__kernel void arithm_s_add_with_mask_C1_D2 (__global ushort *src1, int src1_step, int src1_offset,
|
|
|
|
__global ushort *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
int4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
x = x << 1;
|
2013-08-21 20:44:09 +08:00
|
|
|
|
2013-04-05 08:15:05 +08:00
|
|
|
#ifdef dst_align
|
|
|
|
#undef dst_align
|
|
|
|
#endif
|
2013-05-28 11:12:05 +08:00
|
|
|
#define dst_align ((dst_offset / 2) & 1)
|
2012-10-17 07:18:30 +08:00
|
|
|
int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1));
|
2012-07-17 01:08:14 +08:00
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset - dst_align);
|
|
|
|
|
|
|
|
int dst_start = mad24(y, dst_step, dst_offset);
|
|
|
|
int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
|
|
|
|
int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffffc);
|
|
|
|
|
|
|
|
ushort2 src1_data = vload2(0, (__global ushort *)((__global char *)src1 + src1_index));
|
|
|
|
int2 src2_data = (int2)(src2.x, src2.x);
|
|
|
|
uchar2 mask_data = vload2(0, mask + mask_index);
|
|
|
|
|
|
|
|
ushort2 data = *((__global ushort2 *)((__global uchar *)dst + dst_index));
|
2013-04-06 13:37:36 +08:00
|
|
|
int2 tmp = ARITHM_OP(convert_int2_sat(src1_data), src2_data);
|
2012-07-17 01:08:14 +08:00
|
|
|
ushort2 tmp_data = convert_ushort2_sat(tmp);
|
|
|
|
|
|
|
|
data.x = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.x : data.x;
|
|
|
|
data.y = ((mask_data.y) && (dst_index + 2 < dst_end )) ? tmp_data.y : data.y;
|
|
|
|
|
|
|
|
*((__global ushort2 *)((__global uchar *)dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
__kernel void arithm_s_add_with_mask_C1_D3 (__global short *src1, int src1_step, int src1_offset,
|
|
|
|
__global short *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
int4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
x = x << 1;
|
2013-08-21 20:44:09 +08:00
|
|
|
|
2013-04-05 08:15:05 +08:00
|
|
|
#ifdef dst_align
|
|
|
|
#undef dst_align
|
|
|
|
#endif
|
2013-05-28 11:12:05 +08:00
|
|
|
#define dst_align ((dst_offset / 2) & 1)
|
2012-10-17 07:18:30 +08:00
|
|
|
int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1));
|
2012-07-17 01:08:14 +08:00
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset - dst_align);
|
|
|
|
|
|
|
|
int dst_start = mad24(y, dst_step, dst_offset);
|
|
|
|
int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
|
|
|
|
int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffffc);
|
|
|
|
|
|
|
|
short2 src1_data = vload2(0, (__global short *)((__global char *)src1 + src1_index));
|
|
|
|
int2 src2_data = (int2)(src2.x, src2.x);
|
|
|
|
uchar2 mask_data = vload2(0, mask + mask_index);
|
|
|
|
|
|
|
|
short2 data = *((__global short2 *)((__global uchar *)dst + dst_index));
|
2013-04-06 13:37:36 +08:00
|
|
|
int2 tmp = ARITHM_OP(convert_int2_sat(src1_data), src2_data);
|
2012-07-17 01:08:14 +08:00
|
|
|
short2 tmp_data = convert_short2_sat(tmp);
|
|
|
|
|
|
|
|
data.x = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.x : data.x;
|
|
|
|
data.y = ((mask_data.y) && (dst_index + 2 < dst_end )) ? tmp_data.y : data.y;
|
|
|
|
|
|
|
|
*((__global short2 *)((__global uchar *)dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
__kernel void arithm_s_add_with_mask_C1_D4 (__global int *src1, int src1_step, int src1_offset,
|
|
|
|
__global int *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
int4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
|
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset);
|
|
|
|
int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
|
|
|
|
|
|
|
|
uchar mask_data = *(mask + mask_index);
|
|
|
|
|
|
|
|
int src_data1 = *((__global int *)((__global char *)src1 + src1_index));
|
|
|
|
int src_data2 = src2.x;
|
|
|
|
int dst_data = *((__global int *)((__global char *)dst + dst_index));
|
|
|
|
|
2013-04-06 13:37:36 +08:00
|
|
|
int data = convert_int_sat(ARITHM_OP((long)src_data1, (long)src_data2));
|
2012-10-17 07:18:30 +08:00
|
|
|
data = mask_data ? data : dst_data;
|
2012-07-17 01:08:14 +08:00
|
|
|
|
|
|
|
*((__global int *)((__global char *)dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
__kernel void arithm_s_add_with_mask_C1_D5 (__global float *src1, int src1_step, int src1_offset,
|
|
|
|
__global float *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
float4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
|
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset);
|
|
|
|
int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
|
|
|
|
|
|
|
|
uchar mask_data = *(mask + mask_index);
|
|
|
|
|
|
|
|
float src_data1 = *((__global float *)((__global char *)src1 + src1_index));
|
|
|
|
float src_data2 = src2.x;
|
|
|
|
float dst_data = *((__global float *)((__global char *)dst + dst_index));
|
|
|
|
|
2013-04-06 13:37:36 +08:00
|
|
|
float data = ARITHM_OP(src_data1, src_data2);
|
2012-10-17 07:18:30 +08:00
|
|
|
data = mask_data ? data : dst_data;
|
2012-07-17 01:08:14 +08:00
|
|
|
|
|
|
|
*((__global float *)((__global char *)dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
#if defined (DOUBLE_SUPPORT)
|
|
|
|
__kernel void arithm_s_add_with_mask_C1_D6 (__global double *src1, int src1_step, int src1_offset,
|
|
|
|
__global double *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
double4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
|
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset);
|
|
|
|
int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
|
|
|
|
|
|
|
|
uchar mask_data = *(mask + mask_index);
|
|
|
|
|
|
|
|
double src_data1 = *((__global double *)((__global char *)src1 + src1_index));
|
|
|
|
double src_data2 = src2.x;
|
|
|
|
double dst_data = *((__global double *)((__global char *)dst + dst_index));
|
|
|
|
|
2013-04-06 13:37:36 +08:00
|
|
|
double data = ARITHM_OP(src_data1, src_data2);
|
2012-10-17 07:18:30 +08:00
|
|
|
data = mask_data ? data : dst_data;
|
2012-07-17 01:08:14 +08:00
|
|
|
|
|
|
|
*((__global double *)((__global char *)dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
__kernel void arithm_s_add_with_mask_C2_D0 (__global uchar *src1, int src1_step, int src1_offset,
|
|
|
|
__global uchar *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
int4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
x = x << 1;
|
2013-08-21 20:44:09 +08:00
|
|
|
|
2013-04-05 08:15:05 +08:00
|
|
|
#ifdef dst_align
|
|
|
|
#undef dst_align
|
|
|
|
#endif
|
2013-05-28 11:12:05 +08:00
|
|
|
#define dst_align ((dst_offset / 2) & 1)
|
2012-10-17 07:18:30 +08:00
|
|
|
int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1));
|
2012-07-17 01:08:14 +08:00
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset - dst_align);
|
|
|
|
|
|
|
|
int dst_start = mad24(y, dst_step, dst_offset);
|
|
|
|
int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
|
|
|
|
int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffffc);
|
|
|
|
|
|
|
|
uchar4 src1_data = vload4(0, src1 + src1_index);
|
|
|
|
int4 src2_data = (int4)(src2.x, src2.y, src2.x, src2.y);
|
|
|
|
uchar2 mask_data = vload2(0, mask + mask_index);
|
|
|
|
|
|
|
|
uchar4 data = *((__global uchar4 *)(dst + dst_index));
|
2013-04-06 13:37:36 +08:00
|
|
|
int4 tmp = ARITHM_OP(convert_int4_sat(src1_data), src2_data);
|
2012-07-17 01:08:14 +08:00
|
|
|
uchar4 tmp_data = convert_uchar4_sat(tmp);
|
|
|
|
|
|
|
|
data.xy = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.xy : data.xy;
|
|
|
|
data.zw = ((mask_data.y) && (dst_index + 2 < dst_end )) ? tmp_data.zw : data.zw;
|
|
|
|
|
|
|
|
*((__global uchar4 *)(dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
__kernel void arithm_s_add_with_mask_C2_D2 (__global ushort *src1, int src1_step, int src1_offset,
|
|
|
|
__global ushort *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
int4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
|
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset);
|
|
|
|
int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
|
|
|
|
|
|
|
|
uchar mask_data = *(mask + mask_index);
|
|
|
|
|
|
|
|
ushort2 src_data1 = *((__global ushort2 *)((__global char *)src1 + src1_index));
|
2012-10-17 07:18:30 +08:00
|
|
|
int2 src_data2 = (int2)(src2.x, src2.y);
|
2012-07-17 01:08:14 +08:00
|
|
|
ushort2 dst_data = *((__global ushort2 *)((__global char *)dst + dst_index));
|
|
|
|
|
2013-04-06 13:37:36 +08:00
|
|
|
int2 tmp = ARITHM_OP(convert_int2_sat(src_data1), src_data2);
|
2012-07-17 01:08:14 +08:00
|
|
|
ushort2 data = convert_ushort2_sat(tmp);
|
2012-10-17 07:18:30 +08:00
|
|
|
data = mask_data ? data : dst_data;
|
2012-07-17 01:08:14 +08:00
|
|
|
|
|
|
|
*((__global ushort2 *)((__global char *)dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
__kernel void arithm_s_add_with_mask_C2_D3 (__global short *src1, int src1_step, int src1_offset,
|
|
|
|
__global short *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
int4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
|
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset);
|
|
|
|
int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
|
|
|
|
|
|
|
|
uchar mask_data = *(mask + mask_index);
|
|
|
|
|
|
|
|
short2 src_data1 = *((__global short2 *)((__global char *)src1 + src1_index));
|
2012-10-17 07:18:30 +08:00
|
|
|
int2 src_data2 = (int2)(src2.x, src2.y);
|
2012-07-17 01:08:14 +08:00
|
|
|
short2 dst_data = *((__global short2 *)((__global char *)dst + dst_index));
|
|
|
|
|
2013-04-06 13:37:36 +08:00
|
|
|
int2 tmp = ARITHM_OP(convert_int2_sat(src_data1), src_data2);
|
2012-07-17 01:08:14 +08:00
|
|
|
short2 data = convert_short2_sat(tmp);
|
2012-10-17 07:18:30 +08:00
|
|
|
data = mask_data ? data : dst_data;
|
2012-07-17 01:08:14 +08:00
|
|
|
|
|
|
|
*((__global short2 *)((__global char *)dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
__kernel void arithm_s_add_with_mask_C2_D4 (__global int *src1, int src1_step, int src1_offset,
|
|
|
|
__global int *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
int4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
|
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset);
|
|
|
|
int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
|
|
|
|
|
|
|
|
uchar mask_data = *(mask + mask_index);
|
|
|
|
|
|
|
|
int2 src_data1 = *((__global int2 *)((__global char *)src1 + src1_index));
|
2012-10-17 07:18:30 +08:00
|
|
|
int2 src_data2 = (int2)(src2.x, src2.y);
|
2012-07-17 01:08:14 +08:00
|
|
|
int2 dst_data = *((__global int2 *)((__global char *)dst + dst_index));
|
|
|
|
|
2013-04-06 13:37:36 +08:00
|
|
|
int2 data = convert_int2_sat(ARITHM_OP(convert_long2_sat(src_data1), convert_long2_sat(src_data2)));
|
2012-10-17 07:18:30 +08:00
|
|
|
data = mask_data ? data : dst_data;
|
2012-07-17 01:08:14 +08:00
|
|
|
|
|
|
|
*((__global int2 *)((__global char *)dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
__kernel void arithm_s_add_with_mask_C2_D5 (__global float *src1, int src1_step, int src1_offset,
|
|
|
|
__global float *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
float4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
|
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset);
|
|
|
|
int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
|
|
|
|
|
|
|
|
uchar mask_data = *(mask + mask_index);
|
|
|
|
|
|
|
|
float2 src_data1 = *((__global float2 *)((__global char *)src1 + src1_index));
|
2012-10-17 07:18:30 +08:00
|
|
|
float2 src_data2 = (float2)(src2.x, src2.y);
|
2012-07-17 01:08:14 +08:00
|
|
|
float2 dst_data = *((__global float2 *)((__global char *)dst + dst_index));
|
|
|
|
|
2013-04-06 13:37:36 +08:00
|
|
|
float2 data = ARITHM_OP(src_data1, src_data2);
|
2012-10-17 07:18:30 +08:00
|
|
|
data = mask_data ? data : dst_data;
|
2012-07-17 01:08:14 +08:00
|
|
|
|
|
|
|
*((__global float2 *)((__global char *)dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#if defined (DOUBLE_SUPPORT)
|
|
|
|
__kernel void arithm_s_add_with_mask_C2_D6 (__global double *src1, int src1_step, int src1_offset,
|
|
|
|
__global double *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
double4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
int src1_index = mad24(y, src1_step, (x << 4) + src1_offset);
|
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset);
|
|
|
|
int dst_index = mad24(y, dst_step, (x << 4) + dst_offset);
|
|
|
|
|
|
|
|
uchar mask_data = *(mask + mask_index);
|
|
|
|
|
|
|
|
double2 src_data1 = *((__global double2 *)((__global char *)src1 + src1_index));
|
2012-10-17 07:18:30 +08:00
|
|
|
double2 src_data2 = (double2)(src2.x, src2.y);
|
2012-07-17 01:08:14 +08:00
|
|
|
double2 dst_data = *((__global double2 *)((__global char *)dst + dst_index));
|
|
|
|
|
2013-04-06 13:37:36 +08:00
|
|
|
double2 data = ARITHM_OP(src_data1, src_data2);
|
2012-10-17 07:18:30 +08:00
|
|
|
data = mask_data ? data : dst_data;
|
2012-07-17 01:08:14 +08:00
|
|
|
|
|
|
|
*((__global double2 *)((__global char *)dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
__kernel void arithm_s_add_with_mask_C4_D0 (__global uchar *src1, int src1_step, int src1_offset,
|
|
|
|
__global uchar *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
int4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
|
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset);
|
|
|
|
int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
|
|
|
|
|
|
|
|
uchar mask_data = *(mask + mask_index);
|
|
|
|
|
|
|
|
uchar4 src_data1 = *((__global uchar4 *)(src1 + src1_index));
|
|
|
|
uchar4 dst_data = *((__global uchar4 *)(dst + dst_index));
|
|
|
|
|
2013-04-06 13:37:36 +08:00
|
|
|
uchar4 data = convert_uchar4_sat(ARITHM_OP(convert_int4_sat(src_data1), src2));
|
2012-10-17 07:18:30 +08:00
|
|
|
data = mask_data ? data : dst_data;
|
2012-07-17 01:08:14 +08:00
|
|
|
|
|
|
|
*((__global uchar4 *)(dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
__kernel void arithm_s_add_with_mask_C4_D2 (__global ushort *src1, int src1_step, int src1_offset,
|
|
|
|
__global ushort *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
int4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
|
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset);
|
|
|
|
int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
|
|
|
|
|
|
|
|
uchar mask_data = *(mask + mask_index);
|
|
|
|
|
|
|
|
ushort4 src_data1 = *((__global ushort4 *)((__global char *)src1 + src1_index));
|
|
|
|
ushort4 dst_data = *((__global ushort4 *)((__global char *)dst + dst_index));
|
|
|
|
|
2013-04-06 13:37:36 +08:00
|
|
|
ushort4 data = convert_ushort4_sat(ARITHM_OP(convert_int4_sat(src_data1), src2));
|
2012-10-17 07:18:30 +08:00
|
|
|
data = mask_data ? data : dst_data;
|
2012-07-17 01:08:14 +08:00
|
|
|
|
|
|
|
*((__global ushort4 *)((__global char *)dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
__kernel void arithm_s_add_with_mask_C4_D3 (__global short *src1, int src1_step, int src1_offset,
|
|
|
|
__global short *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
int4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
|
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset);
|
|
|
|
int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
|
|
|
|
|
|
|
|
uchar mask_data = *(mask + mask_index);
|
|
|
|
|
|
|
|
short4 src_data1 = *((__global short4 *)((__global char *)src1 + src1_index));
|
|
|
|
short4 dst_data = *((__global short4 *)((__global char *)dst + dst_index));
|
|
|
|
|
2013-04-06 13:37:36 +08:00
|
|
|
short4 data = convert_short4_sat(ARITHM_OP(convert_int4_sat(src_data1), src2));
|
2012-10-17 07:18:30 +08:00
|
|
|
data = mask_data ? data : dst_data;
|
2012-07-17 01:08:14 +08:00
|
|
|
|
|
|
|
*((__global short4 *)((__global char *)dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
__kernel void arithm_s_add_with_mask_C4_D4 (__global int *src1, int src1_step, int src1_offset,
|
|
|
|
__global int *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
int4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
int src1_index = mad24(y, src1_step, (x << 4) + src1_offset);
|
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset);
|
|
|
|
int dst_index = mad24(y, dst_step, (x << 4) + dst_offset);
|
|
|
|
|
|
|
|
uchar mask_data = *(mask + mask_index);
|
|
|
|
|
|
|
|
int4 src_data1 = *((__global int4 *)((__global char *)src1 + src1_index));
|
|
|
|
int4 dst_data = *((__global int4 *)((__global char *)dst + dst_index));
|
|
|
|
|
2013-04-06 13:37:36 +08:00
|
|
|
int4 data = convert_int4_sat(ARITHM_OP(convert_long4_sat(src_data1), convert_long4_sat(src2)));
|
2012-10-17 07:18:30 +08:00
|
|
|
data = mask_data ? data : dst_data;
|
2012-07-17 01:08:14 +08:00
|
|
|
|
|
|
|
*((__global int4 *)((__global char *)dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
__kernel void arithm_s_add_with_mask_C4_D5 (__global float *src1, int src1_step, int src1_offset,
|
|
|
|
__global float *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
float4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
int src1_index = mad24(y, src1_step, (x << 4) + src1_offset);
|
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset);
|
|
|
|
int dst_index = mad24(y, dst_step, (x << 4) + dst_offset);
|
|
|
|
|
|
|
|
uchar mask_data = *(mask + mask_index);
|
|
|
|
|
|
|
|
float4 src_data1 = *((__global float4 *)((__global char *)src1 + src1_index));
|
|
|
|
float4 dst_data = *((__global float4 *)((__global char *)dst + dst_index));
|
|
|
|
|
2013-04-06 13:37:36 +08:00
|
|
|
float4 data = ARITHM_OP(src_data1, src2);
|
2012-10-17 07:18:30 +08:00
|
|
|
data = mask_data ? data : dst_data;
|
2012-07-17 01:08:14 +08:00
|
|
|
|
|
|
|
*((__global float4 *)((__global char *)dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#if defined (DOUBLE_SUPPORT)
|
|
|
|
__kernel void arithm_s_add_with_mask_C4_D6 (__global double *src1, int src1_step, int src1_offset,
|
|
|
|
__global double *dst, int dst_step, int dst_offset,
|
|
|
|
__global uchar *mask, int mask_step, int mask_offset,
|
|
|
|
double4 src2, int rows, int cols, int dst_step1)
|
|
|
|
{
|
|
|
|
|
|
|
|
int x = get_global_id(0);
|
|
|
|
int y = get_global_id(1);
|
|
|
|
|
|
|
|
if (x < cols && y < rows)
|
|
|
|
{
|
|
|
|
int src1_index = mad24(y, src1_step, (x << 5) + src1_offset);
|
|
|
|
int mask_index = mad24(y, mask_step, x + mask_offset);
|
|
|
|
int dst_index = mad24(y, dst_step, (x << 5) + dst_offset);
|
|
|
|
|
|
|
|
uchar mask_data = *(mask + mask_index);
|
|
|
|
|
|
|
|
double4 src_data1 = *((__global double4 *)((__global char *)src1 + src1_index));
|
|
|
|
double4 dst_data = *((__global double4 *)((__global char *)dst + dst_index));
|
|
|
|
|
2013-04-06 13:37:36 +08:00
|
|
|
double4 data = ARITHM_OP(src_data1, src2);
|
2012-10-17 07:18:30 +08:00
|
|
|
data = mask_data ? data : dst_data;
|
2012-07-17 01:08:14 +08:00
|
|
|
|
|
|
|
*((__global double4 *)((__global char *)dst + dst_index)) = data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif
|