mirror of
https://github.com/opencv/opencv.git
synced 2024-12-16 02:19:12 +08:00
3053 lines
103 KiB
C++
3053 lines
103 KiB
C++
/*M///////////////////////////////////////////////////////////////////////////////////////
|
|
//
|
|
// IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING.
|
|
//
|
|
// By downloading, copying, installing or using the software you agree to this license.
|
|
// If you do not agree to this license, do not download, install,
|
|
// copy or use the software.
|
|
//
|
|
//
|
|
// Intel License Agreement
|
|
// For Open Source Computer Vision Library
|
|
//
|
|
// Copyright (C) 2000, Intel Corporation, all rights reserved.
|
|
// Third party copyrights are property of their respective owners.
|
|
//
|
|
// Redistribution and use in source and binary forms, with or without modification,
|
|
// are permitted provided that the following conditions are met:
|
|
//
|
|
// * Redistribution's of source code must retain the above copyright notice,
|
|
// this list of conditions and the following disclaimer.
|
|
//
|
|
// * Redistribution's in binary form must reproduce the above copyright notice,
|
|
// this list of conditions and the following disclaimer in the documentation
|
|
// and/or other materials provided with the distribution.
|
|
//
|
|
// * The name of Intel Corporation may not be used to endorse or promote products
|
|
// derived from this software without specific prior written permission.
|
|
//
|
|
// This software is provided by the copyright holders and contributors "as is" and
|
|
// any express or implied warranties, including, but not limited to, the implied
|
|
// warranties of merchantability and fitness for a particular purpose are disclaimed.
|
|
// In no event shall the Intel Corporation or contributors be liable for any direct,
|
|
// indirect, incidental, special, exemplary, or consequential damages
|
|
// (including, but not limited to, procurement of substitute goods or services;
|
|
// loss of use, data, or profits; or business interruption) however caused
|
|
// and on any theory of liability, whether in contract, strict liability,
|
|
// or tort (including negligence or otherwise) arising in any way out of
|
|
// the use of this software, even if advised of the possibility of such damage.
|
|
//
|
|
//M*/
|
|
|
|
/*
|
|
* cvhaartraining.cpp
|
|
*
|
|
* training of cascade of boosted classifiers based on haar features
|
|
*/
|
|
|
|
#include "cvhaartraining.h"
|
|
#include "_cvhaartraining.h"
|
|
|
|
#include <cstdio>
|
|
#include <cstdlib>
|
|
#include <cmath>
|
|
#include <climits>
|
|
|
|
#include <highgui.h>
|
|
|
|
#ifdef CV_VERBOSE
|
|
#include <ctime>
|
|
|
|
#ifdef _WIN32
|
|
/* use clock() function insted of time() */
|
|
#define TIME( arg ) (((double) clock()) / CLOCKS_PER_SEC)
|
|
#else
|
|
#define TIME( arg ) (time( arg ))
|
|
#endif /* _WIN32 */
|
|
|
|
#endif /* CV_VERBOSE */
|
|
|
|
#if defined CV_OPENMP && (defined _MSC_VER || defined CV_ICC)
|
|
#define CV_OPENMP 1
|
|
#else
|
|
#undef CV_OPENMP
|
|
#endif
|
|
|
|
typedef struct CvBackgroundData
|
|
{
|
|
int count;
|
|
char** filename;
|
|
int last;
|
|
int round;
|
|
CvSize winsize;
|
|
} CvBackgroundData;
|
|
|
|
typedef struct CvBackgroundReader
|
|
{
|
|
CvMat src;
|
|
CvMat img;
|
|
CvPoint offset;
|
|
float scale;
|
|
float scalefactor;
|
|
float stepfactor;
|
|
CvPoint point;
|
|
} CvBackgroundReader;
|
|
|
|
/*
|
|
* Background reader
|
|
* Created in each thread
|
|
*/
|
|
CvBackgroundReader* cvbgreader = NULL;
|
|
|
|
#if defined CV_OPENMP
|
|
#pragma omp threadprivate(cvbgreader)
|
|
#endif
|
|
|
|
CvBackgroundData* cvbgdata = NULL;
|
|
|
|
|
|
/*
|
|
* get sum image offsets for <rect> corner points
|
|
* step - row step (measured in image pixels!) of sum image
|
|
*/
|
|
#define CV_SUM_OFFSETS( p0, p1, p2, p3, rect, step ) \
|
|
/* (x, y) */ \
|
|
(p0) = (rect).x + (step) * (rect).y; \
|
|
/* (x + w, y) */ \
|
|
(p1) = (rect).x + (rect).width + (step) * (rect).y; \
|
|
/* (x + w, y) */ \
|
|
(p2) = (rect).x + (step) * ((rect).y + (rect).height); \
|
|
/* (x + w, y + h) */ \
|
|
(p3) = (rect).x + (rect).width + (step) * ((rect).y + (rect).height);
|
|
|
|
/*
|
|
* get tilted image offsets for <rect> corner points
|
|
* step - row step (measured in image pixels!) of tilted image
|
|
*/
|
|
#define CV_TILTED_OFFSETS( p0, p1, p2, p3, rect, step ) \
|
|
/* (x, y) */ \
|
|
(p0) = (rect).x + (step) * (rect).y; \
|
|
/* (x - h, y + h) */ \
|
|
(p1) = (rect).x - (rect).height + (step) * ((rect).y + (rect).height);\
|
|
/* (x + w, y + w) */ \
|
|
(p2) = (rect).x + (rect).width + (step) * ((rect).y + (rect).width); \
|
|
/* (x + w - h, y + w + h) */ \
|
|
(p3) = (rect).x + (rect).width - (rect).height \
|
|
+ (step) * ((rect).y + (rect).width + (rect).height);
|
|
|
|
|
|
/*
|
|
* icvCreateIntHaarFeatures
|
|
*
|
|
* Create internal representation of haar features
|
|
*
|
|
* mode:
|
|
* 0 - BASIC = Viola
|
|
* 1 - CORE = All upright
|
|
* 2 - ALL = All features
|
|
*/
|
|
static
|
|
CvIntHaarFeatures* icvCreateIntHaarFeatures( CvSize winsize,
|
|
int mode,
|
|
int symmetric )
|
|
{
|
|
CvIntHaarFeatures* features = NULL;
|
|
CvTHaarFeature haarFeature;
|
|
|
|
CvMemStorage* storage = NULL;
|
|
CvSeq* seq = NULL;
|
|
CvSeqWriter writer;
|
|
|
|
int s0 = 36; /* minimum total area size of basic haar feature */
|
|
int s1 = 12; /* minimum total area size of tilted haar features 2 */
|
|
int s2 = 18; /* minimum total area size of tilted haar features 3 */
|
|
int s3 = 24; /* minimum total area size of tilted haar features 4 */
|
|
|
|
int x = 0;
|
|
int y = 0;
|
|
int dx = 0;
|
|
int dy = 0;
|
|
|
|
float factor = 1.0F;
|
|
|
|
factor = ((float) winsize.width) * winsize.height / (24 * 24);
|
|
#if 0
|
|
s0 = (int) (s0 * factor);
|
|
s1 = (int) (s1 * factor);
|
|
s2 = (int) (s2 * factor);
|
|
s3 = (int) (s3 * factor);
|
|
#else
|
|
s0 = 1;
|
|
s1 = 1;
|
|
s2 = 1;
|
|
s3 = 1;
|
|
#endif
|
|
|
|
/* CV_VECTOR_CREATE( vec, CvIntHaarFeature, size, maxsize ) */
|
|
storage = cvCreateMemStorage();
|
|
cvStartWriteSeq( 0, sizeof( CvSeq ), sizeof( haarFeature ), storage, &writer );
|
|
|
|
for( x = 0; x < winsize.width; x++ )
|
|
{
|
|
for( y = 0; y < winsize.height; y++ )
|
|
{
|
|
for( dx = 1; dx <= winsize.width; dx++ )
|
|
{
|
|
for( dy = 1; dy <= winsize.height; dy++ )
|
|
{
|
|
// haar_x2
|
|
if ( (x+dx*2 <= winsize.width) && (y+dy <= winsize.height) ) {
|
|
if (dx*2*dy < s0) continue;
|
|
if (!symmetric || (x+x+dx*2 <=winsize.width)) {
|
|
haarFeature = cvHaarFeature( "haar_x2",
|
|
x, y, dx*2, dy, -1,
|
|
x+dx, y, dx , dy, +2 );
|
|
/* CV_VECTOR_PUSH( vec, CvIntHaarFeature, haarFeature, size, maxsize, step ) */
|
|
CV_WRITE_SEQ_ELEM( haarFeature, writer );
|
|
}
|
|
}
|
|
|
|
// haar_y2
|
|
if ( (x+dx <= winsize.width) && (y+dy*2 <= winsize.height) ) {
|
|
if (dx*2*dy < s0) continue;
|
|
if (!symmetric || (x+x+dx <= winsize.width)) {
|
|
haarFeature = cvHaarFeature( "haar_y2",
|
|
x, y, dx, dy*2, -1,
|
|
x, y+dy, dx, dy, +2 );
|
|
CV_WRITE_SEQ_ELEM( haarFeature, writer );
|
|
}
|
|
}
|
|
|
|
// haar_x3
|
|
if ( (x+dx*3 <= winsize.width) && (y+dy <= winsize.height) ) {
|
|
if (dx*3*dy < s0) continue;
|
|
if (!symmetric || (x+x+dx*3 <=winsize.width)) {
|
|
haarFeature = cvHaarFeature( "haar_x3",
|
|
x, y, dx*3, dy, -1,
|
|
x+dx, y, dx, dy, +3 );
|
|
CV_WRITE_SEQ_ELEM( haarFeature, writer );
|
|
}
|
|
}
|
|
|
|
// haar_y3
|
|
if ( (x+dx <= winsize.width) && (y+dy*3 <= winsize.height) ) {
|
|
if (dx*3*dy < s0) continue;
|
|
if (!symmetric || (x+x+dx <= winsize.width)) {
|
|
haarFeature = cvHaarFeature( "haar_y3",
|
|
x, y, dx, dy*3, -1,
|
|
x, y+dy, dx, dy, +3 );
|
|
CV_WRITE_SEQ_ELEM( haarFeature, writer );
|
|
}
|
|
}
|
|
|
|
if( mode != 0 /*BASIC*/ ) {
|
|
// haar_x4
|
|
if ( (x+dx*4 <= winsize.width) && (y+dy <= winsize.height) ) {
|
|
if (dx*4*dy < s0) continue;
|
|
if (!symmetric || (x+x+dx*4 <=winsize.width)) {
|
|
haarFeature = cvHaarFeature( "haar_x4",
|
|
x, y, dx*4, dy, -1,
|
|
x+dx, y, dx*2, dy, +2 );
|
|
CV_WRITE_SEQ_ELEM( haarFeature, writer );
|
|
}
|
|
}
|
|
|
|
// haar_y4
|
|
if ( (x+dx <= winsize.width ) && (y+dy*4 <= winsize.height) ) {
|
|
if (dx*4*dy < s0) continue;
|
|
if (!symmetric || (x+x+dx <=winsize.width)) {
|
|
haarFeature = cvHaarFeature( "haar_y4",
|
|
x, y, dx, dy*4, -1,
|
|
x, y+dy, dx, dy*2, +2 );
|
|
CV_WRITE_SEQ_ELEM( haarFeature, writer );
|
|
}
|
|
}
|
|
}
|
|
|
|
// x2_y2
|
|
if ( (x+dx*2 <= winsize.width) && (y+dy*2 <= winsize.height) ) {
|
|
if (dx*4*dy < s0) continue;
|
|
if (!symmetric || (x+x+dx*2 <=winsize.width)) {
|
|
haarFeature = cvHaarFeature( "haar_x2_y2",
|
|
x , y, dx*2, dy*2, -1,
|
|
x , y , dx , dy, +2,
|
|
x+dx, y+dy, dx , dy, +2 );
|
|
CV_WRITE_SEQ_ELEM( haarFeature, writer );
|
|
}
|
|
}
|
|
|
|
if (mode != 0 /*BASIC*/) {
|
|
// point
|
|
if ( (x+dx*3 <= winsize.width) && (y+dy*3 <= winsize.height) ) {
|
|
if (dx*9*dy < s0) continue;
|
|
if (!symmetric || (x+x+dx*3 <=winsize.width)) {
|
|
haarFeature = cvHaarFeature( "haar_point",
|
|
x , y, dx*3, dy*3, -1,
|
|
x+dx, y+dy, dx , dy , +9);
|
|
CV_WRITE_SEQ_ELEM( haarFeature, writer );
|
|
}
|
|
}
|
|
}
|
|
|
|
if (mode == 2 /*ALL*/) {
|
|
// tilted haar_x2 (x, y, w, h, b, weight)
|
|
if ( (x+2*dx <= winsize.width) && (y+2*dx+dy <= winsize.height) && (x-dy>= 0) ) {
|
|
if (dx*2*dy < s1) continue;
|
|
|
|
if (!symmetric || (x <= (winsize.width / 2) )) {
|
|
haarFeature = cvHaarFeature( "tilted_haar_x2",
|
|
x, y, dx*2, dy, -1,
|
|
x, y, dx , dy, +2 );
|
|
CV_WRITE_SEQ_ELEM( haarFeature, writer );
|
|
}
|
|
}
|
|
|
|
// tilted haar_y2 (x, y, w, h, b, weight)
|
|
if ( (x+dx <= winsize.width) && (y+dx+2*dy <= winsize.height) && (x-2*dy>= 0) ) {
|
|
if (dx*2*dy < s1) continue;
|
|
|
|
if (!symmetric || (x <= (winsize.width / 2) )) {
|
|
haarFeature = cvHaarFeature( "tilted_haar_y2",
|
|
x, y, dx, 2*dy, -1,
|
|
x, y, dx, dy, +2 );
|
|
CV_WRITE_SEQ_ELEM( haarFeature, writer );
|
|
}
|
|
}
|
|
|
|
// tilted haar_x3 (x, y, w, h, b, weight)
|
|
if ( (x+3*dx <= winsize.width) && (y+3*dx+dy <= winsize.height) && (x-dy>= 0) ) {
|
|
if (dx*3*dy < s2) continue;
|
|
|
|
if (!symmetric || (x <= (winsize.width / 2) )) {
|
|
haarFeature = cvHaarFeature( "tilted_haar_x3",
|
|
x, y, dx*3, dy, -1,
|
|
x+dx, y+dx, dx , dy, +3 );
|
|
CV_WRITE_SEQ_ELEM( haarFeature, writer );
|
|
}
|
|
}
|
|
|
|
// tilted haar_y3 (x, y, w, h, b, weight)
|
|
if ( (x+dx <= winsize.width) && (y+dx+3*dy <= winsize.height) && (x-3*dy>= 0) ) {
|
|
if (dx*3*dy < s2) continue;
|
|
|
|
if (!symmetric || (x <= (winsize.width / 2) )) {
|
|
haarFeature = cvHaarFeature( "tilted_haar_y3",
|
|
x, y, dx, 3*dy, -1,
|
|
x-dy, y+dy, dx, dy, +3 );
|
|
CV_WRITE_SEQ_ELEM( haarFeature, writer );
|
|
}
|
|
}
|
|
|
|
|
|
// tilted haar_x4 (x, y, w, h, b, weight)
|
|
if ( (x+4*dx <= winsize.width) && (y+4*dx+dy <= winsize.height) && (x-dy>= 0) ) {
|
|
if (dx*4*dy < s3) continue;
|
|
|
|
if (!symmetric || (x <= (winsize.width / 2) )) {
|
|
haarFeature = cvHaarFeature( "tilted_haar_x4",
|
|
|
|
|
|
x, y, dx*4, dy, -1,
|
|
x+dx, y+dx, dx*2, dy, +2 );
|
|
CV_WRITE_SEQ_ELEM( haarFeature, writer );
|
|
}
|
|
}
|
|
|
|
// tilted haar_y4 (x, y, w, h, b, weight)
|
|
if ( (x+dx <= winsize.width) && (y+dx+4*dy <= winsize.height) && (x-4*dy>= 0) ) {
|
|
if (dx*4*dy < s3) continue;
|
|
|
|
if (!symmetric || (x <= (winsize.width / 2) )) {
|
|
haarFeature = cvHaarFeature( "tilted_haar_y4",
|
|
x, y, dx, 4*dy, -1,
|
|
x-dy, y+dy, dx, 2*dy, +2 );
|
|
CV_WRITE_SEQ_ELEM( haarFeature, writer );
|
|
}
|
|
}
|
|
|
|
|
|
/*
|
|
|
|
// tilted point
|
|
if ( (x+dx*3 <= winsize.width - 1) && (y+dy*3 <= winsize.height - 1) && (x-3*dy>= 0)) {
|
|
if (dx*9*dy < 36) continue;
|
|
if (!symmetric || (x <= (winsize.width / 2) )) {
|
|
haarFeature = cvHaarFeature( "tilted_haar_point",
|
|
x, y, dx*3, dy*3, -1,
|
|
x, y+dy, dx , dy, +9 );
|
|
CV_WRITE_SEQ_ELEM( haarFeature, writer );
|
|
}
|
|
}
|
|
*/
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
seq = cvEndWriteSeq( &writer );
|
|
features = (CvIntHaarFeatures*) cvAlloc( sizeof( CvIntHaarFeatures ) +
|
|
( sizeof( CvTHaarFeature ) + sizeof( CvFastHaarFeature ) ) * seq->total );
|
|
features->feature = (CvTHaarFeature*) (features + 1);
|
|
features->fastfeature = (CvFastHaarFeature*) ( features->feature + seq->total );
|
|
features->count = seq->total;
|
|
features->winsize = winsize;
|
|
cvCvtSeqToArray( seq, (CvArr*) features->feature );
|
|
cvReleaseMemStorage( &storage );
|
|
|
|
icvConvertToFastHaarFeature( features->feature, features->fastfeature,
|
|
features->count, (winsize.width + 1) );
|
|
|
|
return features;
|
|
}
|
|
|
|
static
|
|
void icvReleaseIntHaarFeatures( CvIntHaarFeatures** intHaarFeatures )
|
|
{
|
|
if( intHaarFeatures != NULL && (*intHaarFeatures) != NULL )
|
|
{
|
|
cvFree( intHaarFeatures );
|
|
(*intHaarFeatures) = NULL;
|
|
}
|
|
}
|
|
|
|
|
|
void icvConvertToFastHaarFeature( CvTHaarFeature* haarFeature,
|
|
CvFastHaarFeature* fastHaarFeature,
|
|
int size, int step )
|
|
{
|
|
int i = 0;
|
|
int j = 0;
|
|
|
|
for( i = 0; i < size; i++ )
|
|
{
|
|
fastHaarFeature[i].tilted = haarFeature[i].tilted;
|
|
if( !fastHaarFeature[i].tilted )
|
|
{
|
|
for( j = 0; j < CV_HAAR_FEATURE_MAX; j++ )
|
|
{
|
|
fastHaarFeature[i].rect[j].weight = haarFeature[i].rect[j].weight;
|
|
if( fastHaarFeature[i].rect[j].weight == 0.0F )
|
|
{
|
|
break;
|
|
}
|
|
CV_SUM_OFFSETS( fastHaarFeature[i].rect[j].p0,
|
|
fastHaarFeature[i].rect[j].p1,
|
|
fastHaarFeature[i].rect[j].p2,
|
|
fastHaarFeature[i].rect[j].p3,
|
|
haarFeature[i].rect[j].r, step )
|
|
}
|
|
|
|
}
|
|
else
|
|
{
|
|
for( j = 0; j < CV_HAAR_FEATURE_MAX; j++ )
|
|
{
|
|
fastHaarFeature[i].rect[j].weight = haarFeature[i].rect[j].weight;
|
|
if( fastHaarFeature[i].rect[j].weight == 0.0F )
|
|
{
|
|
break;
|
|
}
|
|
CV_TILTED_OFFSETS( fastHaarFeature[i].rect[j].p0,
|
|
fastHaarFeature[i].rect[j].p1,
|
|
fastHaarFeature[i].rect[j].p2,
|
|
fastHaarFeature[i].rect[j].p3,
|
|
haarFeature[i].rect[j].r, step )
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
/*
|
|
* icvCreateHaarTrainingData
|
|
*
|
|
* Create haar training data used in stage training
|
|
*/
|
|
static
|
|
CvHaarTrainigData* icvCreateHaarTrainingData( CvSize winsize, int maxnumsamples )
|
|
{
|
|
CvHaarTrainigData* data;
|
|
|
|
CV_FUNCNAME( "icvCreateHaarTrainingData" );
|
|
|
|
__BEGIN__;
|
|
|
|
data = NULL;
|
|
uchar* ptr = NULL;
|
|
size_t datasize = 0;
|
|
|
|
datasize = sizeof( CvHaarTrainigData ) +
|
|
/* sum and tilted */
|
|
( 2 * (winsize.width + 1) * (winsize.height + 1) * sizeof( sum_type ) +
|
|
sizeof( float ) + /* normfactor */
|
|
sizeof( float ) + /* cls */
|
|
sizeof( float ) /* weight */
|
|
) * maxnumsamples;
|
|
|
|
CV_CALL( data = (CvHaarTrainigData*) cvAlloc( datasize ) );
|
|
memset( (void*)data, 0, datasize );
|
|
data->maxnum = maxnumsamples;
|
|
data->winsize = winsize;
|
|
ptr = (uchar*)(data + 1);
|
|
data->sum = cvMat( maxnumsamples, (winsize.width + 1) * (winsize.height + 1),
|
|
CV_SUM_MAT_TYPE, (void*) ptr );
|
|
ptr += sizeof( sum_type ) * maxnumsamples * (winsize.width+1) * (winsize.height+1);
|
|
data->tilted = cvMat( maxnumsamples, (winsize.width + 1) * (winsize.height + 1),
|
|
CV_SUM_MAT_TYPE, (void*) ptr );
|
|
ptr += sizeof( sum_type ) * maxnumsamples * (winsize.width+1) * (winsize.height+1);
|
|
data->normfactor = cvMat( 1, maxnumsamples, CV_32FC1, (void*) ptr );
|
|
ptr += sizeof( float ) * maxnumsamples;
|
|
data->cls = cvMat( 1, maxnumsamples, CV_32FC1, (void*) ptr );
|
|
ptr += sizeof( float ) * maxnumsamples;
|
|
data->weights = cvMat( 1, maxnumsamples, CV_32FC1, (void*) ptr );
|
|
|
|
data->valcache = NULL;
|
|
data->idxcache = NULL;
|
|
|
|
__END__;
|
|
|
|
return data;
|
|
}
|
|
|
|
static
|
|
void icvReleaseHaarTrainingDataCache( CvHaarTrainigData** haarTrainingData )
|
|
{
|
|
if( haarTrainingData != NULL && (*haarTrainingData) != NULL )
|
|
{
|
|
if( (*haarTrainingData)->valcache != NULL )
|
|
{
|
|
cvReleaseMat( &(*haarTrainingData)->valcache );
|
|
(*haarTrainingData)->valcache = NULL;
|
|
}
|
|
if( (*haarTrainingData)->idxcache != NULL )
|
|
{
|
|
cvReleaseMat( &(*haarTrainingData)->idxcache );
|
|
(*haarTrainingData)->idxcache = NULL;
|
|
}
|
|
}
|
|
}
|
|
|
|
static
|
|
void icvReleaseHaarTrainingData( CvHaarTrainigData** haarTrainingData )
|
|
{
|
|
if( haarTrainingData != NULL && (*haarTrainingData) != NULL )
|
|
{
|
|
icvReleaseHaarTrainingDataCache( haarTrainingData );
|
|
|
|
cvFree( haarTrainingData );
|
|
}
|
|
}
|
|
|
|
static
|
|
void icvGetTrainingDataCallback( CvMat* mat, CvMat* sampleIdx, CvMat*,
|
|
int first, int num, void* userdata )
|
|
{
|
|
int i = 0;
|
|
int j = 0;
|
|
float val = 0.0F;
|
|
float normfactor = 0.0F;
|
|
|
|
CvHaarTrainingData* training_data;
|
|
CvIntHaarFeatures* haar_features;
|
|
|
|
#ifdef CV_COL_ARRANGEMENT
|
|
assert( mat->rows >= num );
|
|
#else
|
|
assert( mat->cols >= num );
|
|
#endif
|
|
|
|
training_data = ((CvUserdata*) userdata)->trainingData;
|
|
haar_features = ((CvUserdata*) userdata)->haarFeatures;
|
|
if( sampleIdx == NULL )
|
|
{
|
|
int num_samples;
|
|
|
|
#ifdef CV_COL_ARRANGEMENT
|
|
num_samples = mat->cols;
|
|
#else
|
|
num_samples = mat->rows;
|
|
#endif
|
|
for( i = 0; i < num_samples; i++ )
|
|
{
|
|
for( j = 0; j < num; j++ )
|
|
{
|
|
val = cvEvalFastHaarFeature(
|
|
( haar_features->fastfeature
|
|
+ first + j ),
|
|
(sum_type*) (training_data->sum.data.ptr
|
|
+ i * training_data->sum.step),
|
|
(sum_type*) (training_data->tilted.data.ptr
|
|
+ i * training_data->tilted.step) );
|
|
normfactor = training_data->normfactor.data.fl[i];
|
|
val = ( normfactor == 0.0F ) ? 0.0F : (val / normfactor);
|
|
|
|
#ifdef CV_COL_ARRANGEMENT
|
|
CV_MAT_ELEM( *mat, float, j, i ) = val;
|
|
#else
|
|
CV_MAT_ELEM( *mat, float, i, j ) = val;
|
|
#endif
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
uchar* idxdata = NULL;
|
|
size_t step = 0;
|
|
int numidx = 0;
|
|
int idx = 0;
|
|
|
|
assert( CV_MAT_TYPE( sampleIdx->type ) == CV_32FC1 );
|
|
|
|
idxdata = sampleIdx->data.ptr;
|
|
if( sampleIdx->rows == 1 )
|
|
{
|
|
step = sizeof( float );
|
|
numidx = sampleIdx->cols;
|
|
}
|
|
else
|
|
{
|
|
step = sampleIdx->step;
|
|
numidx = sampleIdx->rows;
|
|
}
|
|
|
|
for( i = 0; i < numidx; i++ )
|
|
{
|
|
for( j = 0; j < num; j++ )
|
|
{
|
|
idx = (int)( *((float*) (idxdata + i * step)) );
|
|
val = cvEvalFastHaarFeature(
|
|
( haar_features->fastfeature
|
|
+ first + j ),
|
|
(sum_type*) (training_data->sum.data.ptr
|
|
+ idx * training_data->sum.step),
|
|
(sum_type*) (training_data->tilted.data.ptr
|
|
+ idx * training_data->tilted.step) );
|
|
normfactor = training_data->normfactor.data.fl[idx];
|
|
val = ( normfactor == 0.0F ) ? 0.0F : (val / normfactor);
|
|
|
|
#ifdef CV_COL_ARRANGEMENT
|
|
CV_MAT_ELEM( *mat, float, j, idx ) = val;
|
|
#else
|
|
CV_MAT_ELEM( *mat, float, idx, j ) = val;
|
|
#endif
|
|
|
|
}
|
|
}
|
|
}
|
|
#if 0 /*def CV_VERBOSE*/
|
|
if( first % 5000 == 0 )
|
|
{
|
|
fprintf( stderr, "%3d%%\r", (int) (100.0 * first /
|
|
haar_features->count) );
|
|
fflush( stderr );
|
|
}
|
|
#endif /* CV_VERBOSE */
|
|
}
|
|
|
|
static
|
|
void icvPrecalculate( CvHaarTrainingData* data, CvIntHaarFeatures* haarFeatures,
|
|
int numprecalculated )
|
|
{
|
|
CV_FUNCNAME( "icvPrecalculate" );
|
|
|
|
__BEGIN__;
|
|
|
|
icvReleaseHaarTrainingDataCache( &data );
|
|
|
|
numprecalculated -= numprecalculated % CV_STUMP_TRAIN_PORTION;
|
|
numprecalculated = MIN( numprecalculated, haarFeatures->count );
|
|
|
|
if( numprecalculated > 0 )
|
|
{
|
|
//size_t datasize;
|
|
int m;
|
|
CvUserdata userdata;
|
|
|
|
/* private variables */
|
|
#ifdef CV_OPENMP
|
|
CvMat t_data;
|
|
CvMat t_idx;
|
|
int first;
|
|
int t_portion;
|
|
int portion = CV_STUMP_TRAIN_PORTION;
|
|
#endif /* CV_OPENMP */
|
|
|
|
m = data->sum.rows;
|
|
|
|
#ifdef CV_COL_ARRANGEMENT
|
|
CV_CALL( data->valcache = cvCreateMat( numprecalculated, m, CV_32FC1 ) );
|
|
#else
|
|
CV_CALL( data->valcache = cvCreateMat( m, numprecalculated, CV_32FC1 ) );
|
|
#endif
|
|
CV_CALL( data->idxcache = cvCreateMat( numprecalculated, m, CV_IDX_MAT_TYPE ) );
|
|
|
|
userdata = cvUserdata( data, haarFeatures );
|
|
|
|
#ifdef CV_OPENMP
|
|
#pragma omp parallel for private(t_data, t_idx, first, t_portion)
|
|
for( first = 0; first < numprecalculated; first += portion )
|
|
{
|
|
t_data = *data->valcache;
|
|
t_idx = *data->idxcache;
|
|
t_portion = MIN( portion, (numprecalculated - first) );
|
|
|
|
/* indices */
|
|
t_idx.rows = t_portion;
|
|
t_idx.data.ptr = data->idxcache->data.ptr + first * ((size_t)t_idx.step);
|
|
|
|
/* feature values */
|
|
#ifdef CV_COL_ARRANGEMENT
|
|
t_data.rows = t_portion;
|
|
t_data.data.ptr = data->valcache->data.ptr +
|
|
first * ((size_t) t_data.step );
|
|
#else
|
|
t_data.cols = t_portion;
|
|
t_data.data.ptr = data->valcache->data.ptr +
|
|
first * ((size_t) CV_ELEM_SIZE( t_data.type ));
|
|
#endif
|
|
icvGetTrainingDataCallback( &t_data, NULL, NULL, first, t_portion,
|
|
&userdata );
|
|
#ifdef CV_COL_ARRANGEMENT
|
|
cvGetSortedIndices( &t_data, &t_idx, 0 );
|
|
#else
|
|
cvGetSortedIndices( &t_data, &t_idx, 1 );
|
|
#endif
|
|
|
|
#ifdef CV_VERBOSE
|
|
putc( '.', stderr );
|
|
fflush( stderr );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
}
|
|
|
|
#ifdef CV_VERBOSE
|
|
fprintf( stderr, "\n" );
|
|
fflush( stderr );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
#else
|
|
icvGetTrainingDataCallback( data->valcache, NULL, NULL, 0, numprecalculated,
|
|
&userdata );
|
|
#ifdef CV_COL_ARRANGEMENT
|
|
cvGetSortedIndices( data->valcache, data->idxcache, 0 );
|
|
#else
|
|
cvGetSortedIndices( data->valcache, data->idxcache, 1 );
|
|
#endif
|
|
#endif /* CV_OPENMP */
|
|
}
|
|
|
|
__END__;
|
|
}
|
|
|
|
static
|
|
void icvSplitIndicesCallback( int compidx, float threshold,
|
|
CvMat* idx, CvMat** left, CvMat** right,
|
|
void* userdata )
|
|
{
|
|
CvHaarTrainingData* data;
|
|
CvIntHaarFeatures* haar_features;
|
|
int i;
|
|
int m;
|
|
CvFastHaarFeature* fastfeature;
|
|
|
|
data = ((CvUserdata*) userdata)->trainingData;
|
|
haar_features = ((CvUserdata*) userdata)->haarFeatures;
|
|
fastfeature = &haar_features->fastfeature[compidx];
|
|
|
|
m = data->sum.rows;
|
|
*left = cvCreateMat( 1, m, CV_32FC1 );
|
|
*right = cvCreateMat( 1, m, CV_32FC1 );
|
|
(*left)->cols = (*right)->cols = 0;
|
|
if( idx == NULL )
|
|
{
|
|
for( i = 0; i < m; i++ )
|
|
{
|
|
if( cvEvalFastHaarFeature( fastfeature,
|
|
(sum_type*) (data->sum.data.ptr + i * data->sum.step),
|
|
(sum_type*) (data->tilted.data.ptr + i * data->tilted.step) )
|
|
< threshold * data->normfactor.data.fl[i] )
|
|
{
|
|
(*left)->data.fl[(*left)->cols++] = (float) i;
|
|
}
|
|
else
|
|
{
|
|
(*right)->data.fl[(*right)->cols++] = (float) i;
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
uchar* idxdata;
|
|
int idxnum;
|
|
size_t idxstep;
|
|
int index;
|
|
|
|
idxdata = idx->data.ptr;
|
|
idxnum = (idx->rows == 1) ? idx->cols : idx->rows;
|
|
idxstep = (idx->rows == 1) ? CV_ELEM_SIZE( idx->type ) : idx->step;
|
|
for( i = 0; i < idxnum; i++ )
|
|
{
|
|
index = (int) *((float*) (idxdata + i * idxstep));
|
|
if( cvEvalFastHaarFeature( fastfeature,
|
|
(sum_type*) (data->sum.data.ptr + index * data->sum.step),
|
|
(sum_type*) (data->tilted.data.ptr + index * data->tilted.step) )
|
|
< threshold * data->normfactor.data.fl[index] )
|
|
{
|
|
(*left)->data.fl[(*left)->cols++] = (float) index;
|
|
}
|
|
else
|
|
{
|
|
(*right)->data.fl[(*right)->cols++] = (float) index;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* icvCreateCARTStageClassifier
|
|
*
|
|
* Create stage classifier with trees as weak classifiers
|
|
* data - haar training data. It must be created and filled before call
|
|
* minhitrate - desired min hit rate
|
|
* maxfalsealarm - desired max false alarm rate
|
|
* symmetric - if not 0 it is assumed that samples are vertically symmetric
|
|
* numprecalculated - number of features that will be precalculated. Each precalculated
|
|
* feature need (number_of_samples*(sizeof( float ) + sizeof( short ))) bytes of memory
|
|
* weightfraction - weight trimming parameter
|
|
* numsplits - number of binary splits in each tree
|
|
* boosttype - type of applied boosting algorithm
|
|
* stumperror - type of used error if Discrete AdaBoost algorithm is applied
|
|
* maxsplits - maximum total number of splits in all weak classifiers.
|
|
* If it is not 0 then NULL returned if total number of splits exceeds <maxsplits>.
|
|
*/
|
|
static
|
|
CvIntHaarClassifier* icvCreateCARTStageClassifier( CvHaarTrainingData* data,
|
|
CvMat* sampleIdx,
|
|
CvIntHaarFeatures* haarFeatures,
|
|
float minhitrate,
|
|
float maxfalsealarm,
|
|
int symmetric,
|
|
float weightfraction,
|
|
int numsplits,
|
|
CvBoostType boosttype,
|
|
CvStumpError stumperror,
|
|
int maxsplits )
|
|
{
|
|
|
|
#ifdef CV_COL_ARRANGEMENT
|
|
int flags = CV_COL_SAMPLE;
|
|
#else
|
|
int flags = CV_ROW_SAMPLE;
|
|
#endif
|
|
|
|
CvStageHaarClassifier* stage = NULL;
|
|
CvBoostTrainer* trainer;
|
|
CvCARTClassifier* cart = NULL;
|
|
CvCARTTrainParams trainParams;
|
|
CvMTStumpTrainParams stumpTrainParams;
|
|
//CvMat* trainData = NULL;
|
|
//CvMat* sortedIdx = NULL;
|
|
CvMat eval;
|
|
int n = 0;
|
|
int m = 0;
|
|
int numpos = 0;
|
|
int numneg = 0;
|
|
int numfalse = 0;
|
|
float sum_stage = 0.0F;
|
|
float threshold = 0.0F;
|
|
float falsealarm = 0.0F;
|
|
|
|
//CvMat* sampleIdx = NULL;
|
|
CvMat* trimmedIdx;
|
|
//float* idxdata = NULL;
|
|
//float* tempweights = NULL;
|
|
//int idxcount = 0;
|
|
CvUserdata userdata;
|
|
|
|
int i = 0;
|
|
int j = 0;
|
|
int idx;
|
|
int numsamples;
|
|
int numtrimmed;
|
|
|
|
CvCARTHaarClassifier* classifier;
|
|
CvSeq* seq = NULL;
|
|
CvMemStorage* storage = NULL;
|
|
CvMat* weakTrainVals;
|
|
float alpha;
|
|
float sumalpha;
|
|
int num_splits; /* total number of splits in all weak classifiers */
|
|
|
|
#ifdef CV_VERBOSE
|
|
printf( "+----+----+-+---------+---------+---------+---------+\n" );
|
|
printf( "| N |%%SMP|F| ST.THR | HR | FA | EXP. ERR|\n" );
|
|
printf( "+----+----+-+---------+---------+---------+---------+\n" );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
n = haarFeatures->count;
|
|
m = data->sum.rows;
|
|
numsamples = (sampleIdx) ? MAX( sampleIdx->rows, sampleIdx->cols ) : m;
|
|
|
|
userdata = cvUserdata( data, haarFeatures );
|
|
|
|
stumpTrainParams.type = ( boosttype == CV_DABCLASS )
|
|
? CV_CLASSIFICATION_CLASS : CV_REGRESSION;
|
|
stumpTrainParams.error = ( boosttype == CV_LBCLASS || boosttype == CV_GABCLASS )
|
|
? CV_SQUARE : stumperror;
|
|
stumpTrainParams.portion = CV_STUMP_TRAIN_PORTION;
|
|
stumpTrainParams.getTrainData = icvGetTrainingDataCallback;
|
|
stumpTrainParams.numcomp = n;
|
|
stumpTrainParams.userdata = &userdata;
|
|
stumpTrainParams.sortedIdx = data->idxcache;
|
|
|
|
trainParams.count = numsplits;
|
|
trainParams.stumpTrainParams = (CvClassifierTrainParams*) &stumpTrainParams;
|
|
trainParams.stumpConstructor = cvCreateMTStumpClassifier;
|
|
trainParams.splitIdx = icvSplitIndicesCallback;
|
|
trainParams.userdata = &userdata;
|
|
|
|
eval = cvMat( 1, m, CV_32FC1, cvAlloc( sizeof( float ) * m ) );
|
|
|
|
storage = cvCreateMemStorage();
|
|
seq = cvCreateSeq( 0, sizeof( *seq ), sizeof( classifier ), storage );
|
|
|
|
weakTrainVals = cvCreateMat( 1, m, CV_32FC1 );
|
|
trainer = cvBoostStartTraining( &data->cls, weakTrainVals, &data->weights,
|
|
sampleIdx, boosttype );
|
|
num_splits = 0;
|
|
sumalpha = 0.0F;
|
|
do
|
|
{
|
|
|
|
#ifdef CV_VERBOSE
|
|
int v_wt = 0;
|
|
int v_flipped = 0;
|
|
#endif /* CV_VERBOSE */
|
|
|
|
trimmedIdx = cvTrimWeights( &data->weights, sampleIdx, weightfraction );
|
|
numtrimmed = (trimmedIdx) ? MAX( trimmedIdx->rows, trimmedIdx->cols ) : m;
|
|
|
|
#ifdef CV_VERBOSE
|
|
v_wt = 100 * numtrimmed / numsamples;
|
|
v_flipped = 0;
|
|
|
|
#endif /* CV_VERBOSE */
|
|
|
|
cart = (CvCARTClassifier*) cvCreateCARTClassifier( data->valcache,
|
|
flags,
|
|
weakTrainVals, 0, 0, 0, trimmedIdx,
|
|
&(data->weights),
|
|
(CvClassifierTrainParams*) &trainParams );
|
|
|
|
classifier = (CvCARTHaarClassifier*) icvCreateCARTHaarClassifier( numsplits );
|
|
icvInitCARTHaarClassifier( classifier, cart, haarFeatures );
|
|
|
|
num_splits += classifier->count;
|
|
|
|
cart->release( (CvClassifier**) &cart );
|
|
|
|
if( symmetric && (seq->total % 2) )
|
|
{
|
|
float normfactor = 0.0F;
|
|
CvStumpClassifier* stump;
|
|
|
|
/* flip haar features */
|
|
for( i = 0; i < classifier->count; i++ )
|
|
{
|
|
if( classifier->feature[i].desc[0] == 'h' )
|
|
{
|
|
for( j = 0; j < CV_HAAR_FEATURE_MAX &&
|
|
classifier->feature[i].rect[j].weight != 0.0F; j++ )
|
|
{
|
|
classifier->feature[i].rect[j].r.x = data->winsize.width -
|
|
classifier->feature[i].rect[j].r.x -
|
|
classifier->feature[i].rect[j].r.width;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
int tmp = 0;
|
|
|
|
/* (x,y) -> (24-x,y) */
|
|
/* w -> h; h -> w */
|
|
for( j = 0; j < CV_HAAR_FEATURE_MAX &&
|
|
classifier->feature[i].rect[j].weight != 0.0F; j++ )
|
|
{
|
|
classifier->feature[i].rect[j].r.x = data->winsize.width -
|
|
classifier->feature[i].rect[j].r.x;
|
|
CV_SWAP( classifier->feature[i].rect[j].r.width,
|
|
classifier->feature[i].rect[j].r.height, tmp );
|
|
}
|
|
}
|
|
}
|
|
icvConvertToFastHaarFeature( classifier->feature,
|
|
classifier->fastfeature,
|
|
classifier->count, data->winsize.width + 1 );
|
|
|
|
stumpTrainParams.getTrainData = NULL;
|
|
stumpTrainParams.numcomp = 1;
|
|
stumpTrainParams.userdata = NULL;
|
|
stumpTrainParams.sortedIdx = NULL;
|
|
|
|
for( i = 0; i < classifier->count; i++ )
|
|
{
|
|
for( j = 0; j < numtrimmed; j++ )
|
|
{
|
|
idx = icvGetIdxAt( trimmedIdx, j );
|
|
|
|
eval.data.fl[idx] = cvEvalFastHaarFeature( &classifier->fastfeature[i],
|
|
(sum_type*) (data->sum.data.ptr + idx * data->sum.step),
|
|
(sum_type*) (data->tilted.data.ptr + idx * data->tilted.step) );
|
|
normfactor = data->normfactor.data.fl[idx];
|
|
eval.data.fl[idx] = ( normfactor == 0.0F )
|
|
? 0.0F : (eval.data.fl[idx] / normfactor);
|
|
}
|
|
|
|
stump = (CvStumpClassifier*) trainParams.stumpConstructor( &eval,
|
|
CV_COL_SAMPLE,
|
|
weakTrainVals, 0, 0, 0, trimmedIdx,
|
|
&(data->weights),
|
|
trainParams.stumpTrainParams );
|
|
|
|
classifier->threshold[i] = stump->threshold;
|
|
if( classifier->left[i] <= 0 )
|
|
{
|
|
classifier->val[-classifier->left[i]] = stump->left;
|
|
}
|
|
if( classifier->right[i] <= 0 )
|
|
{
|
|
classifier->val[-classifier->right[i]] = stump->right;
|
|
}
|
|
|
|
stump->release( (CvClassifier**) &stump );
|
|
|
|
}
|
|
|
|
stumpTrainParams.getTrainData = icvGetTrainingDataCallback;
|
|
stumpTrainParams.numcomp = n;
|
|
stumpTrainParams.userdata = &userdata;
|
|
stumpTrainParams.sortedIdx = data->idxcache;
|
|
|
|
#ifdef CV_VERBOSE
|
|
v_flipped = 1;
|
|
#endif /* CV_VERBOSE */
|
|
|
|
} /* if symmetric */
|
|
if( trimmedIdx != sampleIdx )
|
|
{
|
|
cvReleaseMat( &trimmedIdx );
|
|
trimmedIdx = NULL;
|
|
}
|
|
|
|
for( i = 0; i < numsamples; i++ )
|
|
{
|
|
idx = icvGetIdxAt( sampleIdx, i );
|
|
|
|
eval.data.fl[idx] = classifier->eval( (CvIntHaarClassifier*) classifier,
|
|
(sum_type*) (data->sum.data.ptr + idx * data->sum.step),
|
|
(sum_type*) (data->tilted.data.ptr + idx * data->tilted.step),
|
|
data->normfactor.data.fl[idx] );
|
|
}
|
|
|
|
alpha = cvBoostNextWeakClassifier( &eval, &data->cls, weakTrainVals,
|
|
&data->weights, trainer );
|
|
sumalpha += alpha;
|
|
|
|
for( i = 0; i <= classifier->count; i++ )
|
|
{
|
|
if( boosttype == CV_RABCLASS )
|
|
{
|
|
classifier->val[i] = cvLogRatio( classifier->val[i] );
|
|
}
|
|
classifier->val[i] *= alpha;
|
|
}
|
|
|
|
cvSeqPush( seq, (void*) &classifier );
|
|
|
|
numpos = 0;
|
|
for( i = 0; i < numsamples; i++ )
|
|
{
|
|
idx = icvGetIdxAt( sampleIdx, i );
|
|
|
|
if( data->cls.data.fl[idx] == 1.0F )
|
|
{
|
|
eval.data.fl[numpos] = 0.0F;
|
|
for( j = 0; j < seq->total; j++ )
|
|
{
|
|
classifier = *((CvCARTHaarClassifier**) cvGetSeqElem( seq, j ));
|
|
eval.data.fl[numpos] += classifier->eval(
|
|
(CvIntHaarClassifier*) classifier,
|
|
(sum_type*) (data->sum.data.ptr + idx * data->sum.step),
|
|
(sum_type*) (data->tilted.data.ptr + idx * data->tilted.step),
|
|
data->normfactor.data.fl[idx] );
|
|
}
|
|
/* eval.data.fl[numpos] = 2.0F * eval.data.fl[numpos] - seq->total; */
|
|
numpos++;
|
|
}
|
|
}
|
|
icvSort_32f( eval.data.fl, numpos, 0 );
|
|
threshold = eval.data.fl[(int) ((1.0F - minhitrate) * numpos)];
|
|
|
|
numneg = 0;
|
|
numfalse = 0;
|
|
for( i = 0; i < numsamples; i++ )
|
|
{
|
|
idx = icvGetIdxAt( sampleIdx, i );
|
|
|
|
if( data->cls.data.fl[idx] == 0.0F )
|
|
{
|
|
numneg++;
|
|
sum_stage = 0.0F;
|
|
for( j = 0; j < seq->total; j++ )
|
|
{
|
|
classifier = *((CvCARTHaarClassifier**) cvGetSeqElem( seq, j ));
|
|
sum_stage += classifier->eval( (CvIntHaarClassifier*) classifier,
|
|
(sum_type*) (data->sum.data.ptr + idx * data->sum.step),
|
|
(sum_type*) (data->tilted.data.ptr + idx * data->tilted.step),
|
|
data->normfactor.data.fl[idx] );
|
|
}
|
|
/* sum_stage = 2.0F * sum_stage - seq->total; */
|
|
if( sum_stage >= (threshold - CV_THRESHOLD_EPS) )
|
|
{
|
|
numfalse++;
|
|
}
|
|
}
|
|
}
|
|
falsealarm = ((float) numfalse) / ((float) numneg);
|
|
|
|
#ifdef CV_VERBOSE
|
|
{
|
|
float v_hitrate = 0.0F;
|
|
float v_falsealarm = 0.0F;
|
|
/* expected error of stage classifier regardless threshold */
|
|
float v_experr = 0.0F;
|
|
|
|
for( i = 0; i < numsamples; i++ )
|
|
{
|
|
idx = icvGetIdxAt( sampleIdx, i );
|
|
|
|
sum_stage = 0.0F;
|
|
for( j = 0; j < seq->total; j++ )
|
|
{
|
|
classifier = *((CvCARTHaarClassifier**) cvGetSeqElem( seq, j ));
|
|
sum_stage += classifier->eval( (CvIntHaarClassifier*) classifier,
|
|
(sum_type*) (data->sum.data.ptr + idx * data->sum.step),
|
|
(sum_type*) (data->tilted.data.ptr + idx * data->tilted.step),
|
|
data->normfactor.data.fl[idx] );
|
|
}
|
|
/* sum_stage = 2.0F * sum_stage - seq->total; */
|
|
if( sum_stage >= (threshold - CV_THRESHOLD_EPS) )
|
|
{
|
|
if( data->cls.data.fl[idx] == 1.0F )
|
|
{
|
|
v_hitrate += 1.0F;
|
|
}
|
|
else
|
|
{
|
|
v_falsealarm += 1.0F;
|
|
}
|
|
}
|
|
if( ( sum_stage >= 0.0F ) != (data->cls.data.fl[idx] == 1.0F) )
|
|
{
|
|
v_experr += 1.0F;
|
|
}
|
|
}
|
|
v_experr /= numsamples;
|
|
printf( "|%4d|%3d%%|%c|%9f|%9f|%9f|%9f|\n",
|
|
seq->total, v_wt, ( (v_flipped) ? '+' : '-' ),
|
|
threshold, v_hitrate / numpos, v_falsealarm / numneg,
|
|
v_experr );
|
|
printf( "+----+----+-+---------+---------+---------+---------+\n" );
|
|
fflush( stdout );
|
|
}
|
|
#endif /* CV_VERBOSE */
|
|
|
|
} while( falsealarm > maxfalsealarm && (!maxsplits || (num_splits < maxsplits) ) );
|
|
cvBoostEndTraining( &trainer );
|
|
|
|
if( falsealarm > maxfalsealarm )
|
|
{
|
|
stage = NULL;
|
|
}
|
|
else
|
|
{
|
|
stage = (CvStageHaarClassifier*) icvCreateStageHaarClassifier( seq->total,
|
|
threshold );
|
|
cvCvtSeqToArray( seq, (CvArr*) stage->classifier );
|
|
}
|
|
|
|
/* CLEANUP */
|
|
cvReleaseMemStorage( &storage );
|
|
cvReleaseMat( &weakTrainVals );
|
|
cvFree( &(eval.data.ptr) );
|
|
|
|
return (CvIntHaarClassifier*) stage;
|
|
}
|
|
|
|
|
|
static
|
|
CvBackgroundData* icvCreateBackgroundData( const char* filename, CvSize winsize )
|
|
{
|
|
CvBackgroundData* data = NULL;
|
|
|
|
const char* dir = NULL;
|
|
char full[PATH_MAX];
|
|
char* imgfilename = NULL;
|
|
size_t datasize = 0;
|
|
int count = 0;
|
|
FILE* input = NULL;
|
|
char* tmp = NULL;
|
|
int len = 0;
|
|
|
|
assert( filename != NULL );
|
|
|
|
dir = strrchr( filename, '\\' );
|
|
if( dir == NULL )
|
|
{
|
|
dir = strrchr( filename, '/' );
|
|
}
|
|
if( dir == NULL )
|
|
{
|
|
imgfilename = &(full[0]);
|
|
}
|
|
else
|
|
{
|
|
strncpy( &(full[0]), filename, (dir - filename + 1) );
|
|
imgfilename = &(full[(dir - filename + 1)]);
|
|
}
|
|
|
|
input = fopen( filename, "r" );
|
|
if( input != NULL )
|
|
{
|
|
count = 0;
|
|
datasize = 0;
|
|
|
|
/* count */
|
|
while( !feof( input ) )
|
|
{
|
|
*imgfilename = '\0';
|
|
if( !fgets( imgfilename, PATH_MAX - (int)(imgfilename - full) - 1, input ))
|
|
break;
|
|
len = (int)strlen( imgfilename );
|
|
if( len > 0 && imgfilename[len-1] == '\n' )
|
|
imgfilename[len-1] = 0, len--;
|
|
if( len > 0 )
|
|
{
|
|
if( (*imgfilename) == '#' ) continue; /* comment */
|
|
count++;
|
|
datasize += sizeof( char ) * (strlen( &(full[0]) ) + 1);
|
|
}
|
|
}
|
|
if( count > 0 )
|
|
{
|
|
//rewind( input );
|
|
fseek( input, 0, SEEK_SET );
|
|
datasize += sizeof( *data ) + sizeof( char* ) * count;
|
|
data = (CvBackgroundData*) cvAlloc( datasize );
|
|
memset( (void*) data, 0, datasize );
|
|
data->count = count;
|
|
data->filename = (char**) (data + 1);
|
|
data->last = 0;
|
|
data->round = 0;
|
|
data->winsize = winsize;
|
|
tmp = (char*) (data->filename + data->count);
|
|
count = 0;
|
|
while( !feof( input ) )
|
|
{
|
|
*imgfilename = '\0';
|
|
if( !fgets( imgfilename, PATH_MAX - (int)(imgfilename - full) - 1, input ))
|
|
break;
|
|
len = (int)strlen( imgfilename );
|
|
if( len > 0 && imgfilename[len-1] == '\n' )
|
|
imgfilename[len-1] = 0, len--;
|
|
if( len > 0 )
|
|
{
|
|
if( (*imgfilename) == '#' ) continue; /* comment */
|
|
data->filename[count++] = tmp;
|
|
strcpy( tmp, &(full[0]) );
|
|
tmp += strlen( &(full[0]) ) + 1;
|
|
}
|
|
}
|
|
}
|
|
fclose( input );
|
|
}
|
|
|
|
return data;
|
|
}
|
|
|
|
static
|
|
void icvReleaseBackgroundData( CvBackgroundData** data )
|
|
{
|
|
assert( data != NULL && (*data) != NULL );
|
|
|
|
cvFree( data );
|
|
}
|
|
|
|
static
|
|
CvBackgroundReader* icvCreateBackgroundReader()
|
|
{
|
|
CvBackgroundReader* reader = NULL;
|
|
|
|
reader = (CvBackgroundReader*) cvAlloc( sizeof( *reader ) );
|
|
memset( (void*) reader, 0, sizeof( *reader ) );
|
|
reader->src = cvMat( 0, 0, CV_8UC1, NULL );
|
|
reader->img = cvMat( 0, 0, CV_8UC1, NULL );
|
|
reader->offset = cvPoint( 0, 0 );
|
|
reader->scale = 1.0F;
|
|
reader->scalefactor = 1.4142135623730950488016887242097F;
|
|
reader->stepfactor = 0.5F;
|
|
reader->point = reader->offset;
|
|
|
|
return reader;
|
|
}
|
|
|
|
static
|
|
void icvReleaseBackgroundReader( CvBackgroundReader** reader )
|
|
{
|
|
assert( reader != NULL && (*reader) != NULL );
|
|
|
|
if( (*reader)->src.data.ptr != NULL )
|
|
{
|
|
cvFree( &((*reader)->src.data.ptr) );
|
|
}
|
|
if( (*reader)->img.data.ptr != NULL )
|
|
{
|
|
cvFree( &((*reader)->img.data.ptr) );
|
|
}
|
|
|
|
cvFree( reader );
|
|
}
|
|
|
|
static
|
|
void icvGetNextFromBackgroundData( CvBackgroundData* data,
|
|
CvBackgroundReader* reader )
|
|
{
|
|
IplImage* img = NULL;
|
|
size_t datasize = 0;
|
|
int round = 0;
|
|
int i = 0;
|
|
CvPoint offset = cvPoint(0,0);
|
|
|
|
assert( data != NULL && reader != NULL );
|
|
|
|
if( reader->src.data.ptr != NULL )
|
|
{
|
|
cvFree( &(reader->src.data.ptr) );
|
|
reader->src.data.ptr = NULL;
|
|
}
|
|
if( reader->img.data.ptr != NULL )
|
|
{
|
|
cvFree( &(reader->img.data.ptr) );
|
|
reader->img.data.ptr = NULL;
|
|
}
|
|
|
|
#ifdef CV_OPENMP
|
|
#pragma omp critical(c_background_data)
|
|
#endif /* CV_OPENMP */
|
|
{
|
|
for( i = 0; i < data->count; i++ )
|
|
{
|
|
round = data->round;
|
|
|
|
//#ifdef CV_VERBOSE
|
|
// printf( "Open background image: %s\n", data->filename[data->last] );
|
|
//#endif /* CV_VERBOSE */
|
|
|
|
img = cvLoadImage( data->filename[data->last++], 0 );
|
|
data->last %= data->count;
|
|
if( !img )
|
|
continue;
|
|
data->round += data->last / data->count;
|
|
data->round = data->round % (data->winsize.width * data->winsize.height);
|
|
|
|
offset.x = round % data->winsize.width;
|
|
offset.y = round / data->winsize.width;
|
|
|
|
offset.x = MIN( offset.x, img->width - data->winsize.width );
|
|
offset.y = MIN( offset.y, img->height - data->winsize.height );
|
|
|
|
if( img != NULL && img->depth == IPL_DEPTH_8U && img->nChannels == 1 &&
|
|
offset.x >= 0 && offset.y >= 0 )
|
|
{
|
|
break;
|
|
}
|
|
if( img != NULL )
|
|
cvReleaseImage( &img );
|
|
img = NULL;
|
|
}
|
|
}
|
|
if( img == NULL )
|
|
{
|
|
/* no appropriate image */
|
|
|
|
#ifdef CV_VERBOSE
|
|
printf( "Invalid background description file.\n" );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
assert( 0 );
|
|
exit( 1 );
|
|
}
|
|
datasize = sizeof( uchar ) * img->width * img->height;
|
|
reader->src = cvMat( img->height, img->width, CV_8UC1, (void*) cvAlloc( datasize ) );
|
|
cvCopy( img, &reader->src, NULL );
|
|
cvReleaseImage( &img );
|
|
img = NULL;
|
|
|
|
//reader->offset.x = round % data->winsize.width;
|
|
//reader->offset.y = round / data->winsize.width;
|
|
reader->offset = offset;
|
|
reader->point = reader->offset;
|
|
reader->scale = MAX(
|
|
((float) data->winsize.width + reader->point.x) / ((float) reader->src.cols),
|
|
((float) data->winsize.height + reader->point.y) / ((float) reader->src.rows) );
|
|
|
|
reader->img = cvMat( (int) (reader->scale * reader->src.rows + 0.5F),
|
|
(int) (reader->scale * reader->src.cols + 0.5F),
|
|
CV_8UC1, (void*) cvAlloc( datasize ) );
|
|
cvResize( &(reader->src), &(reader->img) );
|
|
}
|
|
|
|
|
|
/*
|
|
* icvGetBackgroundImage
|
|
*
|
|
* Get an image from background
|
|
* <img> must be allocated and have size, previously passed to icvInitBackgroundReaders
|
|
*
|
|
* Usage example:
|
|
* icvInitBackgroundReaders( "bg.txt", cvSize( 24, 24 ) );
|
|
* ...
|
|
* #pragma omp parallel
|
|
* {
|
|
* ...
|
|
* icvGetBackgourndImage( cvbgdata, cvbgreader, img );
|
|
* ...
|
|
* }
|
|
* ...
|
|
* icvDestroyBackgroundReaders();
|
|
*/
|
|
static
|
|
void icvGetBackgroundImage( CvBackgroundData* data,
|
|
CvBackgroundReader* reader,
|
|
CvMat* img )
|
|
{
|
|
CvMat mat;
|
|
|
|
assert( data != NULL && reader != NULL && img != NULL );
|
|
assert( CV_MAT_TYPE( img->type ) == CV_8UC1 );
|
|
assert( img->cols == data->winsize.width );
|
|
assert( img->rows == data->winsize.height );
|
|
|
|
if( reader->img.data.ptr == NULL )
|
|
{
|
|
icvGetNextFromBackgroundData( data, reader );
|
|
}
|
|
|
|
mat = cvMat( data->winsize.height, data->winsize.width, CV_8UC1 );
|
|
cvSetData( &mat, (void*) (reader->img.data.ptr + reader->point.y * reader->img.step
|
|
+ reader->point.x * sizeof( uchar )), reader->img.step );
|
|
|
|
cvCopy( &mat, img, 0 );
|
|
if( (int) ( reader->point.x + (1.0F + reader->stepfactor ) * data->winsize.width )
|
|
< reader->img.cols )
|
|
{
|
|
reader->point.x += (int) (reader->stepfactor * data->winsize.width);
|
|
}
|
|
else
|
|
{
|
|
reader->point.x = reader->offset.x;
|
|
if( (int) ( reader->point.y + (1.0F + reader->stepfactor ) * data->winsize.height )
|
|
< reader->img.rows )
|
|
{
|
|
reader->point.y += (int) (reader->stepfactor * data->winsize.height);
|
|
}
|
|
else
|
|
{
|
|
reader->point.y = reader->offset.y;
|
|
reader->scale *= reader->scalefactor;
|
|
if( reader->scale <= 1.0F )
|
|
{
|
|
reader->img = cvMat( (int) (reader->scale * reader->src.rows),
|
|
(int) (reader->scale * reader->src.cols),
|
|
CV_8UC1, (void*) (reader->img.data.ptr) );
|
|
cvResize( &(reader->src), &(reader->img) );
|
|
}
|
|
else
|
|
{
|
|
icvGetNextFromBackgroundData( data, reader );
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
/*
|
|
* icvInitBackgroundReaders
|
|
*
|
|
* Initialize background reading process.
|
|
* <cvbgreader> and <cvbgdata> are initialized.
|
|
* Must be called before any usage of background
|
|
*
|
|
* filename - name of background description file
|
|
* winsize - size of images will be obtained from background
|
|
*
|
|
* return 1 on success, 0 otherwise.
|
|
*/
|
|
static
|
|
int icvInitBackgroundReaders( const char* filename, CvSize winsize )
|
|
{
|
|
if( cvbgdata == NULL && filename != NULL )
|
|
{
|
|
cvbgdata = icvCreateBackgroundData( filename, winsize );
|
|
}
|
|
|
|
if( cvbgdata )
|
|
{
|
|
|
|
#ifdef CV_OPENMP
|
|
#pragma omp parallel
|
|
#endif /* CV_OPENMP */
|
|
{
|
|
#ifdef CV_OPENMP
|
|
#pragma omp critical(c_create_bg_data)
|
|
#endif /* CV_OPENMP */
|
|
{
|
|
if( cvbgreader == NULL )
|
|
{
|
|
cvbgreader = icvCreateBackgroundReader();
|
|
}
|
|
}
|
|
}
|
|
|
|
}
|
|
|
|
return (cvbgdata != NULL);
|
|
}
|
|
|
|
|
|
/*
|
|
* icvDestroyBackgroundReaders
|
|
*
|
|
* Finish backgournd reading process
|
|
*/
|
|
static
|
|
void icvDestroyBackgroundReaders()
|
|
{
|
|
/* release background reader in each thread */
|
|
#ifdef CV_OPENMP
|
|
#pragma omp parallel
|
|
#endif /* CV_OPENMP */
|
|
{
|
|
#ifdef CV_OPENMP
|
|
#pragma omp critical(c_release_bg_data)
|
|
#endif /* CV_OPENMP */
|
|
{
|
|
if( cvbgreader != NULL )
|
|
{
|
|
icvReleaseBackgroundReader( &cvbgreader );
|
|
cvbgreader = NULL;
|
|
}
|
|
}
|
|
}
|
|
|
|
if( cvbgdata != NULL )
|
|
{
|
|
icvReleaseBackgroundData( &cvbgdata );
|
|
cvbgdata = NULL;
|
|
}
|
|
}
|
|
|
|
|
|
/*
|
|
* icvGetAuxImages
|
|
*
|
|
* Get sum, tilted, sqsum images and calculate normalization factor
|
|
* All images must be allocated.
|
|
*/
|
|
static
|
|
void icvGetAuxImages( CvMat* img, CvMat* sum, CvMat* tilted,
|
|
CvMat* sqsum, float* normfactor )
|
|
{
|
|
CvRect normrect;
|
|
int p0, p1, p2, p3;
|
|
sum_type valsum = 0;
|
|
sqsum_type valsqsum = 0;
|
|
double area = 0.0;
|
|
|
|
cvIntegral( img, sum, sqsum, tilted );
|
|
normrect = cvRect( 1, 1, img->cols - 2, img->rows - 2 );
|
|
CV_SUM_OFFSETS( p0, p1, p2, p3, normrect, img->cols + 1 )
|
|
|
|
area = normrect.width * normrect.height;
|
|
valsum = ((sum_type*) (sum->data.ptr))[p0] - ((sum_type*) (sum->data.ptr))[p1]
|
|
- ((sum_type*) (sum->data.ptr))[p2] + ((sum_type*) (sum->data.ptr))[p3];
|
|
valsqsum = ((sqsum_type*) (sqsum->data.ptr))[p0]
|
|
- ((sqsum_type*) (sqsum->data.ptr))[p1]
|
|
- ((sqsum_type*) (sqsum->data.ptr))[p2]
|
|
+ ((sqsum_type*) (sqsum->data.ptr))[p3];
|
|
|
|
/* sqrt( valsqsum / area - ( valsum / are )^2 ) * area */
|
|
(*normfactor) = (float) sqrt( (double) (area * valsqsum - (double)valsum * valsum) );
|
|
}
|
|
|
|
|
|
/* consumed counter */
|
|
typedef uint64 ccounter_t;
|
|
|
|
#define CCOUNTER_MAX CV_BIG_UINT(0xffffffffffffffff)
|
|
#define CCOUNTER_SET_ZERO(cc) ((cc) = 0)
|
|
#define CCOUNTER_INC(cc) ( (CCOUNTER_MAX > (cc) ) ? (++(cc)) : (CCOUNTER_MAX) )
|
|
#define CCOUNTER_ADD(cc0, cc1) ( ((CCOUNTER_MAX-(cc1)) > (cc0) ) ? ((cc0) += (cc1)) : ((cc0) = CCOUNTER_MAX) )
|
|
#define CCOUNTER_DIV(cc0, cc1) ( ((cc1) == 0) ? 0 : ( ((double)(cc0))/(double)(int64)(cc1) ) )
|
|
|
|
|
|
|
|
/*
|
|
* icvGetHaarTrainingData
|
|
*
|
|
* Unified method that can now be used for vec file, bg file and bg vec file
|
|
*
|
|
* Fill <data> with samples, passed <cascade>
|
|
*/
|
|
static
|
|
int icvGetHaarTrainingData( CvHaarTrainingData* data, int first, int count,
|
|
CvIntHaarClassifier* cascade,
|
|
CvGetHaarTrainingDataCallback callback, void* userdata,
|
|
int* consumed, double* acceptance_ratio )
|
|
{
|
|
int i = 0;
|
|
ccounter_t getcount = 0;
|
|
ccounter_t thread_getcount = 0;
|
|
ccounter_t consumed_count;
|
|
ccounter_t thread_consumed_count;
|
|
|
|
/* private variables */
|
|
CvMat img;
|
|
CvMat sum;
|
|
CvMat tilted;
|
|
CvMat sqsum;
|
|
|
|
sum_type* sumdata;
|
|
sum_type* tilteddata;
|
|
float* normfactor;
|
|
|
|
/* end private variables */
|
|
|
|
assert( data != NULL );
|
|
assert( first + count <= data->maxnum );
|
|
assert( cascade != NULL );
|
|
assert( callback != NULL );
|
|
|
|
// if( !cvbgdata ) return 0; this check needs to be done in the callback for BG
|
|
|
|
CCOUNTER_SET_ZERO(getcount);
|
|
CCOUNTER_SET_ZERO(thread_getcount);
|
|
CCOUNTER_SET_ZERO(consumed_count);
|
|
CCOUNTER_SET_ZERO(thread_consumed_count);
|
|
|
|
#ifdef CV_OPENMP
|
|
#pragma omp parallel private(img, sum, tilted, sqsum, sumdata, tilteddata, \
|
|
normfactor, thread_consumed_count, thread_getcount)
|
|
#endif /* CV_OPENMP */
|
|
{
|
|
sumdata = NULL;
|
|
tilteddata = NULL;
|
|
normfactor = NULL;
|
|
|
|
CCOUNTER_SET_ZERO(thread_getcount);
|
|
CCOUNTER_SET_ZERO(thread_consumed_count);
|
|
int ok = 1;
|
|
|
|
img = cvMat( data->winsize.height, data->winsize.width, CV_8UC1,
|
|
cvAlloc( sizeof( uchar ) * data->winsize.height * data->winsize.width ) );
|
|
sum = cvMat( data->winsize.height + 1, data->winsize.width + 1,
|
|
CV_SUM_MAT_TYPE, NULL );
|
|
tilted = cvMat( data->winsize.height + 1, data->winsize.width + 1,
|
|
CV_SUM_MAT_TYPE, NULL );
|
|
sqsum = cvMat( data->winsize.height + 1, data->winsize.width + 1, CV_SQSUM_MAT_TYPE,
|
|
cvAlloc( sizeof( sqsum_type ) * (data->winsize.height + 1)
|
|
* (data->winsize.width + 1) ) );
|
|
|
|
#ifdef CV_OPENMP
|
|
#pragma omp for schedule(static, 1)
|
|
#endif /* CV_OPENMP */
|
|
for( i = first; (i < first + count); i++ )
|
|
{
|
|
if( !ok )
|
|
continue;
|
|
for( ; ; )
|
|
{
|
|
ok = callback( &img, userdata );
|
|
if( !ok )
|
|
break;
|
|
|
|
CCOUNTER_INC(thread_consumed_count);
|
|
|
|
sumdata = (sum_type*) (data->sum.data.ptr + i * data->sum.step);
|
|
tilteddata = (sum_type*) (data->tilted.data.ptr + i * data->tilted.step);
|
|
normfactor = data->normfactor.data.fl + i;
|
|
sum.data.ptr = (uchar*) sumdata;
|
|
tilted.data.ptr = (uchar*) tilteddata;
|
|
icvGetAuxImages( &img, &sum, &tilted, &sqsum, normfactor );
|
|
if( cascade->eval( cascade, sumdata, tilteddata, *normfactor ) != 0.0F )
|
|
{
|
|
CCOUNTER_INC(thread_getcount);
|
|
break;
|
|
}
|
|
}
|
|
|
|
#ifdef CV_VERBOSE
|
|
if( (i - first) % 500 == 0 )
|
|
{
|
|
fprintf( stderr, "%3d%%\r", (int) ( 100.0 * (i - first) / count ) );
|
|
fflush( stderr );
|
|
}
|
|
#endif /* CV_VERBOSE */
|
|
}
|
|
|
|
cvFree( &(img.data.ptr) );
|
|
cvFree( &(sqsum.data.ptr) );
|
|
|
|
#ifdef CV_OPENMP
|
|
#pragma omp critical (c_consumed_count)
|
|
#endif /* CV_OPENMP */
|
|
{
|
|
/* consumed_count += thread_consumed_count; */
|
|
CCOUNTER_ADD(getcount, thread_getcount);
|
|
CCOUNTER_ADD(consumed_count, thread_consumed_count);
|
|
}
|
|
} /* omp parallel */
|
|
|
|
if( consumed != NULL )
|
|
{
|
|
*consumed = (int)consumed_count;
|
|
}
|
|
|
|
if( acceptance_ratio != NULL )
|
|
{
|
|
/* *acceptance_ratio = ((double) count) / consumed_count; */
|
|
*acceptance_ratio = CCOUNTER_DIV(count, consumed_count);
|
|
}
|
|
|
|
return static_cast<int>(getcount);
|
|
}
|
|
|
|
/*
|
|
* icvGetHaarTrainingDataFromBG
|
|
*
|
|
* Fill <data> with background samples, passed <cascade>
|
|
* Background reading process must be initialized before call.
|
|
*/
|
|
//static
|
|
//int icvGetHaarTrainingDataFromBG( CvHaarTrainingData* data, int first, int count,
|
|
// CvIntHaarClassifier* cascade, double* acceptance_ratio )
|
|
//{
|
|
// int i = 0;
|
|
// ccounter_t consumed_count;
|
|
// ccounter_t thread_consumed_count;
|
|
//
|
|
// /* private variables */
|
|
// CvMat img;
|
|
// CvMat sum;
|
|
// CvMat tilted;
|
|
// CvMat sqsum;
|
|
//
|
|
// sum_type* sumdata;
|
|
// sum_type* tilteddata;
|
|
// float* normfactor;
|
|
//
|
|
// /* end private variables */
|
|
//
|
|
// assert( data != NULL );
|
|
// assert( first + count <= data->maxnum );
|
|
// assert( cascade != NULL );
|
|
//
|
|
// if( !cvbgdata ) return 0;
|
|
//
|
|
// CCOUNTER_SET_ZERO(consumed_count);
|
|
// CCOUNTER_SET_ZERO(thread_consumed_count);
|
|
//
|
|
// #ifdef CV_OPENMP
|
|
// #pragma omp parallel private(img, sum, tilted, sqsum, sumdata, tilteddata,
|
|
// normfactor, thread_consumed_count)
|
|
// #endif /* CV_OPENMP */
|
|
// {
|
|
// sumdata = NULL;
|
|
// tilteddata = NULL;
|
|
// normfactor = NULL;
|
|
//
|
|
// CCOUNTER_SET_ZERO(thread_consumed_count);
|
|
//
|
|
// img = cvMat( data->winsize.height, data->winsize.width, CV_8UC1,
|
|
// cvAlloc( sizeof( uchar ) * data->winsize.height * data->winsize.width ) );
|
|
// sum = cvMat( data->winsize.height + 1, data->winsize.width + 1,
|
|
// CV_SUM_MAT_TYPE, NULL );
|
|
// tilted = cvMat( data->winsize.height + 1, data->winsize.width + 1,
|
|
// CV_SUM_MAT_TYPE, NULL );
|
|
// sqsum = cvMat( data->winsize.height + 1, data->winsize.width + 1,
|
|
// CV_SQSUM_MAT_TYPE,
|
|
// cvAlloc( sizeof( sqsum_type ) * (data->winsize.height + 1)
|
|
// * (data->winsize.width + 1) ) );
|
|
//
|
|
// #ifdef CV_OPENMP
|
|
// #pragma omp for schedule(static, 1)
|
|
// #endif /* CV_OPENMP */
|
|
// for( i = first; i < first + count; i++ )
|
|
// {
|
|
// for( ; ; )
|
|
// {
|
|
// icvGetBackgroundImage( cvbgdata, cvbgreader, &img );
|
|
//
|
|
// CCOUNTER_INC(thread_consumed_count);
|
|
//
|
|
// sumdata = (sum_type*) (data->sum.data.ptr + i * data->sum.step);
|
|
// tilteddata = (sum_type*) (data->tilted.data.ptr + i * data->tilted.step);
|
|
// normfactor = data->normfactor.data.fl + i;
|
|
// sum.data.ptr = (uchar*) sumdata;
|
|
// tilted.data.ptr = (uchar*) tilteddata;
|
|
// icvGetAuxImages( &img, &sum, &tilted, &sqsum, normfactor );
|
|
// if( cascade->eval( cascade, sumdata, tilteddata, *normfactor ) != 0.0F )
|
|
// {
|
|
// break;
|
|
// }
|
|
// }
|
|
//
|
|
//#ifdef CV_VERBOSE
|
|
// if( (i - first) % 500 == 0 )
|
|
// {
|
|
// fprintf( stderr, "%3d%%\r", (int) ( 100.0 * (i - first) / count ) );
|
|
// fflush( stderr );
|
|
// }
|
|
//#endif /* CV_VERBOSE */
|
|
//
|
|
// }
|
|
//
|
|
// cvFree( &(img.data.ptr) );
|
|
// cvFree( &(sqsum.data.ptr) );
|
|
//
|
|
// #ifdef CV_OPENMP
|
|
// #pragma omp critical (c_consumed_count)
|
|
// #endif /* CV_OPENMP */
|
|
// {
|
|
// /* consumed_count += thread_consumed_count; */
|
|
// CCOUNTER_ADD(consumed_count, thread_consumed_count);
|
|
// }
|
|
// } /* omp parallel */
|
|
//
|
|
// if( acceptance_ratio != NULL )
|
|
// {
|
|
// /* *acceptance_ratio = ((double) count) / consumed_count; */
|
|
// *acceptance_ratio = CCOUNTER_DIV(count, consumed_count);
|
|
// }
|
|
//
|
|
// return count;
|
|
//}
|
|
|
|
int icvGetHaarTraininDataFromVecCallback( CvMat* img, void* userdata )
|
|
{
|
|
uchar tmp = 0;
|
|
int r = 0;
|
|
int c = 0;
|
|
|
|
assert( img->rows * img->cols == ((CvVecFile*) userdata)->vecsize );
|
|
|
|
fread( &tmp, sizeof( tmp ), 1, ((CvVecFile*) userdata)->input );
|
|
fread( ((CvVecFile*) userdata)->vector, sizeof( short ),
|
|
((CvVecFile*) userdata)->vecsize, ((CvVecFile*) userdata)->input );
|
|
|
|
if( feof( ((CvVecFile*) userdata)->input ) ||
|
|
(((CvVecFile*) userdata)->last)++ >= ((CvVecFile*) userdata)->count )
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
for( r = 0; r < img->rows; r++ )
|
|
{
|
|
for( c = 0; c < img->cols; c++ )
|
|
{
|
|
CV_MAT_ELEM( *img, uchar, r, c ) =
|
|
(uchar) ( ((CvVecFile*) userdata)->vector[r * img->cols + c] );
|
|
}
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
int icvGetHaarTrainingDataFromBGCallback ( CvMat* img, void* /*userdata*/ )
|
|
{
|
|
if (! cvbgdata)
|
|
return 0;
|
|
|
|
if (! cvbgreader)
|
|
return 0;
|
|
|
|
// just in case icvGetBackgroundImage is not thread-safe ...
|
|
#ifdef CV_OPENMP
|
|
#pragma omp critical (get_background_image_callback)
|
|
#endif /* CV_OPENMP */
|
|
{
|
|
icvGetBackgroundImage( cvbgdata, cvbgreader, img );
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* icvGetHaarTrainingDataFromVec
|
|
* Get training data from .vec file
|
|
*/
|
|
static
|
|
int icvGetHaarTrainingDataFromVec( CvHaarTrainingData* data, int first, int count,
|
|
CvIntHaarClassifier* cascade,
|
|
const char* filename,
|
|
int* consumed )
|
|
{
|
|
int getcount = 0;
|
|
|
|
CV_FUNCNAME( "icvGetHaarTrainingDataFromVec" );
|
|
|
|
__BEGIN__;
|
|
|
|
CvVecFile file;
|
|
short tmp = 0;
|
|
|
|
file.input = NULL;
|
|
if( filename ) file.input = fopen( filename, "rb" );
|
|
|
|
if( file.input != NULL )
|
|
{
|
|
fread( &file.count, sizeof( file.count ), 1, file.input );
|
|
fread( &file.vecsize, sizeof( file.vecsize ), 1, file.input );
|
|
fread( &tmp, sizeof( tmp ), 1, file.input );
|
|
fread( &tmp, sizeof( tmp ), 1, file.input );
|
|
if( !feof( file.input ) )
|
|
{
|
|
if( file.vecsize != data->winsize.width * data->winsize.height )
|
|
{
|
|
fclose( file.input );
|
|
CV_ERROR( CV_StsError, "Vec file sample size mismatch" );
|
|
}
|
|
|
|
file.last = 0;
|
|
file.vector = (short*) cvAlloc( sizeof( *file.vector ) * file.vecsize );
|
|
getcount = icvGetHaarTrainingData( data, first, count, cascade,
|
|
icvGetHaarTraininDataFromVecCallback, &file, consumed, NULL);
|
|
cvFree( &file.vector );
|
|
}
|
|
fclose( file.input );
|
|
}
|
|
|
|
__END__;
|
|
|
|
return getcount;
|
|
}
|
|
|
|
/*
|
|
* icvGetHaarTrainingDataFromBG
|
|
*
|
|
* Fill <data> with background samples, passed <cascade>
|
|
* Background reading process must be initialized before call, alternaly, a file
|
|
* name to a vec file may be passed, a NULL filename indicates old behaviour
|
|
*/
|
|
static
|
|
int icvGetHaarTrainingDataFromBG( CvHaarTrainingData* data, int first, int count,
|
|
CvIntHaarClassifier* cascade, double* acceptance_ratio, const char * filename = NULL )
|
|
{
|
|
CV_FUNCNAME( "icvGetHaarTrainingDataFromBG" );
|
|
|
|
__BEGIN__;
|
|
|
|
if (filename)
|
|
{
|
|
CvVecFile file;
|
|
short tmp = 0;
|
|
|
|
file.input = NULL;
|
|
if( filename ) file.input = fopen( filename, "rb" );
|
|
|
|
if( file.input != NULL )
|
|
{
|
|
fread( &file.count, sizeof( file.count ), 1, file.input );
|
|
fread( &file.vecsize, sizeof( file.vecsize ), 1, file.input );
|
|
fread( &tmp, sizeof( tmp ), 1, file.input );
|
|
fread( &tmp, sizeof( tmp ), 1, file.input );
|
|
if( !feof( file.input ) )
|
|
{
|
|
if( file.vecsize != data->winsize.width * data->winsize.height )
|
|
{
|
|
fclose( file.input );
|
|
CV_ERROR( CV_StsError, "Vec file sample size mismatch" );
|
|
}
|
|
|
|
file.last = 0;
|
|
file.vector = (short*) cvAlloc( sizeof( *file.vector ) * file.vecsize );
|
|
icvGetHaarTrainingData( data, first, count, cascade,
|
|
icvGetHaarTraininDataFromVecCallback, &file, NULL, acceptance_ratio);
|
|
cvFree( &file.vector );
|
|
}
|
|
fclose( file.input );
|
|
}
|
|
}
|
|
else
|
|
{
|
|
icvGetHaarTrainingData( data, first, count, cascade,
|
|
icvGetHaarTrainingDataFromBGCallback, NULL, NULL, acceptance_ratio);
|
|
}
|
|
|
|
__END__;
|
|
|
|
return count;
|
|
}
|
|
|
|
void cvCreateCascadeClassifier( const char* dirname,
|
|
const char* vecfilename,
|
|
const char* bgfilename,
|
|
int npos, int nneg, int nstages,
|
|
int numprecalculated,
|
|
int numsplits,
|
|
float minhitrate, float maxfalsealarm,
|
|
float weightfraction,
|
|
int mode, int symmetric,
|
|
int equalweights,
|
|
int winwidth, int winheight,
|
|
int boosttype, int stumperror )
|
|
{
|
|
CvCascadeHaarClassifier* cascade = NULL;
|
|
CvHaarTrainingData* data = NULL;
|
|
CvIntHaarFeatures* haar_features;
|
|
CvSize winsize;
|
|
int i = 0;
|
|
int j = 0;
|
|
int poscount = 0;
|
|
int negcount = 0;
|
|
int consumed = 0;
|
|
double false_alarm = 0;
|
|
char stagename[PATH_MAX];
|
|
float posweight = 1.0F;
|
|
float negweight = 1.0F;
|
|
FILE* file;
|
|
|
|
#ifdef CV_VERBOSE
|
|
double proctime = 0.0F;
|
|
#endif /* CV_VERBOSE */
|
|
|
|
assert( dirname != NULL );
|
|
assert( bgfilename != NULL );
|
|
assert( vecfilename != NULL );
|
|
assert( nstages > 0 );
|
|
|
|
winsize = cvSize( winwidth, winheight );
|
|
|
|
cascade = (CvCascadeHaarClassifier*) icvCreateCascadeHaarClassifier( nstages );
|
|
cascade->count = 0;
|
|
|
|
if( icvInitBackgroundReaders( bgfilename, winsize ) )
|
|
{
|
|
data = icvCreateHaarTrainingData( winsize, npos + nneg );
|
|
haar_features = icvCreateIntHaarFeatures( winsize, mode, symmetric );
|
|
|
|
#ifdef CV_VERBOSE
|
|
printf("Number of features used : %d\n", haar_features->count);
|
|
#endif /* CV_VERBOSE */
|
|
|
|
for( i = 0; i < nstages; i++, cascade->count++ )
|
|
{
|
|
sprintf( stagename, "%s%d/%s", dirname, i, CV_STAGE_CART_FILE_NAME );
|
|
cascade->classifier[i] =
|
|
icvLoadCARTStageHaarClassifier( stagename, winsize.width + 1 );
|
|
|
|
if( !icvMkDir( stagename ) )
|
|
{
|
|
|
|
#ifdef CV_VERBOSE
|
|
printf( "UNABLE TO CREATE DIRECTORY: %s\n", stagename );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
break;
|
|
}
|
|
if( cascade->classifier[i] != NULL )
|
|
{
|
|
|
|
#ifdef CV_VERBOSE
|
|
printf( "STAGE: %d LOADED.\n", i );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
continue;
|
|
}
|
|
|
|
#ifdef CV_VERBOSE
|
|
printf( "STAGE: %d\n", i );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
poscount = icvGetHaarTrainingDataFromVec( data, 0, npos,
|
|
(CvIntHaarClassifier*) cascade, vecfilename, &consumed );
|
|
#ifdef CV_VERBOSE
|
|
printf( "POS: %d %d %f\n", poscount, consumed,
|
|
((float) poscount) / consumed );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
if( poscount <= 0 )
|
|
{
|
|
|
|
#ifdef CV_VERBOSE
|
|
printf( "UNABLE TO OBTAIN POS SAMPLES\n" );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
break;
|
|
}
|
|
|
|
#ifdef CV_VERBOSE
|
|
proctime = -TIME( 0 );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
negcount = icvGetHaarTrainingDataFromBG( data, poscount, nneg,
|
|
(CvIntHaarClassifier*) cascade, &false_alarm );
|
|
#ifdef CV_VERBOSE
|
|
printf( "NEG: %d %g\n", negcount, false_alarm );
|
|
printf( "BACKGROUND PROCESSING TIME: %.2f\n",
|
|
(proctime + TIME( 0 )) );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
if( negcount <= 0 )
|
|
{
|
|
|
|
#ifdef CV_VERBOSE
|
|
printf( "UNABLE TO OBTAIN NEG SAMPLES\n" );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
break;
|
|
}
|
|
|
|
data->sum.rows = data->tilted.rows = poscount + negcount;
|
|
data->normfactor.cols = data->weights.cols = data->cls.cols =
|
|
poscount + negcount;
|
|
|
|
posweight = (equalweights) ? 1.0F / (poscount + negcount) : (0.5F / poscount);
|
|
negweight = (equalweights) ? 1.0F / (poscount + negcount) : (0.5F / negcount);
|
|
for( j = 0; j < poscount; j++ )
|
|
{
|
|
data->weights.data.fl[j] = posweight;
|
|
data->cls.data.fl[j] = 1.0F;
|
|
|
|
}
|
|
for( j = poscount; j < poscount + negcount; j++ )
|
|
{
|
|
data->weights.data.fl[j] = negweight;
|
|
data->cls.data.fl[j] = 0.0F;
|
|
}
|
|
|
|
#ifdef CV_VERBOSE
|
|
proctime = -TIME( 0 );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
icvPrecalculate( data, haar_features, numprecalculated );
|
|
|
|
#ifdef CV_VERBOSE
|
|
printf( "PRECALCULATION TIME: %.2f\n", (proctime + TIME( 0 )) );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
#ifdef CV_VERBOSE
|
|
proctime = -TIME( 0 );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
cascade->classifier[i] = icvCreateCARTStageClassifier( data, NULL,
|
|
haar_features, minhitrate, maxfalsealarm, symmetric, weightfraction,
|
|
numsplits, (CvBoostType) boosttype, (CvStumpError) stumperror, 0 );
|
|
|
|
#ifdef CV_VERBOSE
|
|
printf( "STAGE TRAINING TIME: %.2f\n", (proctime + TIME( 0 )) );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
file = fopen( stagename, "w" );
|
|
if( file != NULL )
|
|
{
|
|
cascade->classifier[i]->save(
|
|
(CvIntHaarClassifier*) cascade->classifier[i], file );
|
|
fclose( file );
|
|
}
|
|
else
|
|
{
|
|
|
|
#ifdef CV_VERBOSE
|
|
printf( "FAILED TO SAVE STAGE CLASSIFIER IN FILE %s\n", stagename );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
}
|
|
|
|
}
|
|
icvReleaseIntHaarFeatures( &haar_features );
|
|
icvReleaseHaarTrainingData( &data );
|
|
|
|
if( i == nstages )
|
|
{
|
|
char xml_path[1024];
|
|
int len = (int)strlen(dirname);
|
|
CvHaarClassifierCascade* cascade = 0;
|
|
strcpy( xml_path, dirname );
|
|
if( xml_path[len-1] == '\\' || xml_path[len-1] == '/' )
|
|
len--;
|
|
strcpy( xml_path + len, ".xml" );
|
|
cascade = cvLoadHaarClassifierCascade( dirname, cvSize(winwidth,winheight) );
|
|
if( cascade )
|
|
cvSave( xml_path, cascade );
|
|
cvReleaseHaarClassifierCascade( &cascade );
|
|
}
|
|
}
|
|
else
|
|
{
|
|
#ifdef CV_VERBOSE
|
|
printf( "FAILED TO INITIALIZE BACKGROUND READERS\n" );
|
|
#endif /* CV_VERBOSE */
|
|
}
|
|
|
|
/* CLEAN UP */
|
|
icvDestroyBackgroundReaders();
|
|
cascade->release( (CvIntHaarClassifier**) &cascade );
|
|
}
|
|
|
|
/* tree cascade classifier */
|
|
|
|
int icvNumSplits( CvStageHaarClassifier* stage )
|
|
{
|
|
int i;
|
|
int num;
|
|
|
|
num = 0;
|
|
for( i = 0; i < stage->count; i++ )
|
|
{
|
|
num += ((CvCARTHaarClassifier*) stage->classifier[i])->count;
|
|
}
|
|
|
|
return num;
|
|
}
|
|
|
|
void icvSetNumSamples( CvHaarTrainingData* training_data, int num )
|
|
{
|
|
assert( num <= training_data->maxnum );
|
|
|
|
training_data->sum.rows = training_data->tilted.rows = num;
|
|
training_data->normfactor.cols = num;
|
|
training_data->cls.cols = training_data->weights.cols = num;
|
|
}
|
|
|
|
void icvSetWeightsAndClasses( CvHaarTrainingData* training_data,
|
|
int num1, float weight1, float cls1,
|
|
int num2, float weight2, float cls2 )
|
|
{
|
|
int j;
|
|
|
|
assert( num1 + num2 <= training_data->maxnum );
|
|
|
|
for( j = 0; j < num1; j++ )
|
|
{
|
|
training_data->weights.data.fl[j] = weight1;
|
|
training_data->cls.data.fl[j] = cls1;
|
|
}
|
|
for( j = num1; j < num1 + num2; j++ )
|
|
{
|
|
training_data->weights.data.fl[j] = weight2;
|
|
training_data->cls.data.fl[j] = cls2;
|
|
}
|
|
}
|
|
|
|
CvMat* icvGetUsedValues( CvHaarTrainingData* training_data,
|
|
int start, int num,
|
|
CvIntHaarFeatures* haar_features,
|
|
CvStageHaarClassifier* stage )
|
|
{
|
|
CvMat* ptr = NULL;
|
|
CvMat* feature_idx = NULL;
|
|
|
|
CV_FUNCNAME( "icvGetUsedValues" );
|
|
|
|
__BEGIN__;
|
|
|
|
int num_splits;
|
|
int i, j;
|
|
int r;
|
|
int total, last;
|
|
|
|
num_splits = icvNumSplits( stage );
|
|
|
|
CV_CALL( feature_idx = cvCreateMat( 1, num_splits, CV_32SC1 ) );
|
|
|
|
total = 0;
|
|
for( i = 0; i < stage->count; i++ )
|
|
{
|
|
CvCARTHaarClassifier* cart;
|
|
|
|
cart = (CvCARTHaarClassifier*) stage->classifier[i];
|
|
for( j = 0; j < cart->count; j++ )
|
|
{
|
|
feature_idx->data.i[total++] = cart->compidx[j];
|
|
}
|
|
}
|
|
icvSort_32s( feature_idx->data.i, total, 0 );
|
|
|
|
last = 0;
|
|
for( i = 1; i < total; i++ )
|
|
{
|
|
if( feature_idx->data.i[i] != feature_idx->data.i[last] )
|
|
{
|
|
feature_idx->data.i[++last] = feature_idx->data.i[i];
|
|
}
|
|
}
|
|
total = last + 1;
|
|
CV_CALL( ptr = cvCreateMat( num, total, CV_32FC1 ) );
|
|
|
|
|
|
#ifdef CV_OPENMP
|
|
#pragma omp parallel for
|
|
#endif
|
|
for( r = start; r < start + num; r++ )
|
|
{
|
|
int c;
|
|
|
|
for( c = 0; c < total; c++ )
|
|
{
|
|
float val, normfactor;
|
|
int fnum;
|
|
|
|
fnum = feature_idx->data.i[c];
|
|
|
|
val = cvEvalFastHaarFeature( haar_features->fastfeature + fnum,
|
|
(sum_type*) (training_data->sum.data.ptr
|
|
+ r * training_data->sum.step),
|
|
(sum_type*) (training_data->tilted.data.ptr
|
|
+ r * training_data->tilted.step) );
|
|
normfactor = training_data->normfactor.data.fl[r];
|
|
val = ( normfactor == 0.0F ) ? 0.0F : (val / normfactor);
|
|
CV_MAT_ELEM( *ptr, float, r - start, c ) = val;
|
|
}
|
|
}
|
|
|
|
__END__;
|
|
|
|
cvReleaseMat( &feature_idx );
|
|
|
|
return ptr;
|
|
}
|
|
|
|
/* possible split in the tree */
|
|
typedef struct CvSplit
|
|
{
|
|
CvTreeCascadeNode* parent;
|
|
CvTreeCascadeNode* single_cluster;
|
|
CvTreeCascadeNode* multiple_clusters;
|
|
int num_clusters;
|
|
float single_multiple_ratio;
|
|
|
|
struct CvSplit* next;
|
|
} CvSplit;
|
|
|
|
|
|
void cvCreateTreeCascadeClassifier( const char* dirname,
|
|
const char* vecfilename,
|
|
const char* bgfilename,
|
|
int npos, int nneg, int nstages,
|
|
int numprecalculated,
|
|
int numsplits,
|
|
float minhitrate, float maxfalsealarm,
|
|
float weightfraction,
|
|
int mode, int symmetric,
|
|
int equalweights,
|
|
int winwidth, int winheight,
|
|
int boosttype, int stumperror,
|
|
int maxtreesplits, int minpos, bool bg_vecfile )
|
|
{
|
|
CvTreeCascadeClassifier* tcc = NULL;
|
|
CvIntHaarFeatures* haar_features = NULL;
|
|
CvHaarTrainingData* training_data = NULL;
|
|
CvMat* vals = NULL;
|
|
CvMat* cluster_idx = NULL;
|
|
CvMat* idx = NULL;
|
|
CvMat* features_idx = NULL;
|
|
|
|
CV_FUNCNAME( "cvCreateTreeCascadeClassifier" );
|
|
|
|
__BEGIN__;
|
|
|
|
int i, k;
|
|
CvTreeCascadeNode* leaves;
|
|
int best_num, cur_num;
|
|
CvSize winsize;
|
|
char stage_name[PATH_MAX];
|
|
char buf[PATH_MAX];
|
|
char* suffix;
|
|
int total_splits;
|
|
|
|
int poscount;
|
|
int negcount;
|
|
int consumed;
|
|
double false_alarm;
|
|
double proctime;
|
|
|
|
int nleaves;
|
|
double required_leaf_fa_rate;
|
|
float neg_ratio;
|
|
|
|
int max_clusters;
|
|
|
|
max_clusters = CV_MAX_CLUSTERS;
|
|
neg_ratio = (float) nneg / npos;
|
|
|
|
nleaves = 1 + MAX( 0, maxtreesplits );
|
|
required_leaf_fa_rate = pow( (double) maxfalsealarm, (double) nstages ) / nleaves;
|
|
|
|
printf( "Required leaf false alarm rate: %g\n", required_leaf_fa_rate );
|
|
|
|
total_splits = 0;
|
|
|
|
winsize = cvSize( winwidth, winheight );
|
|
|
|
CV_CALL( cluster_idx = cvCreateMat( 1, npos + nneg, CV_32SC1 ) );
|
|
CV_CALL( idx = cvCreateMat( 1, npos + nneg, CV_32SC1 ) );
|
|
|
|
CV_CALL( tcc = (CvTreeCascadeClassifier*)
|
|
icvLoadTreeCascadeClassifier( dirname, winwidth + 1, &total_splits ) );
|
|
CV_CALL( leaves = icvFindDeepestLeaves( tcc ) );
|
|
|
|
CV_CALL( icvPrintTreeCascade( tcc->root ) );
|
|
|
|
haar_features = icvCreateIntHaarFeatures( winsize, mode, symmetric );
|
|
|
|
printf( "Number of features used : %d\n", haar_features->count );
|
|
|
|
training_data = icvCreateHaarTrainingData( winsize, npos + nneg );
|
|
|
|
sprintf( stage_name, "%s/", dirname );
|
|
suffix = stage_name + strlen( stage_name );
|
|
|
|
if (! bg_vecfile)
|
|
if( !icvInitBackgroundReaders( bgfilename, winsize ) && nstages > 0 )
|
|
CV_ERROR( CV_StsError, "Unable to read negative images" );
|
|
|
|
if( nstages > 0 )
|
|
{
|
|
/* width-first search in the tree */
|
|
do
|
|
{
|
|
CvSplit* first_split;
|
|
CvSplit* last_split;
|
|
CvSplit* cur_split;
|
|
|
|
CvTreeCascadeNode* parent;
|
|
CvTreeCascadeNode* cur_node;
|
|
CvTreeCascadeNode* last_node;
|
|
|
|
first_split = last_split = cur_split = NULL;
|
|
parent = leaves;
|
|
leaves = NULL;
|
|
do
|
|
{
|
|
int best_clusters; /* best selected number of clusters */
|
|
float posweight, negweight;
|
|
double leaf_fa_rate;
|
|
|
|
if( parent ) sprintf( buf, "%d", parent->idx );
|
|
else sprintf( buf, "NULL" );
|
|
printf( "\nParent node: %s\n\n", buf );
|
|
|
|
printf( "*** 1 cluster ***\n" );
|
|
|
|
tcc->eval = icvEvalTreeCascadeClassifierFilter;
|
|
/* find path from the root to the node <parent> */
|
|
icvSetLeafNode( tcc, parent );
|
|
|
|
/* load samples */
|
|
consumed = 0;
|
|
poscount = icvGetHaarTrainingDataFromVec( training_data, 0, npos,
|
|
(CvIntHaarClassifier*) tcc, vecfilename, &consumed );
|
|
|
|
printf( "POS: %d %d %f\n", poscount, consumed, ((double) poscount)/consumed );
|
|
|
|
if( poscount <= 0 )
|
|
CV_ERROR( CV_StsError, "Unable to obtain positive samples" );
|
|
|
|
fflush( stdout );
|
|
|
|
proctime = -TIME( 0 );
|
|
|
|
nneg = (int) (neg_ratio * poscount);
|
|
negcount = icvGetHaarTrainingDataFromBG( training_data, poscount, nneg,
|
|
(CvIntHaarClassifier*) tcc, &false_alarm, bg_vecfile ? bgfilename : NULL );
|
|
printf( "NEG: %d %g\n", negcount, false_alarm );
|
|
|
|
printf( "BACKGROUND PROCESSING TIME: %.2f\n", (proctime + TIME( 0 )) );
|
|
|
|
if( negcount <= 0 )
|
|
CV_ERROR( CV_StsError, "Unable to obtain negative samples" );
|
|
|
|
leaf_fa_rate = false_alarm;
|
|
if( leaf_fa_rate <= required_leaf_fa_rate )
|
|
{
|
|
printf( "Required leaf false alarm rate achieved. "
|
|
"Branch training terminated.\n" );
|
|
}
|
|
else if( nleaves == 1 && tcc->next_idx == nstages )
|
|
{
|
|
printf( "Required number of stages achieved. "
|
|
"Branch training terminated.\n" );
|
|
}
|
|
else
|
|
{
|
|
CvTreeCascadeNode* single_cluster;
|
|
CvTreeCascadeNode* multiple_clusters;
|
|
CvSplit* cur_split;
|
|
int single_num;
|
|
|
|
icvSetNumSamples( training_data, poscount + negcount );
|
|
posweight = (equalweights) ? 1.0F / (poscount + negcount) : (0.5F/poscount);
|
|
negweight = (equalweights) ? 1.0F / (poscount + negcount) : (0.5F/negcount);
|
|
icvSetWeightsAndClasses( training_data,
|
|
poscount, posweight, 1.0F, negcount, negweight, 0.0F );
|
|
|
|
fflush( stdout );
|
|
|
|
/* precalculate feature values */
|
|
proctime = -TIME( 0 );
|
|
icvPrecalculate( training_data, haar_features, numprecalculated );
|
|
printf( "Precalculation time: %.2f\n", (proctime + TIME( 0 )) );
|
|
|
|
/* train stage classifier using all positive samples */
|
|
CV_CALL( single_cluster = icvCreateTreeCascadeNode() );
|
|
fflush( stdout );
|
|
|
|
proctime = -TIME( 0 );
|
|
single_cluster->stage =
|
|
(CvStageHaarClassifier*) icvCreateCARTStageClassifier(
|
|
training_data, NULL, haar_features,
|
|
minhitrate, maxfalsealarm, symmetric,
|
|
weightfraction, numsplits, (CvBoostType) boosttype,
|
|
(CvStumpError) stumperror, 0 );
|
|
printf( "Stage training time: %.2f\n", (proctime + TIME( 0 )) );
|
|
|
|
single_num = icvNumSplits( single_cluster->stage );
|
|
best_num = single_num;
|
|
best_clusters = 1;
|
|
multiple_clusters = NULL;
|
|
|
|
printf( "Number of used features: %d\n", single_num );
|
|
|
|
if( maxtreesplits >= 0 )
|
|
{
|
|
max_clusters = MIN( max_clusters, maxtreesplits - total_splits + 1 );
|
|
}
|
|
|
|
/* try clustering */
|
|
vals = NULL;
|
|
for( k = 2; k <= max_clusters; k++ )
|
|
{
|
|
int cluster;
|
|
int stop_clustering;
|
|
|
|
printf( "*** %d clusters ***\n", k );
|
|
|
|
/* check whether clusters are big enough */
|
|
stop_clustering = ( k * minpos > poscount );
|
|
if( !stop_clustering )
|
|
{
|
|
int num[CV_MAX_CLUSTERS];
|
|
|
|
if( k == 2 )
|
|
{
|
|
proctime = -TIME( 0 );
|
|
CV_CALL( vals = icvGetUsedValues( training_data, 0, poscount,
|
|
haar_features, single_cluster->stage ) );
|
|
printf( "Getting values for clustering time: %.2f\n", (proctime + TIME(0)) );
|
|
printf( "Value matirx size: %d x %d\n", vals->rows, vals->cols );
|
|
fflush( stdout );
|
|
|
|
cluster_idx->cols = vals->rows;
|
|
for( i = 0; i < negcount; i++ ) idx->data.i[i] = poscount + i;
|
|
}
|
|
|
|
proctime = -TIME( 0 );
|
|
|
|
CV_CALL( cvKMeans2( vals, k, cluster_idx, CV_TERM_CRITERIA() ) );
|
|
|
|
printf( "Clustering time: %.2f\n", (proctime + TIME( 0 )) );
|
|
|
|
for( cluster = 0; cluster < k; cluster++ ) num[cluster] = 0;
|
|
for( i = 0; i < cluster_idx->cols; i++ )
|
|
num[cluster_idx->data.i[i]]++;
|
|
for( cluster = 0; cluster < k; cluster++ )
|
|
{
|
|
if( num[cluster] < minpos )
|
|
{
|
|
stop_clustering = 1;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
if( stop_clustering )
|
|
{
|
|
printf( "Clusters are too small. Clustering aborted.\n" );
|
|
break;
|
|
}
|
|
|
|
cur_num = 0;
|
|
cur_node = last_node = NULL;
|
|
for( cluster = 0; (cluster < k) && (cur_num < best_num); cluster++ )
|
|
{
|
|
CvTreeCascadeNode* new_node;
|
|
|
|
int num_splits;
|
|
int last_pos;
|
|
int total_pos;
|
|
|
|
printf( "Cluster: %d\n", cluster );
|
|
|
|
last_pos = negcount;
|
|
for( i = 0; i < cluster_idx->cols; i++ )
|
|
{
|
|
if( cluster_idx->data.i[i] == cluster )
|
|
{
|
|
idx->data.i[last_pos++] = i;
|
|
}
|
|
}
|
|
idx->cols = last_pos;
|
|
|
|
total_pos = idx->cols - negcount;
|
|
printf( "# pos: %d of %d. (%d%%)\n", total_pos, poscount,
|
|
100 * total_pos / poscount );
|
|
|
|
CV_CALL( new_node = icvCreateTreeCascadeNode() );
|
|
if( last_node ) last_node->next = new_node;
|
|
else cur_node = new_node;
|
|
last_node = new_node;
|
|
|
|
posweight = (equalweights)
|
|
? 1.0F / (total_pos + negcount) : (0.5F / total_pos);
|
|
negweight = (equalweights)
|
|
? 1.0F / (total_pos + negcount) : (0.5F / negcount);
|
|
|
|
icvSetWeightsAndClasses( training_data,
|
|
poscount, posweight, 1.0F, negcount, negweight, 0.0F );
|
|
|
|
/* CV_DEBUG_SAVE( idx ); */
|
|
|
|
fflush( stdout );
|
|
|
|
proctime = -TIME( 0 );
|
|
new_node->stage = (CvStageHaarClassifier*)
|
|
icvCreateCARTStageClassifier( training_data, idx, haar_features,
|
|
minhitrate, maxfalsealarm, symmetric,
|
|
weightfraction, numsplits, (CvBoostType) boosttype,
|
|
(CvStumpError) stumperror, best_num - cur_num );
|
|
printf( "Stage training time: %.2f\n", (proctime + TIME( 0 )) );
|
|
|
|
if( !(new_node->stage) )
|
|
{
|
|
printf( "Stage training aborted.\n" );
|
|
cur_num = best_num + 1;
|
|
}
|
|
else
|
|
{
|
|
num_splits = icvNumSplits( new_node->stage );
|
|
cur_num += num_splits;
|
|
|
|
printf( "Number of used features: %d\n", num_splits );
|
|
}
|
|
} /* for each cluster */
|
|
|
|
if( cur_num < best_num )
|
|
{
|
|
icvReleaseTreeCascadeNodes( &multiple_clusters );
|
|
best_num = cur_num;
|
|
best_clusters = k;
|
|
multiple_clusters = cur_node;
|
|
}
|
|
else
|
|
{
|
|
icvReleaseTreeCascadeNodes( &cur_node );
|
|
}
|
|
} /* try different number of clusters */
|
|
cvReleaseMat( &vals );
|
|
|
|
CV_CALL( cur_split = (CvSplit*) cvAlloc( sizeof( *cur_split ) ) );
|
|
CV_ZERO_OBJ( cur_split );
|
|
|
|
if( last_split ) last_split->next = cur_split;
|
|
else first_split = cur_split;
|
|
last_split = cur_split;
|
|
|
|
cur_split->single_cluster = single_cluster;
|
|
cur_split->multiple_clusters = multiple_clusters;
|
|
cur_split->num_clusters = best_clusters;
|
|
cur_split->parent = parent;
|
|
cur_split->single_multiple_ratio = (float) single_num / best_num;
|
|
}
|
|
|
|
if( parent ) parent = parent->next_same_level;
|
|
} while( parent );
|
|
|
|
/* choose which nodes should be splitted */
|
|
do
|
|
{
|
|
float max_single_multiple_ratio;
|
|
|
|
cur_split = NULL;
|
|
max_single_multiple_ratio = 0.0F;
|
|
last_split = first_split;
|
|
while( last_split )
|
|
{
|
|
if( last_split->single_cluster && last_split->multiple_clusters &&
|
|
last_split->single_multiple_ratio > max_single_multiple_ratio )
|
|
{
|
|
max_single_multiple_ratio = last_split->single_multiple_ratio;
|
|
cur_split = last_split;
|
|
}
|
|
last_split = last_split->next;
|
|
}
|
|
if( cur_split )
|
|
{
|
|
if( maxtreesplits < 0 ||
|
|
cur_split->num_clusters <= maxtreesplits - total_splits + 1 )
|
|
{
|
|
cur_split->single_cluster = NULL;
|
|
total_splits += cur_split->num_clusters - 1;
|
|
}
|
|
else
|
|
{
|
|
icvReleaseTreeCascadeNodes( &(cur_split->multiple_clusters) );
|
|
cur_split->multiple_clusters = NULL;
|
|
}
|
|
}
|
|
} while( cur_split );
|
|
|
|
/* attach new nodes to the tree */
|
|
leaves = last_node = NULL;
|
|
last_split = first_split;
|
|
while( last_split )
|
|
{
|
|
cur_node = (last_split->multiple_clusters)
|
|
? last_split->multiple_clusters : last_split->single_cluster;
|
|
parent = last_split->parent;
|
|
if( parent ) parent->child = cur_node;
|
|
|
|
/* connect leaves via next_same_level and save them */
|
|
for( ; cur_node; cur_node = cur_node->next )
|
|
{
|
|
FILE* file;
|
|
|
|
if( last_node ) last_node->next_same_level = cur_node;
|
|
else leaves = cur_node;
|
|
last_node = cur_node;
|
|
cur_node->parent = parent;
|
|
|
|
cur_node->idx = tcc->next_idx;
|
|
tcc->next_idx++;
|
|
sprintf( suffix, "%d/%s", cur_node->idx, CV_STAGE_CART_FILE_NAME );
|
|
file = NULL;
|
|
if( icvMkDir( stage_name ) && (file = fopen( stage_name, "w" )) != 0 )
|
|
{
|
|
cur_node->stage->save( (CvIntHaarClassifier*) cur_node->stage, file );
|
|
fprintf( file, "\n%d\n%d\n",
|
|
((parent) ? parent->idx : -1),
|
|
((cur_node->next) ? tcc->next_idx : -1) );
|
|
}
|
|
else
|
|
{
|
|
printf( "Failed to save classifier into %s\n", stage_name );
|
|
}
|
|
if( file ) fclose( file );
|
|
}
|
|
|
|
if( parent ) sprintf( buf, "%d", parent->idx );
|
|
else sprintf( buf, "NULL" );
|
|
printf( "\nParent node: %s\n", buf );
|
|
printf( "Chosen number of splits: %d\n\n", (last_split->multiple_clusters)
|
|
? (last_split->num_clusters - 1) : 0 );
|
|
|
|
cur_split = last_split;
|
|
last_split = last_split->next;
|
|
cvFree( &cur_split );
|
|
} /* for each split point */
|
|
|
|
printf( "Total number of splits: %d\n", total_splits );
|
|
|
|
if( !(tcc->root) ) tcc->root = leaves;
|
|
CV_CALL( icvPrintTreeCascade( tcc->root ) );
|
|
|
|
} while( leaves );
|
|
|
|
/* save the cascade to xml file */
|
|
{
|
|
char xml_path[1024];
|
|
int len = (int)strlen(dirname);
|
|
CvHaarClassifierCascade* cascade = 0;
|
|
strcpy( xml_path, dirname );
|
|
if( xml_path[len-1] == '\\' || xml_path[len-1] == '/' )
|
|
len--;
|
|
strcpy( xml_path + len, ".xml" );
|
|
cascade = cvLoadHaarClassifierCascade( dirname, cvSize(winwidth,winheight) );
|
|
if( cascade )
|
|
cvSave( xml_path, cascade );
|
|
cvReleaseHaarClassifierCascade( &cascade );
|
|
}
|
|
|
|
} /* if( nstages > 0 ) */
|
|
|
|
/* check cascade performance */
|
|
printf( "\nCascade performance\n" );
|
|
|
|
tcc->eval = icvEvalTreeCascadeClassifier;
|
|
|
|
/* load samples */
|
|
consumed = 0;
|
|
poscount = icvGetHaarTrainingDataFromVec( training_data, 0, npos,
|
|
(CvIntHaarClassifier*) tcc, vecfilename, &consumed );
|
|
|
|
printf( "POS: %d %d %f\n", poscount, consumed,
|
|
(consumed > 0) ? (((float) poscount)/consumed) : 0 );
|
|
|
|
if( poscount <= 0 )
|
|
fprintf( stderr, "Warning: unable to obtain positive samples\n" );
|
|
|
|
proctime = -TIME( 0 );
|
|
|
|
negcount = icvGetHaarTrainingDataFromBG( training_data, poscount, nneg,
|
|
(CvIntHaarClassifier*) tcc, &false_alarm, bg_vecfile ? bgfilename : NULL );
|
|
|
|
printf( "NEG: %d %g\n", negcount, false_alarm );
|
|
|
|
printf( "BACKGROUND PROCESSING TIME: %.2f\n", (proctime + TIME( 0 )) );
|
|
|
|
if( negcount <= 0 )
|
|
fprintf( stderr, "Warning: unable to obtain negative samples\n" );
|
|
|
|
__END__;
|
|
|
|
if (! bg_vecfile)
|
|
icvDestroyBackgroundReaders();
|
|
|
|
if( tcc ) tcc->release( (CvIntHaarClassifier**) &tcc );
|
|
icvReleaseIntHaarFeatures( &haar_features );
|
|
icvReleaseHaarTrainingData( &training_data );
|
|
cvReleaseMat( &cluster_idx );
|
|
cvReleaseMat( &idx );
|
|
cvReleaseMat( &vals );
|
|
cvReleaseMat( &features_idx );
|
|
}
|
|
|
|
|
|
|
|
void cvCreateTrainingSamples( const char* filename,
|
|
const char* imgfilename, int bgcolor, int bgthreshold,
|
|
const char* bgfilename, int count,
|
|
int invert, int maxintensitydev,
|
|
double maxxangle, double maxyangle, double maxzangle,
|
|
int showsamples,
|
|
int winwidth, int winheight )
|
|
{
|
|
CvSampleDistortionData data;
|
|
|
|
assert( filename != NULL );
|
|
assert( imgfilename != NULL );
|
|
|
|
if( !icvMkDir( filename ) )
|
|
{
|
|
fprintf( stderr, "Unable to create output file: %s\n", filename );
|
|
return;
|
|
}
|
|
if( icvStartSampleDistortion( imgfilename, bgcolor, bgthreshold, &data ) )
|
|
{
|
|
FILE* output = NULL;
|
|
|
|
output = fopen( filename, "wb" );
|
|
if( output != NULL )
|
|
{
|
|
int hasbg;
|
|
int i;
|
|
CvMat sample;
|
|
int inverse;
|
|
|
|
hasbg = 0;
|
|
hasbg = (bgfilename != NULL && icvInitBackgroundReaders( bgfilename,
|
|
cvSize( winwidth,winheight ) ) );
|
|
|
|
sample = cvMat( winheight, winwidth, CV_8UC1, cvAlloc( sizeof( uchar ) *
|
|
winheight * winwidth ) );
|
|
|
|
icvWriteVecHeader( output, count, sample.cols, sample.rows );
|
|
|
|
if( showsamples )
|
|
{
|
|
cvNamedWindow( "Sample", CV_WINDOW_AUTOSIZE );
|
|
}
|
|
|
|
inverse = invert;
|
|
for( i = 0; i < count; i++ )
|
|
{
|
|
if( hasbg )
|
|
{
|
|
icvGetBackgroundImage( cvbgdata, cvbgreader, &sample );
|
|
}
|
|
else
|
|
{
|
|
cvSet( &sample, cvScalar( bgcolor ) );
|
|
}
|
|
|
|
if( invert == CV_RANDOM_INVERT )
|
|
{
|
|
inverse = (rand() > (RAND_MAX/2));
|
|
}
|
|
icvPlaceDistortedSample( &sample, inverse, maxintensitydev,
|
|
maxxangle, maxyangle, maxzangle,
|
|
0 /* nonzero means placing image without cut offs */,
|
|
0.0 /* nozero adds random shifting */,
|
|
0.0 /* nozero adds random scaling */,
|
|
&data );
|
|
|
|
if( showsamples )
|
|
{
|
|
cvShowImage( "Sample", &sample );
|
|
if( cvWaitKey( 0 ) == 27 )
|
|
{
|
|
showsamples = 0;
|
|
}
|
|
}
|
|
|
|
icvWriteVecSample( output, &sample );
|
|
|
|
#ifdef CV_VERBOSE
|
|
if( i % 500 == 0 )
|
|
{
|
|
printf( "\r%3d%%", 100 * i / count );
|
|
}
|
|
#endif /* CV_VERBOSE */
|
|
}
|
|
icvDestroyBackgroundReaders();
|
|
cvFree( &(sample.data.ptr) );
|
|
fclose( output );
|
|
} /* if( output != NULL ) */
|
|
|
|
icvEndSampleDistortion( &data );
|
|
}
|
|
|
|
#ifdef CV_VERBOSE
|
|
printf( "\r \r" );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
}
|
|
|
|
#define CV_INFO_FILENAME "info.dat"
|
|
|
|
|
|
void cvCreateTestSamples( const char* infoname,
|
|
const char* imgfilename, int bgcolor, int bgthreshold,
|
|
const char* bgfilename, int count,
|
|
int invert, int maxintensitydev,
|
|
double maxxangle, double maxyangle, double maxzangle,
|
|
int showsamples,
|
|
int winwidth, int winheight )
|
|
{
|
|
CvSampleDistortionData data;
|
|
|
|
assert( infoname != NULL );
|
|
assert( imgfilename != NULL );
|
|
assert( bgfilename != NULL );
|
|
|
|
if( !icvMkDir( infoname ) )
|
|
{
|
|
|
|
#if CV_VERBOSE
|
|
fprintf( stderr, "Unable to create directory hierarchy: %s\n", infoname );
|
|
#endif /* CV_VERBOSE */
|
|
|
|
return;
|
|
}
|
|
if( icvStartSampleDistortion( imgfilename, bgcolor, bgthreshold, &data ) )
|
|
{
|
|
char fullname[PATH_MAX];
|
|
char* filename;
|
|
CvMat win;
|
|
FILE* info;
|
|
|
|
if( icvInitBackgroundReaders( bgfilename, cvSize( 10, 10 ) ) )
|
|
{
|
|
int i;
|
|
int x, y, width, height;
|
|
float scale;
|
|
float maxscale;
|
|
int inverse;
|
|
|
|
if( showsamples )
|
|
{
|
|
cvNamedWindow( "Image", CV_WINDOW_AUTOSIZE );
|
|
}
|
|
|
|
info = fopen( infoname, "w" );
|
|
strcpy( fullname, infoname );
|
|
filename = strrchr( fullname, '\\' );
|
|
if( filename == NULL )
|
|
{
|
|
filename = strrchr( fullname, '/' );
|
|
}
|
|
if( filename == NULL )
|
|
{
|
|
filename = fullname;
|
|
}
|
|
else
|
|
{
|
|
filename++;
|
|
}
|
|
|
|
count = MIN( count, cvbgdata->count );
|
|
inverse = invert;
|
|
for( i = 0; i < count; i++ )
|
|
{
|
|
icvGetNextFromBackgroundData( cvbgdata, cvbgreader );
|
|
|
|
maxscale = MIN( 0.7F * cvbgreader->src.cols / winwidth,
|
|
0.7F * cvbgreader->src.rows / winheight );
|
|
if( maxscale < 1.0F ) continue;
|
|
|
|
scale = (maxscale - 1.0F) * rand() / RAND_MAX + 1.0F;
|
|
width = (int) (scale * winwidth);
|
|
height = (int) (scale * winheight);
|
|
x = (int) ((0.1+0.8 * rand()/RAND_MAX) * (cvbgreader->src.cols - width));
|
|
y = (int) ((0.1+0.8 * rand()/RAND_MAX) * (cvbgreader->src.rows - height));
|
|
|
|
cvGetSubArr( &cvbgreader->src, &win, cvRect( x, y ,width, height ) );
|
|
if( invert == CV_RANDOM_INVERT )
|
|
{
|
|
inverse = (rand() > (RAND_MAX/2));
|
|
}
|
|
icvPlaceDistortedSample( &win, inverse, maxintensitydev,
|
|
maxxangle, maxyangle, maxzangle,
|
|
1, 0.0, 0.0, &data );
|
|
|
|
|
|
sprintf( filename, "%04d_%04d_%04d_%04d_%04d.jpg",
|
|
(i + 1), x, y, width, height );
|
|
|
|
if( info )
|
|
{
|
|
fprintf( info, "%s %d %d %d %d %d\n",
|
|
filename, 1, x, y, width, height );
|
|
}
|
|
|
|
cvSaveImage( fullname, &cvbgreader->src );
|
|
if( showsamples )
|
|
{
|
|
cvShowImage( "Image", &cvbgreader->src );
|
|
if( cvWaitKey( 0 ) == 27 )
|
|
{
|
|
showsamples = 0;
|
|
}
|
|
}
|
|
}
|
|
if( info ) fclose( info );
|
|
icvDestroyBackgroundReaders();
|
|
}
|
|
icvEndSampleDistortion( &data );
|
|
}
|
|
}
|
|
|
|
|
|
/* End of file. */
|