2014-11-27 20:39:05 +08:00
|
|
|
AKAZE and ORB planar tracking {#tutorial_akaze_tracking}
|
|
|
|
=============================
|
|
|
|
|
|
|
|
Introduction
|
|
|
|
------------
|
|
|
|
|
|
|
|
In this tutorial we will compare *AKAZE* and *ORB* local features using them to find matches between
|
|
|
|
video frames and track object movements.
|
|
|
|
|
|
|
|
The algorithm is as follows:
|
|
|
|
|
|
|
|
- Detect and describe keypoints on the first frame, manually set object boundaries
|
|
|
|
- For every next frame:
|
2014-11-28 21:21:28 +08:00
|
|
|
-# Detect and describe keypoints
|
|
|
|
-# Match them using bruteforce matcher
|
|
|
|
-# Estimate homography transformation using RANSAC
|
|
|
|
-# Filter inliers from all the matches
|
|
|
|
-# Apply homography transformation to the bounding box to find the object
|
|
|
|
-# Draw bounding box and inliers, compute inlier ratio as evaluation metric
|
2014-11-27 20:39:05 +08:00
|
|
|
|
2014-11-28 21:21:28 +08:00
|
|
|
![](images/frame.png)
|
2014-11-27 20:39:05 +08:00
|
|
|
|
2014-11-28 21:21:28 +08:00
|
|
|
Data
|
|
|
|
----
|
2014-11-27 20:39:05 +08:00
|
|
|
|
|
|
|
To do the tracking we need a video and object position on the first frame.
|
|
|
|
|
|
|
|
You can download our example video and data from
|
|
|
|
[here](https://docs.google.com/file/d/0B72G7D4snftJandBb0taLVJHMFk).
|
|
|
|
|
2016-03-19 16:55:18 +08:00
|
|
|
To run the code you have to specify input (camera id or video_file). Then, select a bounding box with the mouse, and press any key to start tracking
|
2014-11-27 20:39:05 +08:00
|
|
|
@code{.none}
|
2016-03-19 16:55:18 +08:00
|
|
|
./planar_tracking blais.mp4
|
2014-11-27 20:39:05 +08:00
|
|
|
@endcode
|
2014-11-28 21:21:28 +08:00
|
|
|
|
|
|
|
Source Code
|
|
|
|
-----------
|
2014-11-27 20:39:05 +08:00
|
|
|
|
2015-04-29 15:31:53 +08:00
|
|
|
@include cpp/tutorial_code/features2D/AKAZE_tracking/planar_tracking.cpp
|
2014-11-27 20:39:05 +08:00
|
|
|
|
2014-11-28 21:21:28 +08:00
|
|
|
Explanation
|
|
|
|
-----------
|
2014-11-27 20:39:05 +08:00
|
|
|
|
2014-11-28 21:21:28 +08:00
|
|
|
### Tracker class
|
2014-11-27 20:39:05 +08:00
|
|
|
|
|
|
|
This class implements algorithm described abobve using given feature detector and descriptor
|
|
|
|
matcher.
|
|
|
|
|
|
|
|
- **Setting up the first frame**
|
|
|
|
@code{.cpp}
|
|
|
|
void Tracker::setFirstFrame(const Mat frame, vector<Point2f> bb, string title, Stats& stats)
|
|
|
|
{
|
|
|
|
first_frame = frame.clone();
|
|
|
|
(*detector)(first_frame, noArray(), first_kp, first_desc);
|
|
|
|
stats.keypoints = (int)first_kp.size();
|
|
|
|
drawBoundingBox(first_frame, bb);
|
|
|
|
putText(first_frame, title, Point(0, 60), FONT_HERSHEY_PLAIN, 5, Scalar::all(0), 4);
|
|
|
|
object_bb = bb;
|
|
|
|
}
|
|
|
|
@endcode
|
|
|
|
We compute and store keypoints and descriptors from the first frame and prepare it for the
|
|
|
|
output.
|
|
|
|
|
|
|
|
We need to save number of detected keypoints to make sure both detectors locate roughly the same
|
|
|
|
number of those.
|
|
|
|
|
|
|
|
- **Processing frames**
|
|
|
|
|
2014-11-28 21:21:28 +08:00
|
|
|
-# Locate keypoints and compute descriptors
|
|
|
|
@code{.cpp}
|
|
|
|
(*detector)(frame, noArray(), kp, desc);
|
|
|
|
@endcode
|
2014-11-27 20:39:05 +08:00
|
|
|
|
2014-11-28 21:21:28 +08:00
|
|
|
To find matches between frames we have to locate the keypoints first.
|
2014-11-27 20:39:05 +08:00
|
|
|
|
2014-11-28 21:21:28 +08:00
|
|
|
In this tutorial detectors are set up to find about 1000 keypoints on each frame.
|
2014-11-27 20:39:05 +08:00
|
|
|
|
2014-11-28 21:21:28 +08:00
|
|
|
-# Use 2-nn matcher to find correspondences
|
|
|
|
@code{.cpp}
|
|
|
|
matcher->knnMatch(first_desc, desc, matches, 2);
|
|
|
|
for(unsigned i = 0; i < matches.size(); i++) {
|
|
|
|
if(matches[i][0].distance < nn_match_ratio * matches[i][1].distance) {
|
|
|
|
matched1.push_back(first_kp[matches[i][0].queryIdx]);
|
|
|
|
matched2.push_back( kp[matches[i][0].trainIdx]);
|
|
|
|
}
|
2014-11-27 20:39:05 +08:00
|
|
|
}
|
2014-11-28 21:21:28 +08:00
|
|
|
@endcode
|
|
|
|
If the closest match is *nn_match_ratio* closer than the second closest one, then it's a
|
|
|
|
match.
|
|
|
|
|
|
|
|
-# Use *RANSAC* to estimate homography transformation
|
|
|
|
@code{.cpp}
|
|
|
|
homography = findHomography(Points(matched1), Points(matched2),
|
|
|
|
RANSAC, ransac_thresh, inlier_mask);
|
|
|
|
@endcode
|
|
|
|
If there are at least 4 matches we can use random sample consensus to estimate image
|
|
|
|
transformation.
|
|
|
|
|
|
|
|
-# Save the inliers
|
|
|
|
@code{.cpp}
|
|
|
|
for(unsigned i = 0; i < matched1.size(); i++) {
|
|
|
|
if(inlier_mask.at<uchar>(i)) {
|
|
|
|
int new_i = static_cast<int>(inliers1.size());
|
|
|
|
inliers1.push_back(matched1[i]);
|
|
|
|
inliers2.push_back(matched2[i]);
|
|
|
|
inlier_matches.push_back(DMatch(new_i, new_i, 0));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
@endcode
|
|
|
|
Since *findHomography* computes the inliers we only have to save the chosen points and
|
|
|
|
matches.
|
2014-11-27 20:39:05 +08:00
|
|
|
|
2014-11-28 21:21:28 +08:00
|
|
|
-# Project object bounding box
|
|
|
|
@code{.cpp}
|
|
|
|
perspectiveTransform(object_bb, new_bb, homography);
|
|
|
|
@endcode
|
2014-11-27 20:39:05 +08:00
|
|
|
|
2014-11-28 21:21:28 +08:00
|
|
|
If there is a reasonable number of inliers we can use estimated transformation to locate the
|
|
|
|
object.
|
|
|
|
|
|
|
|
Results
|
|
|
|
-------
|
2014-11-27 20:39:05 +08:00
|
|
|
|
|
|
|
You can watch the resulting [video on youtube](http://www.youtube.com/watch?v=LWY-w8AGGhE).
|
|
|
|
|
|
|
|
*AKAZE* statistics:
|
|
|
|
@code{.none}
|
|
|
|
Matches 626
|
|
|
|
Inliers 410
|
|
|
|
Inlier ratio 0.58
|
|
|
|
Keypoints 1117
|
|
|
|
@endcode
|
2014-11-28 21:21:28 +08:00
|
|
|
|
2014-11-27 20:39:05 +08:00
|
|
|
*ORB* statistics:
|
|
|
|
@code{.none}
|
|
|
|
Matches 504
|
|
|
|
Inliers 319
|
|
|
|
Inlier ratio 0.56
|
|
|
|
Keypoints 1112
|
|
|
|
@endcode
|