2#include <visp3/core/vpConfig.h>
3#include <visp3/core/vpIoTools.h>
4#include <visp3/gui/vpDisplayFactory.h>
5#include <visp3/io/vpVideoReader.h>
6#include <visp3/mbt/vpMbEdgeTracker.h>
7#include <visp3/vision/vpKeyPoint.h>
9#ifdef ENABLE_VISP_NAMESPACE
13#if defined(VISP_HAVE_OPENCV) && defined(HAVE_OPENCV_IMGPROC) && defined(VISP_HAVE_DISPLAY) && \
14 (((VISP_HAVE_OPENCV_VERSION < 0x050000) && defined(HAVE_OPENCV_CALIB3D) && defined(HAVE_OPENCV_FEATURES2D)) || \
15 ((VISP_HAVE_OPENCV_VERSION >= 0x050000) && defined(HAVE_OPENCV_3D) && defined(HAVE_OPENCV_FEATURES)))
22 std::vector<cv::KeyPoint> trainKeyPoints;
24 keypoint_learning.
detect(I, trainKeyPoints, elapsedTime);
28 std::vector<vpPolygon> polygons;
29 std::vector<std::vector<vpPoint> > roisPt;
30 std::pair<std::vector<vpPolygon>, std::vector<std::vector<vpPoint> > > pair =
tracker.getPolygonFaces();
31 polygons = pair.first;
34 std::vector<cv::Point3f> points3f;
38 tracker.getCameraParameters(cam);
43 keypoint_learning.
buildReference(I, trainKeyPoints, points3f,
true,
id);
47 for (std::vector<cv::KeyPoint>::const_iterator it = trainKeyPoints.begin(); it != trainKeyPoints.end(); ++it) {
54int main(
int argc,
char **argv)
56#if defined(HAVE_OPENCV_IMGPROC) && defined(VISP_HAVE_OPENCV) && \
57 (((VISP_HAVE_OPENCV_VERSION < 0x050000) && defined(HAVE_OPENCV_FEATURES2D)) || \
58 ((VISP_HAVE_OPENCV_VERSION >= 0x050000) && defined(HAVE_OPENCV_FEATURES)))
61#if (VISP_CXX_STANDARD >= VISP_CXX_STANDARD_11)
62 std::shared_ptr<vpDisplay> display, display2;
68 std::string videoname =
"cube.mp4";
70 for (
int i = 1;
i < argc;
i++) {
71 if (std::string(argv[i]) ==
"--name" && i + 1 < argc) {
72 videoname = std::string(argv[++i]);
74 else if (std::string(argv[i]) ==
"--help" || std::string(argv[i]) ==
"-h") {
75 std::cout <<
"\nUsage: " << argv[0]
76 <<
" [--name <video name>]"
77 <<
"[--help] [-h]\n" << std::endl;
84 if (!parentname.empty())
85 objectname = parentname +
"/" + objectname;
87 std::cout <<
"Video name: " << videoname << std::endl;
88 std::cout <<
"Tracker requested config files: " << objectname <<
".[init,"
90 <<
"cao or wrl]" << std::endl;
91 std::cout <<
"Tracker optional config files: " << objectname <<
".[ppm]" << std::endl;
99#if defined(VISP_HAVE_PUGIXML)
101 tracker.loadConfigFile(objectname +
".xml");
102 tracker.getCameraParameters(cam);
118 cam.initPersProjWithoutDistortion(547, 542, 339, 235);
119 tracker.setCameraParameters(cam);
122 tracker.setNearClippingDistance(0.01);
123 tracker.setFarClippingDistance(10.0);
127 tracker.setOgreVisibilityTest(
false);
129 tracker.loadModel(objectname +
".cao");
131 tracker.loadModel(objectname +
".wrl");
132 tracker.setDisplayFeatures(
true);
136 vpKeyPoint keypoint_learning(
"ORB",
"ORB",
"BruteForce-Hamming");
137#if ((VISP_HAVE_OPENCV_VERSION < 0x050000) && defined(HAVE_OPENCV_FEATURES2D)) || ((VISP_HAVE_OPENCV_VERSION >= 0x050000) && defined(HAVE_OPENCV_FEATURES))
138#if (VISP_HAVE_OPENCV_VERSION < 0x030000)
139 keypoint_learning.setDetectorParameter(
"ORB",
"nLevels", 1);
141 cv::Ptr<cv::ORB> orb_learning = keypoint_learning.
getDetector(
"ORB").dynamicCast<cv::ORB>();
143 orb_learning->setNLevels(1);
152 std::string imageName[] = {
"cube0001.png",
"cube0150.png",
"cube0200.png" };
154 vpHomogeneousMatrix(0.02143385294, 0.1098083886, 0.5127439561, 2.087159614, 1.141775176, -0.4701291124),
155 vpHomogeneousMatrix(0.02651282185, -0.03713587374, 0.6873765919, 2.314744454, 0.3492296488, -0.1226054828),
156 vpHomogeneousMatrix(0.02965448956, -0.07283091786, 0.7253526051, 2.300529617, -0.4286674806, 0.1788761025) };
157 for (
int i = 0;
i < 3;
i++) {
160#if (VISP_CXX_STANDARD >= VISP_CXX_STANDARD_11)
166 std::stringstream title;
167 title <<
"Learning cube on image: " << imageName[
i];
173 tracker.setPose(I, initPoseTab[i]);
186 learnCube(I, tracker, keypoint_learning, i);
209 vpKeyPoint keypoint_detection(
"ORB",
"ORB",
"BruteForce-Hamming");
210#if ((VISP_HAVE_OPENCV_VERSION < 0x050000) && defined(HAVE_OPENCV_FEATURES2D)) || ((VISP_HAVE_OPENCV_VERSION >= 0x050000) && defined(HAVE_OPENCV_FEATURES))
211#if (VISP_HAVE_OPENCV_VERSION < 0x030000)
212 keypoint_detection.setDetectorParameter(
"ORB",
"nLevels", 1);
214 cv::Ptr<cv::ORB> orb_detector = keypoint_detection.getDetector(
"ORB").dynamicCast<cv::ORB>();
215 orb_detector = keypoint_detection.getDetector(
"ORB").dynamicCast<cv::ORB>();
217 orb_detector->setNLevels(1);
224 keypoint_detection.loadLearningData(
"cube_learning_data.bin",
true);
229 keypoint_detection.createImageMatching(I, IMatching);
236#if (VISP_CXX_STANDARD >= VISP_CXX_STANDARD_11)
244 bool click_done =
false;
251 keypoint_detection.insertImageMatching(I, IMatching);
259 if (keypoint_detection.matchPoint(I, cam, cMo, error, elapsedTime)) {
271 keypoint_detection.displayMatching(I, IMatching);
274 std::vector<vpImagePoint> ransacInliers = keypoint_detection.getRansacInliers();
275 std::vector<vpImagePoint> ransacOutliers = keypoint_detection.getRansacOutliers();
279 for (std::vector<vpImagePoint>::const_iterator it = ransacInliers.begin(); it != ransacInliers.end(); ++it) {
282 imPt.set_u(imPt.get_u() + I.getWidth());
283 imPt.set_v(imPt.get_v() + I.getHeight());
289 for (std::vector<vpImagePoint>::const_iterator it = ransacOutliers.begin(); it != ransacOutliers.end(); ++it) {
292 imPt.set_u(imPt.get_u() + I.getWidth());
293 imPt.set_v(imPt.get_v() + I.getHeight());
299 keypoint_detection.displayMatching(I, IMatching);
305 cam.get_v0() + I.getHeight());
306 tracker.setCameraParameters(cam2);
307 tracker.setPose(IMatching, cMo);
330 std::cout <<
"Catch an exception: " <<
e << std::endl;
333#if (VISP_CXX_STANDARD < VISP_CXX_STANDARD_11)
334 if (display !=
nullptr) {
338 if (display2 !=
nullptr) {
345 std::cout <<
"Install OpenCV and rebuild ViSP to use this example." << std::endl;
Generic class defining intrinsic camera parameters.
void initPersProjWithoutDistortion(double px, double py, double u0, double v0)
static const vpColor none
static const vpColor green
Class that defines generic functionalities for display.
static bool getClick(const vpImage< unsigned char > &I, bool blocking=true)
static void displayCircle(const vpImage< unsigned char > &I, const vpImageCircle &circle, const vpColor &color, bool fill=false, unsigned int thickness=1)
static void display(const vpImage< unsigned char > &I)
static void displayFrame(const vpImage< unsigned char > &I, const vpHomogeneousMatrix &cMo, const vpCameraParameters &cam, double size, const vpColor &color=vpColor::none, unsigned int thickness=1, const vpImagePoint &offset=vpImagePoint(0, 0), const std::string &frameName="", const vpColor &textColor=vpColor::black, const vpImagePoint &textOffset=vpImagePoint(15, 15))
static void displayCross(const vpImage< unsigned char > &I, const vpImagePoint &ip, unsigned int size, const vpColor &color, unsigned int thickness=1)
static void setTitle(const vpImage< unsigned char > &I, const std::string &windowtitle)
static void flush(const vpImage< unsigned char > &I)
static void displayText(const vpImage< unsigned char > &I, const vpImagePoint &ip, const std::string &s, const vpColor &color)
error that can be emitted by ViSP classes.
Implementation of an homogeneous matrix and operations on such kind of matrices.
static void read(vpImage< unsigned char > &I, const std::string &filename, int backend=IO_DEFAULT_BACKEND)
Class that defines a 2D point in an image. This class is useful for image processing and stores only ...
Definition of the vpImage class member functions.
Class that allows keypoints 2D features detection (and descriptors extraction) and matching thanks to...
unsigned int buildReference(const vpImage< unsigned char > &I) VP_OVERRIDE
void detect(const vpImage< unsigned char > &I, std::vector< cv::KeyPoint > &keyPoints, const vpRect &rectangle=vpRect())
static void compute3DForPointsInPolygons(const vpHomogeneousMatrix &cMo, const vpCameraParameters &cam, std::vector< cv::KeyPoint > &candidates, const std::vector< vpPolygon > &polygons, const std::vector< std::vector< vpPoint > > &roisPt, std::vector< cv::Point3f > &points, cv::Mat *descriptors=nullptr)
void saveLearningData(const std::string &filename, bool binaryMode=false, bool saveTrainingImages=true)
cv::Ptr< cv::FeatureDetector > getDetector(const vpFeatureDetectorType &type) const
static double rad(double deg)
Make the complete tracking of an object by using its CAD model.
void setMu1(const double &mu_1)
void setRange(const unsigned int &range)
void setLikelihoodThresholdType(const vpLikelihoodThresholdType likelihood_threshold_type)
void setNbTotalSample(const int &ntotal_sample)
void setMaskNumber(const unsigned int &mask_number)
void setThreshold(const double &threshold)
void setSampleStep(const double &sample_step)
void setMaskSize(const unsigned int &mask_size)
void setMu2(const double &mu_2)
Class that enables to manipulate easily a video file or a sequence of images. As it inherits from the...
void open(vpImage< vpRGBa > &I) VP_OVERRIDE
void setFileName(const std::string &filename)
void acquire(vpImage< vpRGBa > &I) VP_OVERRIDE
std::shared_ptr< vpDisplay > createDisplay()
Return a smart pointer vpDisplay specialization if a GUI library is available or nullptr otherwise.
vpDisplay * allocateDisplay()
Return a newly allocated vpDisplay specialization if a GUI library is available or nullptr otherwise.