SampleMarkerlessDetector.cpp

This is an example that demonstrates the use of FernImageDetector to detect an image as a marker.

#include "CvTestbed.h"
#include "GlutViewer.h"
#include "Shared.h"
using namespace alvar;
using namespace std;
bool init = true;
std::stringstream calibrationFilename;
cv::Mat gray;
bool reset = false;
void videocallback(IplImage *image)
{
bool flip_image = (image->origin?true:false);
if (flip_image) {
cvFlip(image);
image->origin = !image->origin;
}
if (init) {
init = false;
cout << "Loading calibration: " << calibrationFilename.str();
if (fernEstimator.setCalibration(calibrationFilename.str(), image->width, image->height)) {
cout << " [Ok]" << endl;
} else {
fernEstimator.setResolution(image->width, image->height);
cout << " [Fail]" << endl;
}
double p[16];
fernEstimator.camera().GetOpenglProjectionMatrix(p, image->width, image->height);
d.SetScale(10);
gray = cv::Mat(image);
}
if (image->nChannels == 3) {
cv::Mat img = cvarrToMat(image);
cv::cvtColor(img, gray, CV_RGB2GRAY);
}
else {
}
vector<CvPoint2D64f> ipts;
vector<CvPoint3D64f> mpts;
if (test > 0.15 && mpts.size() > 4) {
fernEstimator.calculateFromPointCorrespondences(mpts, ipts);
}
Pose pose = fernEstimator.pose();
if (flip_image) {
cvFlip(image);
image->origin = !image->origin;
}
}
int main(int argc, char *argv[])
{
try {
// Output usage message
std::string filename(argv[0]);
filename = filename.substr(filename.find_last_of('\\') + 1);
std::cout << "SampleMarkerlessDetector" << std::endl;
std::cout << "========================" << std::endl;
std::cout << std::endl;
std::cout << "Description:" << std::endl;
std::cout << " This is an example of how to use the 'FernImageDetector' and" << std::endl;
std::cout << " 'FernPoseEstimator' classes to detect and track an image and" << std::endl;
std::cout << " visualize it using 'GlutViewer'. The classification must first" << std::endl;
std::cout << " be trained with the SampleMarkerlessCreator sample and the" << std::endl;
std::cout << " resulting file passed as an argument to this sample." << std::endl;
std::cout << std::endl;
std::cout << " For optimal results, a high quality USB camera or a Firewire" << std::endl;
std::cout << " camera is necessary. It is also advised to calibrate the camera" << std::endl;
std::cout << " using the SampleCamCalib sample. It should be noted that the size" << std::endl;
std::cout << " of the trained image will affect the optimal distance for detection." << std::endl;
std::cout << std::endl;
std::cout << "Usage:" << std::endl;
std::cout << " " << filename << " filename [device]" << std::endl;
std::cout << std::endl;
std::cout << " filename the filename of classifier (.dat)" << std::endl;
std::cout << " device integer selecting device from enumeration list (default 0)" << std::endl;
std::cout << " highgui capture devices are prefered" << std::endl;
std::cout << std::endl;
std::cout << "Keyboard Shortcuts:" << std::endl;
std::cout << " q: quit" << std::endl;
std::cout << std::endl;
if (argc < 2) {
std::cout << "Filename not specified." << std::endl;
return 0;
}
std::string classifierFilename(argv[1]);
// Initialise GlutViewer and CvTestbed
GlutViewer::Start(argc, argv, 640, 480);
// Enumerate possible capture plugins
if (plugins.size() < 1) {
std::cout << "Could not find any capture plugins." << std::endl;
return 0;
}
// Display capture plugins
std::cout << "Available Plugins: ";
std::cout << std::endl;
// Enumerate possible capture devices
if (devices.size() < 1) {
std::cout << "Could not find any capture devices." << std::endl;
return 0;
}
// Check command line argument for which device to use
int selectedDevice = defaultDevice(devices);
if (argc > 2) {
selectedDevice = atoi(argv[2]);
}
if (selectedDevice >= (int)devices.size()) {
selectedDevice = defaultDevice(devices);
}
// Display capture devices
std::cout << "Enumerated Capture Devices:" << std::endl;
outputEnumeratedDevices(devices, selectedDevice);
std::cout << std::endl;
// Create capture object from camera
Capture *cap = CaptureFactory::instance()->createCapture(devices[selectedDevice]);
std::string uniqueName = devices[selectedDevice].uniqueName();
// Handle capture lifecycle and start video capture
// Note that loadSettings/saveSettings are not supported by all plugins
if (cap) {
std::cout << "Loading classifier." << std::endl;
if (!fernDetector.read(classifierFilename)) {
std::cout << "Loading classifier failed." << std::endl;
cap->stop();
delete cap;
return 1;
}
std::stringstream settingsFilename;
settingsFilename << "camera_settings_" << uniqueName << ".xml";
calibrationFilename << "camera_calibration_" << uniqueName << ".xml";
cap->start();
cap->setResolution(640, 480);
if (cap->loadSettings(settingsFilename.str())) {
std::cout << "Loading settings: " << settingsFilename.str() << std::endl;
}
std::stringstream title;
title << "SampleMarkerDetector (" << cap->captureDevice().captureType() << ")";
CvTestbed::Instance().StartVideo(cap, title.str().c_str());
if (cap->saveSettings(settingsFilename.str())) {
std::cout << "Saving settings: " << settingsFilename.str() << std::endl;
}
cap->stop();
delete cap;
}
else if (CvTestbed::Instance().StartVideo(0, argv[0])) {
}
else {
std::cout << "Could not initialize the selected capture backend." << std::endl;
}
return 0;
}
catch (const std::exception &e) {
std::cout << "Exception: " << e.what() << endl;
}
catch (...) {
std::cout << "Exception: unknown" << std::endl;
}
}


ar_track_alvar
Author(s): Scott Niekum
autogenerated on Thu Jun 6 2019 19:27:23