SamplePointcloud.cpp

This is an example showing how to use SimpleSfM for tracking the environment using features in addition to MultiMarker.

#include "CvTestbed.h"
#include "GlutViewer.h"
#include "SfM.h"
#include "Shared.h"
using namespace alvar;
using namespace std;
const double marker_size=1;
bool init=true;
std::stringstream calibrationFilename;
// Own drawable 3D cross on features
struct OwnDrawable : public Drawable {
virtual void Draw() {
const double scale = 0.2;
glPushMatrix();
glMultMatrixd(gl_mat);
glColor3d(color[0], color[1], color[2]);
glBegin(GL_LINES);
glVertex3f(0.0, 0.0, -scale);
glVertex3f(0.0, 0.0, scale);
glVertex3f(0.0, -scale, 0.0);
glVertex3f(0.0, scale, 0.0);
glVertex3f(-scale, 0.0, 0.0);
glVertex3f(scale, 0.0, 0.0);
glEnd();
glPopMatrix();
}
};
bool reset=false;
void videocallback(IplImage *image)
{
static IplImage *rgb = 0;
static IplImage* bg_image = 0;
bool flip_image = (image->origin?true:false);
if (flip_image) {
cvFlip(image);
image->origin = !image->origin;
}
if (init) {
init = false;
sfm->Clear();
cout<<"Loading calibration: "<<calibrationFilename.str();
if (sfm->GetCamera()->SetCalib(calibrationFilename.str().c_str(), image->width, image->height)) {
cout<<" [Ok]"<<endl;
} else {
sfm->GetCamera()->SetRes(image->width, image->height);
cout<<" [Fail]"<<endl;
}
double p[16];
sfm->GetCamera()->GetOpenglProjectionMatrix(p,image->width,image->height);
d_marker.SetScale(marker_size*2);
rgb = CvTestbed::Instance().CreateImageWithProto("RGB", image, 0, 3);
bg_image = CvTestbed::Instance().CreateImage("BG texture", cvSize(512,512),8, 3);
bg_image->origin = 0;
sfm->SetScale(10);
if (sfm->AddMultiMarker("mmarker.xml", FILE_FORMAT_XML)) {
std::cout<<"Using MultiMarker defined in mmarker.xml."<<std::endl;
} else {
std::cout<<"Couldn't load mmarker.xml. Using default 'SampleMultiMarker' setup."<<std::endl;
Pose pose;
pose.Reset();
sfm->AddMarker(0, marker_size*2, pose);
pose.SetTranslation(-marker_size*2.5, +marker_size*1.5, 0);
sfm->AddMarker(1, marker_size, pose);
pose.SetTranslation(+marker_size*2.5, +marker_size*1.5, 0);
sfm->AddMarker(2, marker_size, pose);
pose.SetTranslation(-marker_size*2.5, -marker_size*1.5, 0);
sfm->AddMarker(3, marker_size, pose);
pose.SetTranslation(+marker_size*2.5, -marker_size*1.5, 0);
sfm->AddMarker(4, marker_size, pose);
}
sfm->SetResetPoint();
}
if (reset) {
sfm->Reset();
reset = false;
}
//if (sfm->UpdateRotationsOnly(image)) {
//if (sfm->UpdateTriangulateOnly(image)) {
if (sfm->Update(image, false, true, 7.f, 15.f)) {
// Draw the camera (The GlutViewer has little weirdness here...)q
Pose pose = *(sfm->GetPose());
double gl[16];
pose.GetMatrixGL(gl, true);
pose.Invert();
pose.GetMatrixGL(d_marker.gl_mat, false);
// Draw features
std::map<int, SimpleSfM::Feature>::iterator iter;
iter = sfm->container.begin();
for(;iter != sfm->container.end(); iter++) {
if (sfm->container_triangulated.find(iter->first) != sfm->container_triangulated.end()) continue;
if (iter->second.has_p3d)
{
if (own_drawable_count < 1000) {
memset(d_points[own_drawable_count].gl_mat, 0, 16*sizeof(double));
d_points[own_drawable_count].gl_mat[0] = 1;
d_points[own_drawable_count].gl_mat[5] = 1;
d_points[own_drawable_count].gl_mat[10] = 1;
d_points[own_drawable_count].gl_mat[15] = 1;
d_points[own_drawable_count].gl_mat[12] = iter->second.p3d.x;
d_points[own_drawable_count].gl_mat[13] = iter->second.p3d.y;
d_points[own_drawable_count].gl_mat[14] = iter->second.p3d.z;
if (iter->second.type_id == 0) d_points[own_drawable_count].SetColor(1,0,0);
else d_points[own_drawable_count].SetColor(0,1,0);
own_drawable_count++;
}
}
}
// Draw triangulated features
iter = sfm->container_triangulated.begin();
for(;iter != sfm->container_triangulated.end(); iter++) {
if (iter->second.has_p3d)
{
if (own_drawable_count < 1000) {
memset(d_points[own_drawable_count].gl_mat, 0, 16*sizeof(double));
d_points[own_drawable_count].gl_mat[0] = 1;
d_points[own_drawable_count].gl_mat[5] = 1;
d_points[own_drawable_count].gl_mat[10] = 1;
d_points[own_drawable_count].gl_mat[15] = 1;
d_points[own_drawable_count].gl_mat[12] = iter->second.p3d.x;
d_points[own_drawable_count].gl_mat[13] = iter->second.p3d.y;
d_points[own_drawable_count].gl_mat[14] = iter->second.p3d.z;
/*if (iter->second.type_id == 0) d_points[own_drawable_count].SetColor(1,0,1);
else*/ d_points[own_drawable_count].SetColor(0,0,1);
own_drawable_count++;
}
}
}
}
if (image->nChannels == 1) cvCvtColor(image, rgb, CV_GRAY2RGB);
else if (image->nChannels == 3) cvCopy(image, rgb);
// Draw video on GlutViewer background
cvResize(rgb, bg_image);
// Draw debug info to the rgb
sfm->Draw(rgb);
if (flip_image) {
cvFlip(image);
image->origin = !image->origin;
}
}
int keycallback(int key)
{
if(key == 'r')
{
reset = true;
}
else return key;
return 0;
}
int main(int argc, char *argv[])
{
try {
// Output usage message
std::string filename(argv[0]);
filename = filename.substr(filename.find_last_of('\\') + 1);
std::cout << "SamplePointcloud" << std::endl;
std::cout << "================" << std::endl;
std::cout << std::endl;
std::cout << "Description:" << std::endl;
std::cout << " This example shows simple structure from motion approach that can be "<< std::endl;
std::cout << " used to track environment beyond an multimarker setup. To get this "<< std::endl;
std::cout << " example work properly be sure to calibrate your camera and tune it "<< std::endl;
std::cout << " to have fast framerate without motion blur. "<< std::endl;
std::cout << std::endl;
std::cout << " There are two possible approaches Update() and UpdateRotationsOnly()."<< std::endl;
std::cout << " By default the Update() is used but you can easily uncomment the "<< std::endl;
std::cout << " other one if needed."<< std::endl;
std::cout << std::endl;
std::cout << "Usage:" << std::endl;
std::cout << " " << filename << " [device]" << std::endl;
std::cout << std::endl;
std::cout << " device integer selecting device from enumeration list (default 0)" << std::endl;
std::cout << " highgui capture devices are prefered" << std::endl;
std::cout << std::endl;
std::cout << "Keyboard Shortcuts:" << std::endl;
std::cout << " r: reset" << std::endl;
std::cout << " q: quit" << std::endl;
std::cout << std::endl;
// Initialise GlutViewer and CvTestbed
GlutViewer::Start(argc, argv, 640, 480, 100);
sfm = new SimpleSfM();
// Enumerate possible capture plugins
if (plugins.size() < 1) {
std::cout << "Could not find any capture plugins." << std::endl;
return 0;
}
// Display capture plugins
std::cout << "Available Plugins: ";
std::cout << std::endl;
// Enumerate possible capture devices
if (devices.size() < 1) {
std::cout << "Could not find any capture devices." << std::endl;
return 0;
}
// Check command line argument for which device to use
int selectedDevice = defaultDevice(devices);
if (argc > 1) {
selectedDevice = atoi(argv[1]);
}
if (selectedDevice >= (int)devices.size()) {
selectedDevice = defaultDevice(devices);
}
// Display capture devices
std::cout << "Enumerated Capture Devices:" << std::endl;
outputEnumeratedDevices(devices, selectedDevice);
std::cout << std::endl;
// Create capture object from camera
Capture *cap = CaptureFactory::instance()->createCapture(devices[selectedDevice]);
std::string uniqueName = devices[selectedDevice].uniqueName();
// Handle capture lifecycle and start video capture
// Note that loadSettings/saveSettings are not supported by all plugins
if (cap) {
std::stringstream settingsFilename;
settingsFilename << "camera_settings_" << uniqueName << ".xml";
calibrationFilename << "camera_calibration_" << uniqueName << ".xml";
cap->start();
cap->setResolution(640, 480);
if (cap->loadSettings(settingsFilename.str())) {
std::cout << "Loading settings: " << settingsFilename.str() << std::endl;
}
std::stringstream title;
title << "SamplePointcloud (" << cap->captureDevice().captureType() << ")";
CvTestbed::Instance().StartVideo(cap, title.str().c_str());
if (cap->saveSettings(settingsFilename.str())) {
std::cout << "Saving settings: " << settingsFilename.str() << std::endl;
}
cap->stop();
delete cap; cap = NULL;
}
else if (CvTestbed::Instance().StartVideo(0, argv[0])) {
}
else {
std::cout << "Could not initialize the selected capture backend." << std::endl;
}
delete sfm; sfm = NULL;
return 0;
}
catch (const std::exception &e) {
std::cout << "Exception: " << e.what() << endl;
}
catch (...) {
std::cout << "Exception: unknown" << std::endl;
}
}


ar_track_alvar
Author(s): Scott Niekum
autogenerated on Thu Jun 6 2019 19:27:23