nerian_stereo_node_base.cpp
Go to the documentation of this file.
1 /*******************************************************************************
2  * Copyright (c) 2022 Nerian Vision GmbH
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a copy
5  * of this software and associated documentation files (the "Software"), to deal
6  * in the Software without restriction, including without limitation the rights
7  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
8  * copies of the Software, and to permit persons to whom the Software is
9  * furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice shall be included in
12  * all copies or substantial portions of the Software.
13  *******************************************************************************/
14 
16 
17 namespace nerian_stereo {
18 
19 void StereoNodeBase::dynamicReconfigureCallback(nerian_stereo::NerianStereoConfig &config, uint32_t level) {
21  ROS_INFO("Received a new configuration via dynamic_reconfigure");
22  // Unfortunately, we have to check for each potential change (no configuration deltas provided).
23  // This is done in the autogenerated external code.
25  } else {
26  initialConfigReceived = true;
27  }
28  lastKnownConfig = config;
29 }
30 
31 void StereoNodeBase::updateParameterServerFromDevice(param::ParameterSet& cfg) {
32  // Publish the current config to the parameter server
34  // Publish reboot flag to definitely be set to false in the parameter server
35  getNH().setParam("/nerian_stereo/reboot", false);
36 }
37 
40 }
41 /*
42  * \brief Initialize and publish configuration with a dynamic_reconfigure server
43  */
45  // Connect to parameter server on device
46  ROS_INFO("Connecting to %s for parameter service", remoteHost.c_str());
47  try {
48  deviceParameters.reset(new DeviceParameters(remoteHost.c_str()));
49  } catch(visiontransfer::ParameterException& e) {
50  ROS_ERROR("ParameterException while connecting to parameter service: %s", e.what());
51  throw;
52  }
53  param::ParameterSet ssParams;
54  try {
55  ssParams = deviceParameters->getParameterSet();
56  } catch(visiontransfer::TransferException& e) {
57  ROS_ERROR("TransferException while obtaining parameter set: %s", e.what());
58  throw;
59  } catch(visiontransfer::ParameterException& e) {
60  ROS_ERROR("ParameterException while obtaining parameter set: %s", e.what());
61  throw;
62  }
63  // First make sure that the parameter server gets all *current* values
65  // Initialize (and publish) initial configuration from compile-time generated header
66  dynReconfServer.reset(new dynamic_reconfigure::Server<nerian_stereo::NerianStereoConfig>());
67  // Obtain and publish the default, min, and max values from the device to dyn_reconf
69  // Callback for future changes requested from the ROS side
70  dynReconfServer->setCallback(boost::bind(&StereoNodeBase::dynamicReconfigureCallback, this, _1, _2));
71 }
72 
77  ros::NodeHandle& privateNh = getPrivateNH();
78 
79  // Read all ROS parameters
80  std::string intensityChannel = "mono8";
81  privateNh.getParam("point_cloud_intensity_channel", intensityChannel);
82  if(intensityChannel == "none") {
84  } else if(intensityChannel == "rgb8") {
86  } else if(intensityChannel == "rgb32f") {
88  } else {
90  }
91 
92  if (!privateNh.getParam("color_code_disparity_map", colorCodeDispMap)) {
93  colorCodeDispMap = "";
94  }
95 
96  if (!privateNh.getParam("color_code_legend", colorCodeLegend)) {
97  colorCodeLegend = false;
98  }
99 
100  if (!privateNh.getParam("top_level_frame", frame)) {
101  if (!privateNh.getParam("frame", frame)) {
102  frame = "world";
103  }
104  }
105 
106  if (!privateNh.getParam("internal_frame", internalFrame)) {
107  internalFrame = "nerian_stereo";
108  }
109 
110  if (!privateNh.getParam("remote_port", remotePort)) {
111  remotePort = "7681";
112  }
113 
114  if (!privateNh.getParam("remote_host", remoteHost)) {
115  remoteHost = "0.0.0.0";
116  }
117 
118  if (!privateNh.getParam("use_tcp", useTcp)) {
119  useTcp = false;
120  }
121 
122  if (!privateNh.getParam("ros_coordinate_system", rosCoordinateSystem)) {
123  rosCoordinateSystem = true;
124  }
125 
126  if (!privateNh.getParam("ros_timestamps", rosTimestamps)) {
127  rosTimestamps = true;
128  }
129 
130  if (!privateNh.getParam("calibration_file", calibFile)) {
131  calibFile = "";
132  }
133 
134  if (!privateNh.getParam("delay_execution", execDelay)) {
135  execDelay = 0;
136  }
137 
138  if (!privateNh.getParam("max_depth", maxDepth)) {
139  maxDepth = -1;
140  }
141 
142  if (!privateNh.getParam("q_from_calib_file", useQFromCalibFile)) {
143  useQFromCalibFile = false;
144  }
145 
146  // Apply an initial delay if configured
148 
149  // Create publishers
150  disparityPublisher.reset(new ros::Publisher(getNH().advertise<sensor_msgs::Image>(
151  "/nerian_stereo/disparity_map", 5)));
152  leftImagePublisher.reset(new ros::Publisher(getNH().advertise<sensor_msgs::Image>(
153  "/nerian_stereo/left_image", 5)));
154  rightImagePublisher.reset(new ros::Publisher(getNH().advertise<sensor_msgs::Image>(
155  "/nerian_stereo/right_image", 5)));
156  thirdImagePublisher.reset(new ros::Publisher(getNH().advertise<sensor_msgs::Image>(
157  "/nerian_stereo/color_image", 5)));
158 
160 
161  cameraInfoPublisher.reset(new ros::Publisher(getNH().advertise<nerian_stereo::StereoCameraInfo>(
162  "/nerian_stereo/stereo_camera_info", 1)));
163  cloudPublisher.reset(new ros::Publisher(getNH().advertise<sensor_msgs::PointCloud2>(
164  "/nerian_stereo/point_cloud", 5)));
165 
167  currentTransform.header.stamp = ros::Time::now();
168  currentTransform.header.frame_id = frame;
169  currentTransform.child_frame_id = internalFrame;
170  currentTransform.transform.translation.x = 0.0;
171  currentTransform.transform.translation.y = 0.0;
172  currentTransform.transform.translation.z = 0.0;
173  currentTransform.transform.rotation.x = 0.0;
174  currentTransform.transform.rotation.y = 0.0;
175  currentTransform.transform.rotation.z = 0.0;
176  currentTransform.transform.rotation.w = 1.0;
177 
178 }
179 
181  dataChannelService.reset(new DataChannelService(remoteHost.c_str()));
182 }
183 
185  ROS_INFO("Connecting to %s:%s for data transfer", remoteHost.c_str(), remotePort.c_str());
186  asyncTransfer.reset(new AsyncTransfer(remoteHost.c_str(), remotePort.c_str(),
187  useTcp ? ImageProtocol::PROTOCOL_TCP : ImageProtocol::PROTOCOL_UDP));
188 }
189 
191  // Receive image data
192  ImageSet imageSet;
193  if(asyncTransfer->collectReceivedImageSet(imageSet, 0.005)) {
194 
195  // Get time stamp
196  ros::Time stamp;
197  if(rosTimestamps) {
198  stamp = ros::Time::now();
199  } else {
200  int secs = 0, microsecs = 0;
201  imageSet.getTimestamp(secs, microsecs);
202  stamp = ros::Time(secs, microsecs*1000);
203  }
204 
205  bool hasLeft = false, hasRight = false, hasColor = false, hasDisparity = false;
206 
207  // Publish image data messages for all images included in the set
208  if (imageSet.hasImageType(ImageSet::IMAGE_LEFT)) {
209  publishImageMsg(imageSet, imageSet.getIndexOf(ImageSet::IMAGE_LEFT), stamp, false, leftImagePublisher.get());
210  hasLeft = true;
211  }
212  if (imageSet.hasImageType(ImageSet::IMAGE_DISPARITY)) {
213  publishImageMsg(imageSet, imageSet.getIndexOf(ImageSet::IMAGE_DISPARITY), stamp, true, disparityPublisher.get());
214  hasDisparity = true;
215  }
216  if (imageSet.hasImageType(ImageSet::IMAGE_RIGHT)) {
217  publishImageMsg(imageSet, imageSet.getIndexOf(ImageSet::IMAGE_RIGHT), stamp, false, rightImagePublisher.get());
218  hasRight = true;
219  }
220  if (imageSet.hasImageType(ImageSet::IMAGE_COLOR)) {
221  publishImageMsg(imageSet, imageSet.getIndexOf(ImageSet::IMAGE_COLOR), stamp, false, thirdImagePublisher.get());
222  hasColor = true;
223  }
224 
225  // Dump info about currently available topics (this can change when output channels are toggled)
226  if ((frameNum==0) || (hasLeft!=hadLeft) || (hasRight!=hadRight) || (hasColor!=hadColor) || (hasDisparity!=hadDisparity)) {
227  ROS_INFO("Topics currently being served, based on the device \"Output Channels\" settings:");
228  if (hasLeft) ROS_INFO( " /nerian_stereo/left_image");
229  if (hasRight) ROS_INFO(" /nerian_stereo/right_image");
230  if (hasColor) ROS_INFO(" /nerian_stereo/color_image");
231  if (hasDisparity) {
232  ROS_INFO(" /nerian_stereo/disparity_map");
233  ROS_INFO(" /nerian_stereo/point_cloud");
234  } else {
235  ROS_WARN("Disparity channel deactivated on device -> no disparity or point cloud data!");
236  }
237  hadLeft = hasLeft;
238  hadRight = hasRight;
239  hadColor = hasColor;
240  hadDisparity = hasDisparity;
241  }
242 
243  if(cloudPublisher->getNumSubscribers() > 0) {
244  if(recon3d == nullptr) {
245  // First initialize
246  initPointCloud();
247  }
248 
249  publishPointCloudMsg(imageSet, stamp);
250  }
251 
252  if(cameraInfoPublisher != NULL && cameraInfoPublisher->getNumSubscribers() > 0) {
253  publishCameraInfo(stamp, imageSet);
254  }
255 
256  // Display some simple statistics
257  frameNum++;
258  if(stamp.sec != lastLogTime.sec) {
259  if(lastLogTime != ros::Time()) {
260  double dt = (stamp - lastLogTime).toSec();
261  double fps = (frameNum - lastLogFrames) / dt;
262  ROS_INFO("%.1f fps", fps);
263  }
265  lastLogTime = stamp;
266  }
267  }
268 }
269 
271  if(calibFile == "" ) {
272  ROS_WARN("No camera calibration file configured. Cannot publish detailed camera information!");
273  } else {
274  bool success = false;
275  try {
276  if (calibStorage.open(calibFile, cv::FileStorage::READ)) {
277  success = true;
278  }
279  } catch(...) {
280  }
281 
282  if(!success) {
283  ROS_WARN("Error reading calibration file: %s\n"
284  "Cannot publish detailed camera information!", calibFile.c_str());
285  }
286  }
287 }
288 
289 void StereoNodeBase::publishImageMsg(const ImageSet& imageSet, int imageIndex, ros::Time stamp, bool allowColorCode,
290  ros::Publisher* publisher) {
291 
292  if(publisher->getNumSubscribers() <= 0) {
293  return; //No subscribers
294  }
295 
296  cv_bridge::CvImage cvImg;
297  cvImg.header.frame_id = internalFrame;
298  cvImg.header.stamp = stamp;
299  cvImg.header.seq = imageSet.getSequenceNumber(); // Actually ROS will overwrite this
300 
301  bool format12Bit = (imageSet.getPixelFormat(imageIndex) == ImageSet::FORMAT_12_BIT_MONO);
302  string encoding = "";
303  bool ok = true;
304 
305  if(colorCodeDispMap == "" || colorCodeDispMap == "none" || !allowColorCode || !format12Bit) {
306  switch (imageSet.getPixelFormat(imageIndex)) {
307  case ImageSet::FORMAT_8_BIT_RGB: {
308  cv::Mat rgbImg(imageSet.getHeight(), imageSet.getWidth(),
309  CV_8UC3,
310  imageSet.getPixelData(imageIndex), imageSet.getRowStride(imageIndex));
311  cvImg.image = rgbImg;
312  encoding = "rgb8";
313  break;
314  }
315  case ImageSet::FORMAT_8_BIT_MONO:
316  case ImageSet::FORMAT_12_BIT_MONO: {
317  cv::Mat monoImg(imageSet.getHeight(), imageSet.getWidth(),
318  format12Bit ? CV_16UC1 : CV_8UC1,
319  imageSet.getPixelData(imageIndex), imageSet.getRowStride(imageIndex));
320  cvImg.image = monoImg;
321  encoding = (format12Bit ? "mono16": "mono8");
322  break;
323  }
324  default: {
325  ROS_WARN("Omitting an image with unhandled pixel format");
326  ok = false;
327  }
328  }
329  } else {
330  cv::Mat monoImg(imageSet.getHeight(), imageSet.getWidth(),
331  format12Bit ? CV_16UC1 : CV_8UC1,
332  imageSet.getPixelData(imageIndex), imageSet.getRowStride(imageIndex));
333 
334  if(colCoder == NULL) {
335  int dispMin = 0, dispMax = 0;
336  imageSet.getDisparityRange(dispMin, dispMax);
337 
338  colCoder.reset(new ColorCoder(
339  colorCodeDispMap == "rainbow" ? ColorCoder::COLOR_RAINBOW_BGR : ColorCoder::COLOR_RED_BLUE_BGR,
340  dispMin*16, dispMax*16, true, true));
341  if(colorCodeLegend) {
342  // Create the legend
343  colDispMap = colCoder->createLegendBorder(monoImg.cols, monoImg.rows, 1.0/16.0);
344  } else {
345  colDispMap = cv::Mat_<cv::Vec3b>(monoImg.rows, monoImg.cols);
346  }
347  }
348 
349  cv::Mat_<cv::Vec3b> dispSection = colDispMap(cv::Rect(0, 0, monoImg.cols, monoImg.rows));
350 
351  colCoder->codeImage(cv::Mat_<unsigned short>(monoImg), dispSection);
352  cvImg.image = colDispMap;
353  encoding = "bgr8";
354  }
355 
356  if (ok) {
357  sensor_msgs::ImagePtr msg = cvImg.toImageMsg();
358  msg->encoding = encoding;
359  publisher->publish(msg);
360  }
361 }
362 
363 void StereoNodeBase::qMatrixToRosCoords(const float* src, float* dst) {
364  dst[0] = src[8]; dst[1] = src[9];
365  dst[2] = src[10]; dst[3] = src[11];
366 
367  dst[4] = -src[0]; dst[5] = -src[1];
368  dst[6] = -src[2]; dst[7] = -src[3];
369 
370  dst[8] = -src[4]; dst[9] = -src[5];
371  dst[10] = -src[6]; dst[11] = -src[7];
372 
373  dst[12] = src[12]; dst[13] = src[13];
374  dst[14] = src[14]; dst[15] = src[15];
375 }
376 
377 void StereoNodeBase::publishPointCloudMsg(ImageSet& imageSet, ros::Time stamp) {
378  if ((!imageSet.hasImageType(ImageSet::IMAGE_DISPARITY))
379  || (imageSet.getPixelFormat(ImageSet::IMAGE_DISPARITY) != ImageSet::FORMAT_12_BIT_MONO)) {
380  return; // This is not a disparity map
381  }
382 
383  // Set static q matrix if desired
384  if(useQFromCalibFile) {
385  static std::vector<float> q;
386  calibStorage["Q"] >> q;
387  imageSet.setQMatrix(&q[0]);
388  }
389 
390  // Transform Q-matrix if desired
391  float qRos[16];
392  if(rosCoordinateSystem) {
393  qMatrixToRosCoords(imageSet.getQMatrix(), qRos);
394  imageSet.setQMatrix(qRos);
395  }
396 
397  // Get 3D points
398  float* pointMap = nullptr;
399  try {
400  pointMap = recon3d->createPointMap(imageSet, 0);
401  } catch(std::exception& ex) {
402  cerr << "Error creating point cloud: " << ex.what() << endl;
403  return;
404  }
405 
406  // Create message object and set header
407  pointCloudMsg->header.stamp = stamp;
408  pointCloudMsg->header.frame_id = internalFrame;
409  pointCloudMsg->header.seq = imageSet.getSequenceNumber(); // Actually ROS will overwrite this
410 
411  // Copy 3D points
412  if(pointCloudMsg->data.size() != imageSet.getWidth()*imageSet.getHeight()*4*sizeof(float)) {
413  // Allocate buffer
414  pointCloudMsg->data.resize(imageSet.getWidth()*imageSet.getHeight()*4*sizeof(float));
415 
416  // Set basic data
417  pointCloudMsg->width = imageSet.getWidth();
418  pointCloudMsg->height = imageSet.getHeight();
419  pointCloudMsg->is_bigendian = false;
420  pointCloudMsg->point_step = 4*sizeof(float);
421  pointCloudMsg->row_step = imageSet.getWidth() * pointCloudMsg->point_step;
422  pointCloudMsg->is_dense = false;
423  }
424 
425  if(maxDepth < 0) {
426  // Just copy everything
427  memcpy(&pointCloudMsg->data[0], pointMap,
428  imageSet.getWidth()*imageSet.getHeight()*4*sizeof(float));
429  } else {
430  // Only copy points up to maximum depth
431  if(rosCoordinateSystem) {
432  copyPointCloudClamped<0>(pointMap, reinterpret_cast<float*>(&pointCloudMsg->data[0]),
433  imageSet.getWidth()*imageSet.getHeight());
434  } else {
435  copyPointCloudClamped<2>(pointMap, reinterpret_cast<float*>(&pointCloudMsg->data[0]),
436  imageSet.getWidth()*imageSet.getHeight());
437  }
438  }
439 
440  if (imageSet.hasImageType(ImageSet::IMAGE_LEFT) || imageSet.hasImageType(ImageSet::IMAGE_COLOR)) {
441  // Copy intensity values as well (if we received any image data)
442  switch(pointCloudColorMode) {
443  case INTENSITY:
444  copyPointCloudIntensity<INTENSITY>(imageSet);
445  break;
446  case RGB_COMBINED:
447  copyPointCloudIntensity<RGB_COMBINED>(imageSet);
448  break;
449  case RGB_SEPARATE:
450  copyPointCloudIntensity<RGB_SEPARATE>(imageSet);
451  break;
452  case NONE:
453  break;
454  }
455  }
456 
457  cloudPublisher->publish(pointCloudMsg);
458 }
459 
460 template <StereoNodeBase::PointCloudColorMode colorMode> void StereoNodeBase::copyPointCloudIntensity(ImageSet& imageSet) {
461  auto imageIndex = imageSet.hasImageType(ImageSet::IMAGE_COLOR) ? ImageSet::IMAGE_COLOR : ImageSet::IMAGE_LEFT;
462  // Get pointers to the beginning and end of the point cloud
463  unsigned char* cloudStart = &pointCloudMsg->data[0];
464  unsigned char* cloudEnd = &pointCloudMsg->data[0]
465  + imageSet.getWidth()*imageSet.getHeight()*4*sizeof(float);
466 
467  if(imageSet.getPixelFormat(imageIndex) == ImageSet::FORMAT_8_BIT_MONO) {
468  // Get pointer to the current pixel and end of current row
469  unsigned char* imagePtr = imageSet.getPixelData(imageIndex);
470  unsigned char* rowEndPtr = imagePtr + imageSet.getWidth();
471  int rowIncrement = imageSet.getRowStride(imageIndex) - imageSet.getWidth();
472 
473  for(unsigned char* cloudPtr = cloudStart + 3*sizeof(float);
474  cloudPtr < cloudEnd; cloudPtr+= 4*sizeof(float)) {
475  if(colorMode == RGB_SEPARATE) {// RGB as float
476  *reinterpret_cast<float*>(cloudPtr) = static_cast<float>(*imagePtr) / 255.0F;
477  } else if(colorMode == RGB_COMBINED) {// RGB as integer
478  const unsigned char intensity = *imagePtr;
479  *reinterpret_cast<unsigned int*>(cloudPtr) = (intensity << 16) | (intensity << 8) | intensity;
480  } else {
481  *cloudPtr = *imagePtr;
482  }
483 
484  imagePtr++;
485  if(imagePtr == rowEndPtr) {
486  // Progress to next row
487  imagePtr += rowIncrement;
488  rowEndPtr = imagePtr + imageSet.getWidth();
489  }
490  }
491  } else if(imageSet.getPixelFormat(imageIndex) == ImageSet::FORMAT_12_BIT_MONO) {
492  // Get pointer to the current pixel and end of current row
493  unsigned short* imagePtr = reinterpret_cast<unsigned short*>(imageSet.getPixelData(imageIndex));
494  unsigned short* rowEndPtr = imagePtr + imageSet.getWidth();
495  int rowIncrement = imageSet.getRowStride(imageIndex) - 2*imageSet.getWidth();
496 
497  for(unsigned char* cloudPtr = cloudStart + 3*sizeof(float);
498  cloudPtr < cloudEnd; cloudPtr+= 4*sizeof(float)) {
499 
500  if(colorMode == RGB_SEPARATE) {// RGB as float
501  *reinterpret_cast<float*>(cloudPtr) = static_cast<float>(*imagePtr) / 4095.0F;
502  } else if(colorMode == RGB_COMBINED) {// RGB as integer
503  const unsigned char intensity = *imagePtr/16;
504  *reinterpret_cast<unsigned int*>(cloudPtr) = (intensity << 16) | (intensity << 8) | intensity;
505  } else {
506  *cloudPtr = *imagePtr/16;
507  }
508 
509  imagePtr++;
510  if(imagePtr == rowEndPtr) {
511  // Progress to next row
512  imagePtr += rowIncrement;
513  rowEndPtr = imagePtr + imageSet.getWidth();
514  }
515  }
516  } else if(imageSet.getPixelFormat(imageIndex) == ImageSet::FORMAT_8_BIT_RGB) {
517  // Get pointer to the current pixel and end of current row
518  unsigned char* imagePtr = imageSet.getPixelData(imageIndex);
519  unsigned char* rowEndPtr = imagePtr + 3*imageSet.getWidth();
520  int rowIncrement = imageSet.getRowStride(imageIndex) - 3*imageSet.getWidth();
521 
522  static bool warned = false;
523  if(colorMode == RGB_SEPARATE && !warned) {
524  warned = true;
525  ROS_WARN("RGBF32 is not supported for color images. Please use RGB8!");
526  }
527 
528  for(unsigned char* cloudPtr = cloudStart + 3*sizeof(float);
529  cloudPtr < cloudEnd; cloudPtr+= 4*sizeof(float)) {
530  if(colorMode == RGB_SEPARATE) {// RGB as float
531  *reinterpret_cast<float*>(cloudPtr) = static_cast<float>(imagePtr[2]) / 255.0F;
532  } else if(colorMode == RGB_COMBINED) {// RGB as integer
533  *reinterpret_cast<unsigned int*>(cloudPtr) = (imagePtr[0] << 16) | (imagePtr[1] << 8) | imagePtr[2];
534  } else {
535  *cloudPtr = (imagePtr[0] + imagePtr[1]*2 + imagePtr[2])/4;
536  }
537 
538  imagePtr+=3;
539  if(imagePtr == rowEndPtr) {
540  // Progress to next row
541  imagePtr += rowIncrement;
542  rowEndPtr = imagePtr + imageSet.getWidth();
543  }
544  }
545  } else {
546  throw std::runtime_error("Invalid pixel format!");
547  }
548 }
549 
550 template <int coord> void StereoNodeBase::copyPointCloudClamped(float* src, float* dst, int size) {
551  // Only copy points that are below the minimum depth
552  float* endPtr = src + 4*size;
553  for(float *srcPtr = src, *dstPtr = dst; srcPtr < endPtr; srcPtr+=4, dstPtr+=4) {
554  if(srcPtr[coord] > maxDepth) {
555  dstPtr[0] = std::numeric_limits<float>::quiet_NaN();
556  dstPtr[1] = std::numeric_limits<float>::quiet_NaN();
557  dstPtr[2] = std::numeric_limits<float>::quiet_NaN();
558  } else {
559  dstPtr[0] = srcPtr[0];
560  dstPtr[1] = srcPtr[1];
561  dstPtr[2] = srcPtr[2];
562  }
563  }
564 }
565 
567  //ros::NodeHandle privateNh("~"); // RYT TODO check
568 
569  // Initialize 3D reconstruction class
570  recon3d.reset(new Reconstruct3D);
571 
572  // Initialize message
573  pointCloudMsg.reset(new sensor_msgs::PointCloud2);
574 
575  // Set channel information.
576  sensor_msgs::PointField fieldX;
577  fieldX.name ="x";
578  fieldX.offset = 0;
579  fieldX.datatype = sensor_msgs::PointField::FLOAT32;
580  fieldX.count = 1;
581  pointCloudMsg->fields.push_back(fieldX);
582 
583  sensor_msgs::PointField fieldY;
584  fieldY.name ="y";
585  fieldY.offset = sizeof(float);
586  fieldY.datatype = sensor_msgs::PointField::FLOAT32;
587  fieldY.count = 1;
588  pointCloudMsg->fields.push_back(fieldY);
589 
590  sensor_msgs::PointField fieldZ;
591  fieldZ.name ="z";
592  fieldZ.offset = 2*sizeof(float);
593  fieldZ.datatype = sensor_msgs::PointField::FLOAT32;
594  fieldZ.count = 1;
595  pointCloudMsg->fields.push_back(fieldZ);
596 
598  sensor_msgs::PointField fieldI;
599  fieldI.name ="intensity";
600  fieldI.offset = 3*sizeof(float);
601  fieldI.datatype = sensor_msgs::PointField::UINT8;
602  fieldI.count = 1;
603  pointCloudMsg->fields.push_back(fieldI);
604  }
605  else if(pointCloudColorMode == RGB_SEPARATE) {
606  sensor_msgs::PointField fieldRed;
607  fieldRed.name ="r";
608  fieldRed.offset = 3*sizeof(float);
609  fieldRed.datatype = sensor_msgs::PointField::FLOAT32;
610  fieldRed.count = 1;
611  pointCloudMsg->fields.push_back(fieldRed);
612 
613  sensor_msgs::PointField fieldGreen;
614  fieldGreen.name ="g";
615  fieldGreen.offset = 3*sizeof(float);
616  fieldGreen.datatype = sensor_msgs::PointField::FLOAT32;
617  fieldGreen.count = 1;
618  pointCloudMsg->fields.push_back(fieldGreen);
619 
620  sensor_msgs::PointField fieldBlue;
621  fieldBlue.name ="b";
622  fieldBlue.offset = 3*sizeof(float);
623  fieldBlue.datatype = sensor_msgs::PointField::FLOAT32;
624  fieldBlue.count = 1;
625  pointCloudMsg->fields.push_back(fieldBlue);
626  } else if(pointCloudColorMode == RGB_COMBINED) {
627  sensor_msgs::PointField fieldRGB;
628  fieldRGB.name ="rgb";
629  fieldRGB.offset = 3*sizeof(float);
630  fieldRGB.datatype = sensor_msgs::PointField::UINT32;
631  fieldRGB.count = 1;
632  pointCloudMsg->fields.push_back(fieldRGB);
633  }
634 }
635 
636 void StereoNodeBase::publishCameraInfo(ros::Time stamp, const ImageSet& imageSet) {
637  if(camInfoMsg == NULL) {
638  // Initialize the camera info structure
639  camInfoMsg.reset(new nerian_stereo::StereoCameraInfo);
640 
641  camInfoMsg->header.frame_id = internalFrame;
642  camInfoMsg->header.seq = imageSet.getSequenceNumber(); // Actually ROS will overwrite this
643 
644  if(calibFile != "") {
645  std::vector<int> sizeVec;
646  calibStorage["size"] >> sizeVec;
647  if(sizeVec.size() != 2) {
648  std::runtime_error("Calibration file format error!");
649  }
650 
651  camInfoMsg->left_info.header = camInfoMsg->header;
652  camInfoMsg->left_info.width = sizeVec[0];
653  camInfoMsg->left_info.height = sizeVec[1];
654  camInfoMsg->left_info.distortion_model = "plumb_bob";
655  calibStorage["D1"] >> camInfoMsg->left_info.D;
656  readCalibrationArray("M1", camInfoMsg->left_info.K);
657  readCalibrationArray("R1", camInfoMsg->left_info.R);
658  readCalibrationArray("P1", camInfoMsg->left_info.P);
659  camInfoMsg->left_info.binning_x = 1;
660  camInfoMsg->left_info.binning_y = 1;
661  camInfoMsg->left_info.roi.do_rectify = false;
662  camInfoMsg->left_info.roi.height = 0;
663  camInfoMsg->left_info.roi.width = 0;
664  camInfoMsg->left_info.roi.x_offset = 0;
665  camInfoMsg->left_info.roi.y_offset = 0;
666 
667  camInfoMsg->right_info.header = camInfoMsg->header;
668  camInfoMsg->right_info.width = sizeVec[0];
669  camInfoMsg->right_info.height = sizeVec[1];
670  camInfoMsg->right_info.distortion_model = "plumb_bob";
671  calibStorage["D2"] >> camInfoMsg->right_info.D;
672  readCalibrationArray("M2", camInfoMsg->right_info.K);
673  readCalibrationArray("R2", camInfoMsg->right_info.R);
674  readCalibrationArray("P2", camInfoMsg->right_info.P);
675  camInfoMsg->right_info.binning_x = 1;
676  camInfoMsg->right_info.binning_y = 1;
677  camInfoMsg->right_info.roi.do_rectify = false;
678  camInfoMsg->right_info.roi.height = 0;
679  camInfoMsg->right_info.roi.width = 0;
680  camInfoMsg->right_info.roi.x_offset = 0;
681  camInfoMsg->right_info.roi.y_offset = 0;
682 
684  readCalibrationArray("T", camInfoMsg->T_left_right);
685  readCalibrationArray("R", camInfoMsg->R_left_right);
686  }
687  }
688 
689  double dt = (stamp - lastCamInfoPublish).toSec();
690  if(dt > 1.0) {
691  // Rather use the Q-matrix that we received over the network if it is valid
692  const float* qMatrix = imageSet.getQMatrix();
693  if(qMatrix[0] != 0.0) {
694  for(int i=0; i<16; i++) {
695  camInfoMsg->Q[i] = static_cast<double>(qMatrix[i]);
696  }
697  }
698 
699  // Publish once per second
700  camInfoMsg->header.stamp = stamp;
701  camInfoMsg->left_info.header.stamp = stamp;
702  camInfoMsg->right_info.header.stamp = stamp;
704 
705  lastCamInfoPublish = stamp;
706  }
707 }
708 
709 template<class T> void StereoNodeBase::readCalibrationArray(const char* key, T& dest) {
710  std::vector<double> doubleVec;
711  calibStorage[key] >> doubleVec;
712 
713  if(doubleVec.size() != dest.size()) {
714  std::runtime_error("Calibration file format error!");
715  }
716 
717  std::copy(doubleVec.begin(), doubleVec.end(), dest.begin());
718 }
719 
721  auto now = ros::Time::now();
722  if ((now - currentTransform.header.stamp).toSec() < 0.01) {
723  // Limit to 100 Hz transform update frequency
724  return;
725  }
726  if (dataChannelService->imuAvailable()) {
727  // Obtain and publish the most recent orientation
728  TimestampedQuaternion tsq = dataChannelService->imuGetRotationQuaternion();
729  currentTransform.header.stamp = now;
730  if(rosCoordinateSystem) {
731  currentTransform.transform.rotation.x = tsq.x();
732  currentTransform.transform.rotation.y = -tsq.z();
733  currentTransform.transform.rotation.z = tsq.y();
734  } else {
735  currentTransform.transform.rotation.x = tsq.x();
736  currentTransform.transform.rotation.y = tsq.y();
737  currentTransform.transform.rotation.z = tsq.z();
738  }
739  currentTransform.transform.rotation.w = tsq.w();
740 
741  /*
742  // DEBUG: Quaternion->Euler + debug output
743  double roll, pitch, yaw;
744  tf2::Quaternion q(tsq.x(), rosCoordinateSystem?(-tsq.z()):tsq.y(), rosCoordinateSystem?tsq.y():tsq.z(), tsq.w());
745  tf2::Matrix3x3 m(q);
746  m.getRPY(roll, pitch, yaw);
747  std::cout << "Orientation:" << std::setprecision(2) << std::fixed << " Roll " << (180.0*roll/M_PI) << " Pitch " << (180.0*pitch/M_PI) << " Yaw " << (180.0*yaw/M_PI) << std::endl;
748  */
749 
751  } else {
752  // We must periodically republish due to ROS interval constraints
753  /*
754  // DEBUG: Impart a (fake) periodic horizontal swaying motion
755  static double DEBUG_t = 0.0;
756  DEBUG_t += 0.1;
757  tf2::Quaternion q;
758  q.setRPY(0, 0, 0.3*sin(DEBUG_t));
759  currentTransform.header.stamp = ros::Time::now();
760  currentTransform.transform.rotation.x = q.x();
761  currentTransform.transform.rotation.y = q.y();
762  currentTransform.transform.rotation.z = q.z();
763  currentTransform.transform.rotation.w = q.w();
764  */
765  currentTransform.header.stamp = now;
767  }
768 }
769 
771  transformBroadcaster->sendTransform(currentTransform);
772 }
773 
774 } // namespace
virtual ros::NodeHandle & getPrivateNH()=0
void prepareAsyncTransfer()
Connects to the image service to request the stream of image sets.
sensor_msgs::PointCloud2Ptr pointCloudMsg
boost::scoped_ptr< dynamic_reconfigure::Server< nerian_stereo::NerianStereoConfig > > dynReconfServer
boost::scoped_ptr< ColorCoder > colCoder
boost::scoped_ptr< DataChannelService > dataChannelService
void copyPointCloudIntensity(ImageSet &imageSet)
Copies the intensity or RGB data to the point cloud.
virtual ros::NodeHandle & getNH()=0
void loadCameraCalibration()
Loads a camera calibration file if configured.
nerian_stereo::StereoCameraInfoPtr camInfoMsg
boost::scoped_ptr< ros::Publisher > rightImagePublisher
void updateParameterServerFromDevice(param::ParameterSet &cfg)
void updateDynamicReconfigureFromDevice(param::ParameterSet &cfg)
void dynamicReconfigureCallback(nerian_stereo::NerianStereoConfig &config, uint32_t level)
#define ROS_WARN(...)
void autogen_updateDynamicReconfigureFromDevice(param::ParameterSet &cfg)
Auto-generated code to override the dynamic_reconfigure limits and defaults for all parameters...
geometry_msgs::TransformStamped currentTransform
void autogen_updateParameterServerFromDevice(param::ParameterSet &cfg)
Auto-generated code to set initial parameters according to those obtained from the device...
sensor_msgs::ImagePtr toImageMsg() const
void publish(const boost::shared_ptr< M > &message) const
void publishImageMsg(const ImageSet &imageSet, int imageIndex, ros::Time stamp, bool allowColorCode, ros::Publisher *publisher)
Publishes the disparity map as 16-bit grayscale image or color coded RGB image.
boost::scoped_ptr< ros::Publisher > thirdImagePublisher
#define ROS_INFO(...)
bool getParam(const std::string &key, std::string &s) const
ROSCPP_DECL bool ok()
void initPointCloud()
Performs all neccessary initializations for point cloud+ publishing.
void readCalibrationArray(const char *key, T &dest)
Reads a vector from the calibration file to a boost:array.
boost::scoped_ptr< Reconstruct3D > recon3d
boost::scoped_ptr< ros::Publisher > cloudPublisher
boost::scoped_ptr< ros::Publisher > disparityPublisher
boost::scoped_ptr< tf2_ros::TransformBroadcaster > transformBroadcaster
void setParam(const std::string &key, const XmlRpc::XmlRpcValue &v) const
void publishPointCloudMsg(ImageSet &imageSet, ros::Time stamp)
Reconstructs the 3D locations form the disparity map and publishes them as point cloud.
A driver node that receives data from Nerian stereo devices and forwards it to ROS.
boost::scoped_ptr< ros::Publisher > leftImagePublisher
static Time now()
boost::scoped_ptr< DeviceParameters > deviceParameters
void copyPointCloudClamped(float *src, float *dst, int size)
Copies all points in a point cloud that have a depth smaller than maxDepth. Other points are set to N...
bool sleep() const
uint32_t getNumSubscribers() const
void publishCameraInfo(ros::Time stamp, const ImageSet &imageSet)
Publishes the camera info once per second.
void qMatrixToRosCoords(const float *src, float *dst)
Transform Q matrix to match the ROS coordinate system: Swap y/z axis, then swap x/y axis...
#define ROS_ERROR(...)
void init()
Performs general initializations.
void autogen_dynamicReconfigureCallback(nerian_stereo::NerianStereoConfig &config, uint32_t level)
Auto-generated code to check for parameter changes and forward them to the device.
std_msgs::Header header
boost::scoped_ptr< ros::Publisher > cameraInfoPublisher
nerian_stereo::NerianStereoConfig lastKnownConfig
boost::scoped_ptr< AsyncTransfer > asyncTransfer


nerian_stereo
Author(s): Nerian Vision Technologies
autogenerated on Thu Jan 12 2023 03:44:09