alb_helper.cpp
Go to the documentation of this file.
1 // Standard headers
2 #include <algorithm>
3 #include <map>
4 
5 // ROS headers
6 #include "ros/ros.h"
8 
9 // Osef headers
10 #include "osefTypes.h"
11 #include "typeCaster.h"
12 
13 // Local headers
14 #include "alb_common.h"
15 #include "alb_helper.h"
16 
18 
19 // Constant definition.
20 const std::map<int, std::string> k_object_class = { { 0, "UNKNOWN" }, { 1, "PERSON" }, { 2, "LUGGAGE" },
21  { 3, "TROLLEY" }, { 4, "TRUCK" }, { 5, "BUS" },
22  { 6, "CAR" }, { 7, "VAN" }, { 8, "TWO_WHEELER" },
23  { 9, "MASK" }, { 10, "NO_MASK" } };
24 
25 namespace Helper
26 {
27 void define_point_fields(sensor_msgs::PointCloud2 &pointCloud)
28 {
29  sensor_msgs::PointField x_field;
30  x_field.name = "x";
31  x_field.offset = 0;
32  x_field.datatype = sensor_msgs::PointField::FLOAT32;
33  x_field.count = 1;
34 
35  sensor_msgs::PointField y_field;
36  y_field.name = "y";
37  y_field.offset = 4;
38  y_field.datatype = sensor_msgs::PointField::FLOAT32;
39  y_field.count = 1;
40 
41  sensor_msgs::PointField z_field;
42  z_field.name = "z";
43  z_field.offset = 8;
44  z_field.datatype = sensor_msgs::PointField::FLOAT32;
45  z_field.count = 1;
46 
47  pointCloud.fields.push_back(x_field);
48  pointCloud.fields.push_back(y_field);
49  pointCloud.fields.push_back(z_field);
50 }
51 
52 void define_points_data(sensor_msgs::PointCloud2 &pointCloud, const uint32_t layers, const uint32_t points,
53  float *pointData, bool use_colwise_order)
54 {
55  // Height and width define the number of points ("width" points in "height" dimension);
56  if (!layers) {
57  pointCloud.height = 1;
58  pointCloud.width = points;
59  } else {
60  pointCloud.height = layers;
61  pointCloud.width = points / layers;
62  }
63 
64  // Steps define the data storage.
65  constexpr size_t point_size = 3 * sizeof(float); // Size of a point in bytes
66  pointCloud.point_step = point_size;
67  pointCloud.row_step = pointCloud.width * pointCloud.point_step;
68 
69  // Point data have to be reinterpreted into uint8_t.
70  const uint8_t *point_data = reinterpret_cast<uint8_t *>(pointData);
71  const size_t total_size = point_size * points;
72  if (use_colwise_order || !layers) {
73  pointCloud.data.assign(point_data, point_data + total_size);
74  } else {
75  pointCloud.data.resize(total_size);
76  size_t colwise_idx = 0;
77  for (uint32_t w = 0; w < pointCloud.width; ++w) {
78  for (uint32_t h = 0; h < layers; ++h, colwise_idx += point_size) {
79  size_t rowwise_idx = (h * pointCloud.width + w) * point_size;
80  std::copy(&point_data[colwise_idx], &point_data[colwise_idx] + point_size,
81  &(pointCloud.data[rowwise_idx]));
82  }
83  }
84  }
85  pointCloud.is_dense = true;
86 }
87 
88 void define_pose(geometry_msgs::Pose &pose, const std::array<float, 3> &position, const std::array<float, 9> &rotation)
89 {
90  pose.position.x = position[0];
91  pose.position.y = position[1];
92  pose.position.z = position[2];
93 
94  // Convert rotation matrix to quaternion orientation. (Stuelpnagel 1964)
95  float t = rotation[0] + rotation[4] + rotation[8];
96 
97  if (1 + t > 0.0f) {
98  float r = sqrt(1 + t);
99  float s = 1 / (2 * r);
100 
101  pose.orientation.x = (rotation[5] - rotation[7]) * s;
102  pose.orientation.y = (rotation[6] - rotation[2]) * s;
103  pose.orientation.z = (rotation[1] - rotation[3]) * s;
104  pose.orientation.w = r / 2;
105  } else if (rotation[0] > rotation[4] && rotation[0] > rotation[8]) {
106  float r = sqrt(1.0f + rotation[0] - rotation[4] - rotation[8]);
107  float s = 1 / (2 * r);
108  pose.orientation.x = r / 2;
109  pose.orientation.y = (rotation[1] + rotation[3]) * s;
110  pose.orientation.z = (rotation[2] + rotation[6]) * s;
111  pose.orientation.w = (rotation[5] - rotation[7]) * s;
112  } else if (rotation[4] > rotation[8]) {
113  float r = sqrt(1.0f + rotation[4] - rotation[0] - rotation[8]);
114  float s = 1 / (2 * r);
115  pose.orientation.x = (rotation[1] + rotation[3]) * s;
116  pose.orientation.y = r / 2;
117  pose.orientation.z = (rotation[5] + rotation[7]) * s;
118  pose.orientation.w = (rotation[6] - rotation[2]) * s;
119  } else {
120  float r = sqrt(1.0f + rotation[8] - rotation[0] - rotation[4]);
121  float s = 1 / (2 * r);
122  pose.orientation.x = (rotation[6] + rotation[2]) * s;
123  pose.orientation.y = (rotation[5] + rotation[7]) * s;
124  pose.orientation.z = r / 2;
125  pose.orientation.w = (rotation[1] - rotation[3]) * s;
126  }
127 
128  // Normalize the quaternion.
129  tf2::Quaternion current_quat;
130  tf2::fromMsg(pose.orientation, current_quat);
131  current_quat.normalize();
132  pose.orientation = tf2::toMsg(current_quat);
133 }
134 
135 void init_pose(geometry_msgs::Pose &pose)
136 {
137  pose.position.x = 0.0f;
138  pose.position.y = 0.0f;
139  pose.position.z = 0.0f;
140  pose.orientation.x = 0.0f;
141  pose.orientation.y = 0.0f;
142  pose.orientation.z = 0.0f;
143  pose.orientation.w = 1.0f;
144 }
145 
146 void parse_zone_data(const std::string &raw_zones, outsight_alb_driver::Zones &zones_msg)
147 {
148  const json parsed_zones = json::parse(raw_zones);
149 
150  for (auto zone_it = parsed_zones[ALBCommon::k_zone_config_zones].begin();
151  zone_it != parsed_zones[ALBCommon::k_zone_config_zones].end(); ++zone_it) {
152  if ((*zone_it)[ALBCommon::k_zone_config_role].get<std::string>() == ALBCommon::k_zone_config_event) {
153  outsight_alb_driver::ZoneData zone;
154  zone.name = (*zone_it)[ALBCommon::k_zone_config_name].get<std::string>();
155  zone.id = (*zone_it)[ALBCommon::k_zone_config_id].get<std::string>();
156 
157  for (auto pt_it = (*zone_it)[ALBCommon::k_zone_config_points].begin();
158  pt_it != (*zone_it)[ALBCommon::k_zone_config_points].end(); pt_it++) {
159  geometry_msgs::Point point;
160  point.x = (*pt_it)[0];
161  point.y = (*pt_it)[1];
162  zone.points.push_back(point);
163  }
164 
165  zone.role = (*zone_it)[ALBCommon::k_zone_config_role].get<std::string>();
166  zones_msg.zones.push_back(zone);
167  }
168  }
169 }
170 
171 bool get_pose_from_tlv(const Tlv::tlv_s *frame, geometry_msgs::PoseStamped &pose)
172 {
173  Tlv::Parser parser(frame->getValue(), frame->getLength());
174  const Tlv::tlv_s *frameTlv = parser.findTlv(OSEF_TYPE_SCAN_FRAME);
175  if (!frameTlv) {
176  ROS_ERROR("[AlbHelper] Scan frame not found, unable to publish PoseStamped message.");
177  return false;
178  }
179 
180  Tlv::Parser frameParser(frameTlv->getValue(), frameTlv->getLength());
181  const Tlv::tlv_s *pose_tlv = frameParser.findTlv(OSEF_TYPE_POSE);
182  if (!pose_tlv) {
183  ROS_ERROR("[AlbHelper] Pose not found, unable to publish PoseStamped message.");
184  return false;
185  }
186 
187  std::array<float, 3> pose_translation(Pose::castTranslationValue(pose_tlv));
188  std::array<float, 9> pose_rotation(Pose::castRotationValue(pose_tlv));
189 
190  define_pose(pose.pose, pose_translation, pose_rotation);
191 
192  return true;
193 }
194 
195 bool get_egomotion_from_tlv(const Tlv::tlv_s *frame, geometry_msgs::PoseStamped &pose)
196 {
197  Tlv::Parser parser(frame->getValue(), frame->getLength());
198  const Tlv::tlv_s *frameTlv = parser.findTlv(OSEF_TYPE_SCAN_FRAME);
199  if (!frameTlv) {
200  ROS_ERROR("[AlbHelper] Scan frame not found, unable to publish PoseStamped message.");
201  return false;
202  }
203 
204  Tlv::Parser frameParser(frameTlv->getValue(), frameTlv->getLength());
205  const Tlv::tlv_s *egomotion_tlv = frameParser.findTlv(OSEF_TYPE_EGO_MOTION);
206  if (!egomotion_tlv) {
207  ROS_ERROR("[AlbHelper] Egomotion not found, unable to publish PoseStamped message.");
208  return false;
209  }
210 
211  Tlv::Parser egomotionParser(egomotion_tlv->getValue(), egomotion_tlv->getLength());
212  const Tlv::tlv_s *pose_tlv = egomotionParser.findTlv(OSEF_TYPE_POSE_RELATIVE);
213  if (!pose_tlv) {
214  ROS_ERROR("[AlbHelper] Pose not found, unable to publish PoseStamped message.");
215  return false;
216  }
217 
218  std::array<float, 3> pose_translation(Pose::castTranslationValue(pose_tlv));
219  std::array<float, 9> pose_rotation(Pose::castRotationValue(pose_tlv));
220 
221  define_pose(pose.pose, pose_translation, pose_rotation);
222 
223  return true;
224 }
225 
226 void computeOdometry(nav_msgs::Odometry &odom, const geometry_msgs::Pose &current_pose,
227  const geometry_msgs::Pose &last_pose, const geometry_msgs::Pose &first_pose, float dt)
228 {
229  // Compute Pose displacement between first pose and current pose.
230  geometry_msgs::PoseStamped global_pose;
231  geometry_msgs::PoseStamped global_pose_from_first;
232  global_pose.pose.position.x = current_pose.position.x - first_pose.position.x;
233  global_pose.pose.position.y = current_pose.position.y - first_pose.position.y;
234  global_pose.pose.position.z = current_pose.position.z - first_pose.position.z;
235 
236  // Define transform from first pose.
237  geometry_msgs::TransformStamped local_transform;
238  local_transform.transform.rotation = first_pose.orientation;
239  local_transform.transform.rotation.w = -local_transform.transform.rotation.w;
240 
241  tf2::doTransform(global_pose, global_pose_from_first, local_transform);
242 
243  // Odometry is the pose displacement.
244  odom.pose.pose.position.x = global_pose_from_first.pose.position.x;
245  odom.pose.pose.position.y = global_pose_from_first.pose.position.y;
246  odom.pose.pose.position.z = global_pose_from_first.pose.position.z;
247 
248  // Compute rotation between first and current pose.
249  tf2::Quaternion current_orientation;
250  tf2::Quaternion first_orientation;
251  tf2::Quaternion odom_orientation;
252  tf2::fromMsg(current_pose.orientation, current_orientation);
253  tf2::fromMsg(first_pose.orientation, first_orientation);
254  odom_orientation = current_orientation * first_orientation.inverse();
255  odom_orientation.normalize();
256  odom.pose.pose.orientation = tf2::toMsg(odom_orientation);
257 
258  // Compute displacement between the current and last frame in the fixed frame
259  geometry_msgs::PoseStamped local_pose;
260  geometry_msgs::PoseStamped local_pose_from_first;
261  local_pose.pose.position.x = current_pose.position.x - last_pose.position.x;
262  local_pose.pose.position.y = current_pose.position.y - last_pose.position.y;
263  local_pose.pose.position.z = current_pose.position.z - last_pose.position.z;
264 
265  // Compute the linear instant speed in the fixed frame
266  tf2::doTransform(local_pose, local_pose_from_first, local_transform);
267 
268  odom.twist.twist.linear.x = local_pose_from_first.pose.position.x / dt;
269  odom.twist.twist.linear.y = local_pose_from_first.pose.position.y / dt;
270  odom.twist.twist.linear.z = local_pose_from_first.pose.position.z / dt;
271 
272  // Compute the angular instant speed
273  tf2::Quaternion egomotion_orientation;
274  tf2::Quaternion last_orientation;
275  tf2::fromMsg(last_pose.orientation, last_orientation);
276  egomotion_orientation = current_orientation * last_orientation.inverse();
277  egomotion_orientation.normalize();
278  tf2::Matrix3x3 m(egomotion_orientation);
279  double roll = 0.0;
280  double pitch = 0.0;
281  double yaw = 0.0;
282  m.getRPY(roll, pitch, yaw);
283 
284  odom.twist.twist.angular.x = roll / dt;
285  odom.twist.twist.angular.y = pitch / dt;
286  odom.twist.twist.angular.z = yaw / dt;
287 }
288 
289 std::string get_object_class(uint32_t classId)
290 {
291  std::string object_class("UNKNOWN");
292 
293  if (classId < k_object_class.size()) {
294  object_class = k_object_class.at(classId);
295  }
296 
297  return object_class;
298 }
299 
300 void define_box_size(outsight_alb_driver::ObjectData &tracked, const Tlv::tlv_s *bBoxSizes, size_t objectIndex)
301 {
302  float *bbox_size_raw = (float *)(bBoxSizes->getValue() + 3 * sizeof(float) * objectIndex);
303 
304  tracked.box_size.x = bbox_size_raw[0];
305  tracked.box_size.y = bbox_size_raw[1];
306  tracked.box_size.z = bbox_size_raw[2];
307 }
308 
309 void define_box_pose(outsight_alb_driver::ObjectData &tracked, const Tlv::tlv_s *bBoxPoses, size_t objectIndex)
310 {
311  std::array<float, 3> pose_translation(Pose::castTranslationValue(bBoxPoses, objectIndex));
312  std::array<float, 9> pose_rotation(Pose::castRotationValue(bBoxPoses, objectIndex));
313  define_pose(tracked.pose, pose_translation, pose_rotation);
314 }
315 
316 void define_object_speed(outsight_alb_driver::ObjectData &tracked, const Tlv::tlv_s *objectSpeeds, size_t objectIndex)
317 {
318  float *speed_raw = (float *)(objectSpeeds->getValue() + 3 * sizeof(float) * objectIndex);
319 
320  tracked.speed.linear.x = speed_raw[0];
321  tracked.speed.linear.y = speed_raw[1];
322  tracked.speed.linear.z = speed_raw[2];
323 }
324 
325 void define_augmented_cloud(outsight_alb_driver::AugmentedCloud &message, const Tlv::tlv_s *frame)
326 {
327  Tlv::Parser aug_cloud_parser(frame->getValue(), frame->getLength());
328 
329  const Tlv::tlv_s *number_object_tlv = aug_cloud_parser.findTlv(OSEF_TYPE_NUMBER_OF_POINTS);
330  const Tlv::tlv_s *number_layer_tlv = aug_cloud_parser.findTlv(OSEF_TYPE_NUMBER_OF_LAYERS);
331  message.number_of_points = NumberOfObjects::castValue(number_object_tlv);
332  message.number_of_layers = NumberOfObjects::castValue(number_layer_tlv);
333 
334  const Tlv::tlv_s *reflectivities_tlv = aug_cloud_parser.findTlv(OSEF_TYPE_REFLECTIVITIES);
335  if (reflectivities_tlv) {
336  message.reflectivities = std::vector<uint8_t>(
337  reflectivities_tlv->getValue(), reflectivities_tlv->getValue() + message.number_of_points);
338  }
339 
340  const Tlv::tlv_s *cartesian_tlv = aug_cloud_parser.findTlv(OSEF_TYPE_CARTESIAN_COORDINATES);
341  if (cartesian_tlv) {
342  float *cartesian = (float *)(cartesian_tlv->getValue());
343  message.cartesian_coordinates = std::vector<float>(cartesian, cartesian + 3 * message.number_of_points);
344  }
345 
346  const Tlv::tlv_s *object_ids_tlv = aug_cloud_parser.findTlv(OSEF_TYPE_OBJECT_ID_32_BITS);
347  if (object_ids_tlv) {
348  message.object_ids = ObjectIds32Bits::castValue(object_ids_tlv);
349  }
350 }
351 } // namespace Helper
Helper
Namespace to define helper functions.
Definition: alb_helper.h:22
OSEF_TYPE_REFLECTIVITIES
#define OSEF_TYPE_REFLECTIVITIES
Definition: osefTypes.h:44
Helper::get_egomotion_from_tlv
bool get_egomotion_from_tlv(const Tlv::tlv_s *frame, geometry_msgs::PoseStamped &pose)
Get a ROS egomotion pose from the TLV data.
Definition: alb_helper.cpp:195
ALBCommon::k_zone_config_id
constexpr const char * k_zone_config_id
Definition: alb_common.h:53
OSEF_TYPE_NUMBER_OF_POINTS
#define OSEF_TYPE_NUMBER_OF_POINTS
Definition: osefTypes.h:29
tf2::fromMsg
void fromMsg(const A &, B &b)
s
XmlRpcServer s
Helper::define_augmented_cloud
void define_augmented_cloud(outsight_alb_driver::AugmentedCloud &message, const Tlv::tlv_s *frame)
Define the AugmentedCloud message.
Definition: alb_helper.cpp:325
ros.h
Helper::define_point_fields
void define_point_fields(sensor_msgs::PointCloud2 &pointCloud)
Define the point fields for the ALB PointCloud2 message.
Definition: alb_helper.cpp:27
Helper::define_box_pose
void define_box_pose(outsight_alb_driver::ObjectData &tracked, const Tlv::tlv_s *bBoxPoses, size_t objectIndex)
Define the bounding box pose for the tracked object.
Definition: alb_helper.cpp:309
Helper::define_object_speed
void define_object_speed(outsight_alb_driver::ObjectData &tracked, const Tlv::tlv_s *objectSpeeds, size_t objectIndex)
Define the tracked object speed.
Definition: alb_helper.cpp:316
ALBCommon::k_zone_config_points
constexpr const char * k_zone_config_points
Definition: alb_common.h:54
tf2::Quaternion::normalize
Quaternion & normalize()
Helper::get_pose_from_tlv
bool get_pose_from_tlv(const Tlv::tlv_s *frame, geometry_msgs::PoseStamped &pose)
Get a ROS pose from the TLV data.
Definition: alb_helper.cpp:171
OSEF_TYPE_EGO_MOTION
#define OSEF_TYPE_EGO_MOTION
Definition: osefTypes.h:306
ALBCommon::k_zone_config_zones
constexpr const char * k_zone_config_zones
Definition: alb_common.h:51
Helper::define_points_data
void define_points_data(sensor_msgs::PointCloud2 &pointCloud, const uint32_t layers, const uint32_t points, float *pointData, bool use_colwise_order)
Define the PointCloud2 point data from raw points coming from the ALB.
Definition: alb_helper.cpp:52
OSEF_TYPE_NUMBER_OF_LAYERS
#define OSEF_TYPE_NUMBER_OF_LAYERS
Definition: osefTypes.h:88
f
f
tf2::Matrix3x3::getRPY
void getRPY(tf2Scalar &roll, tf2Scalar &pitch, tf2Scalar &yaw, unsigned int solution_number=1) const
NumberOfObjects::castValue
uint32_t castValue(const Tlv::tlv_s *leafTlv)
Definition: typeCaster.cpp:43
Helper::define_box_size
void define_box_size(outsight_alb_driver::ObjectData &tracked, const Tlv::tlv_s *bBoxSizes, size_t objectIndex)
Define the bounding box size for the tracked object.
Definition: alb_helper.cpp:300
Pose::castTranslationValue
const std::array< float, 3 > castTranslationValue(const Tlv::tlv_s *leafTlv, size_t index=0)
Definition: typeCaster.cpp:81
typeCaster.h
json
nlohmann::json json
Definition: alb_helper.cpp:17
alb_common.h
OSEF_TYPE_CARTESIAN_COORDINATES
#define OSEF_TYPE_CARTESIAN_COORDINATES
Definition: osefTypes.h:67
OSEF_TYPE_POSE_RELATIVE
#define OSEF_TYPE_POSE_RELATIVE
Definition: osefTypes.h:299
alb_helper.h
Helper::init_pose
void init_pose(geometry_msgs::Pose &pose)
Initialize a ROS Pose message.
Definition: alb_helper.cpp:135
Pose::castRotationValue
const std::array< float, 9 > castRotationValue(const Tlv::tlv_s *leafTlv, size_t index=0)
Definition: typeCaster.cpp:87
Tlv::Parser
Definition: tlvParser.h:15
k_object_class
const std::map< int, std::string > k_object_class
Definition: alb_helper.cpp:20
OSEF_TYPE_SCAN_FRAME
#define OSEF_TYPE_SCAN_FRAME
Definition: osefTypes.h:143
tf2::Quaternion::inverse
Quaternion inverse() const
ALBCommon::k_zone_config_name
constexpr const char * k_zone_config_name
Definition: alb_common.h:52
tf2::doTransform
void doTransform(const T &data_in, T &data_out, const geometry_msgs::TransformStamped &transform)
ALBCommon::k_zone_config_event
constexpr const char * k_zone_config_event
Definition: alb_common.h:56
ROS_ERROR
#define ROS_ERROR(...)
Tlv::Parser::findTlv
tlv_s * findTlv(type_t type) const
Definition: tlvParser.cpp:76
tf2::Quaternion
OSEF_TYPE_POSE
#define OSEF_TYPE_POSE
Definition: osefTypes.h:129
tf2_geometry_msgs.h
tf2::toMsg
B toMsg(const A &a)
OSEF_TYPE_OBJECT_ID_32_BITS
#define OSEF_TYPE_OBJECT_ID_32_BITS
Definition: osefTypes.h:337
Helper::get_object_class
std::string get_object_class(uint32_t classId)
Get the string class of the tracked object from the ALB.
Definition: alb_helper.cpp:289
ALBCommon::k_zone_config_role
constexpr const char * k_zone_config_role
Definition: alb_common.h:55
Helper::parse_zone_data
void parse_zone_data(const std::string &raw_zones, outsight_alb_driver::Zones &zones_msg)
Parse the zone data in the raw string and set them in the Zones msg.
Definition: alb_helper.cpp:146
tf2::Matrix3x3
ObjectIds32Bits::castValue
std::vector< uint32_t > castValue(const Tlv::tlv_s *leafTlv)
Definition: typeCaster.cpp:63
Helper::computeOdometry
void computeOdometry(nav_msgs::Odometry &odom, const geometry_msgs::Pose &current_pose, const geometry_msgs::Pose &last_pose, const geometry_msgs::Pose &first_pose, float dt)
Compute the odometry between the last two poses.
Definition: alb_helper.cpp:226
osefTypes.h
t
geometry_msgs::TransformStamped t
Helper::define_pose
void define_pose(geometry_msgs::Pose &pose, const std::array< float, 3 > &position, const std::array< float, 9 > &rotation)
Define a ROS Pose message from ALB position and rotation.
Definition: alb_helper.cpp:88


outsight_alb_driver
Author(s): Outsight
autogenerated on Thu Oct 13 2022 02:21:45