|
| example3 - opencv deploy.aligned_stream = rs.align(rs.stream.color) |
|
tuple | example3 - opencv deploy.bbox = (int(left), int(top), int(width), int(height)) |
|
| example3 - opencv deploy.bottom = detection[6]*H |
|
tuple | example3 - opencv deploy.bottomLeftCornerOfText = (p1[0], p1[1]+20) |
|
| example3 - opencv deploy.color_frame = frames.get_color_frame() |
|
| example3 - opencv deploy.color_image = np.asanyarray(color_frame.get_data()) |
|
| example3 - opencv deploy.config = rs.config() |
|
| example3 - opencv deploy.depth_frame = frames.get_depth_frame().as_depth_frame() |
|
| example3 - opencv deploy.depth_image = np.asanyarray(depth_frame.get_data()) |
|
| example3 - opencv deploy.detections = net.forward() |
|
| example3 - opencv deploy.font = cv2.FONT_HERSHEY_SIMPLEX |
|
tuple | example3 - opencv deploy.fontColor = (255, 255, 255) |
|
int | example3 - opencv deploy.fontScale = 1 |
|
| example3 - opencv deploy.frames = pipeline.wait_for_frames() |
|
int | example3 - opencv deploy.H = 480 |
|
| example3 - opencv deploy.height = bottom-top |
|
string | example3 - opencv deploy.height_txt = str(height)+"[m]" |
|
| example3 - opencv deploy.idx = int(detection[1]) |
|
| example3 - opencv deploy.left = detection[3]*W |
|
int | example3 - opencv deploy.lineType = 2 |
|
| example3 - opencv deploy.my = np.amin(ys, initial=1) |
|
| example3 - opencv deploy.My = np.amax(ys, initial=-1) |
|
| example3 - opencv deploy.net = cv2.dnn.readNetFromTensorflow("frozen_inference_graph.pb", "faster_rcnn_inception_v2_coco_2018_01_28.pbtxt") |
|
| example3 - opencv deploy.obj_points = verts[int(bbox[1]):int(bbox[1] + bbox[3]), int(bbox[0]):int(bbox[0] + bbox[2])].reshape(-1, 3) |
|
tuple | example3 - opencv deploy.p1 = (int(bbox[0]), int(bbox[1])) |
|
tuple | example3 - opencv deploy.p2 = (int(bbox[0] + bbox[2]), int(bbox[1] + bbox[3])) |
|
| example3 - opencv deploy.pipeline = rs.pipeline() |
|
| example3 - opencv deploy.point_cloud = rs.pointcloud() |
|
| example3 - opencv deploy.points = point_cloud.calculate(depth_frame) |
|
| example3 - opencv deploy.right = detection[5]*W |
|
tuple | example3 - opencv deploy.scaled_size = (int(W), int(H)) |
|
| example3 - opencv deploy.score = float(detection[2]) |
|
| example3 - opencv deploy.top = detection[4]*H |
|
| example3 - opencv deploy.verts = np.asanyarray(points.get_vertices()).view(np.float32).reshape(-1, W, 3) |
|
int | example3 - opencv deploy.W = 848 |
|
| example3 - opencv deploy.width = right-left |
|
| example3 - opencv deploy.ys = obj_points[:,1] |
|
| example3 - opencv deploy.z = np.median(zs) |
|
| example3 - opencv deploy.zs = obj_points[:,2] |
|