===== T1b-react - Reactive obstacle avoidance ====== The main task is to implement a function that will steer the simulated robot towards a given goal while reactively avoiding obstacles. |**Due date** | October 15, 2023, 23:59 PST | |**Deadline** | January 13, 2024, 23:59 PST | |**Points** | 3 | |**Label in BRUTE** | t1b-react | |**Files to submit** | archive with ''HexapodController.py'' file | |**Resources** | {{ :courses:uir:hw:b4m36uir_23_t1_resource_pack.zip | B4M36UIR_t1_resource_pack}} | ---- ===Assignment=== In class ''HexapodController.py'' implement the ''goto_reactive'' function. The purpose of the function is to produce the steering command that will drive the simulated hexapod robot towards the goal position based on the current odometry of the robot, the current collision state of the robot, and the laser scan perceived by the robot. The input parameters of the function are: - ''goal'' - [[courses:uir:hw:support:messages|Pose message]] - only ''position.x'' and ''position.y'' parameters are used as the goal position - ''odometry'' - [[courses:uir:hw:support:messages|Odometry message]] - ''pose.position.x'', ''pose.position.y'' and ''pose.orientation.quaternion'' encodes the current robot absolute position in the environment - ''collision'' - boolean - ''True'' if the robot collides with some obstacle, ''False'' otherwise - ''laser_scan'' - [[courses:uir:hw:support:messages|LaserScan message]] - contains the current laser scan data perceived by the robot. The function returns: - Zero command ''cmd_msg = Twist()'' when any of the input data are invalid - ''None'' when the robot currently collides with any obstacle. - ''None'' when the robot has reached the given goal (note that in real world the exact position is impossible to reach by the robot, hence the goal is considered reached when the robot is in the ''DELTA_DISTANCE'' vicinity of the goal position). - Otherwise, the function returns the steering command in form of a velocity command ([[courses:uir:hw:support:messages|Twist message]]) consisting of linear and angular parts that steers the robot towards the goal position while avoiding obstacles. The ''goto_reactive'' function has the following prescription def goto_reactive(self, goal, odometry, collision, laser_scan): """Method to steer the robot towards the goal position while avoiding contact with the obstacles given its current odometry, collision status and laser scan data Args: goal: Pose of the robot goal odometry: Perceived odometry of the robot collision: bool of the robot collision status laser_scan: LaserScan data perceived by the robot Returns: cmd: Twist steering command """ In class ''HexapodController.py'' you can change whatever you want. In evaluation, the given interfaces are fixed and the evaluation script is fixed. ---- === Approach === The recommended approach for the reactive obstacle avoidance uses simple AI cognitive model of Braitenberg vehicles described in [[courses:uir:labs:lab02|Lab02 - Exteroceptive sensing, Mapping and Reactive-based Obstacle Avoidance]]. The direct sensory-motor mapping has to combine: * navigation towards the goal location (implemented in the task [[courses:uir:hw:t1a-ctrl|T1a-ctrl - Open-loop locomotion control]]), * reactive obstacle avoidance based on the currently perceived environment. While the continuous navigation function towards the goal location can be represented by //love// (Vehicle 3a) or //aggression// (Vehicle 2b), the reactive obstacle avoidance is best modelled as the robot //fear// (Vehicle 2a) the obstacles in the environment perceived by the laser scanner with nonlinear activation functions. Hence the reactive obstacle avoidance can be achieved using the following continuous navigation function (pseudocode). while not goal_reached: dphi = the difference between the current heading and the heading towards the target scan_left = distance to the closest obstacle to the left of the robot scan_right = distance to the closest obstacle to the right of the robot repulsive_force = 1/scan_left - 1/scan_right linear_speed = distance towards target angular_speed_navigation_component = dphi*C_TURNING_SPEED angular_speed_avoidance_component = repulsive_force*C_AVOID_SPEED angular_speed = angular_speed_navigation_component + angular_speed_avoidance_component Where ''C_AVOID_SPEED'' is a constant that defines the robot //fear// of the obstacle and ''C_TURNING_SPEED'' is a constant that defines the //aggression// with which the robot will turn towards the desired heading. An example operation of the controller can be seen in the following video (4x speed up): {{:courses:uir:hw:t1b-react.gif?600|}} ---- === Evaluation === The evaluation focus on the ability of the robot to reach the given goal locations while not crash into the obstacles. It is supposed to introduce you to the work with the robot exteroceptive sensing. The code can be evaluated using the following script (also attached as ''t1b-react.py''). #!/usr/bin/env python3 # -*- coding: utf-8 -*- import matplotlib.pyplot as plt import sys import os import math import numpy as np sys.path.append('hexapod_robot') #import hexapod robot import HexapodRobot as hexapod #import communication messages from messages import * if __name__=="__main__": robot = hexapod.HexapodRobot(0) #turn on the robot robot.turn_on() #start navigation thread robot.start_navigation() #assign goal for navigation goals = [ Pose(Vector3(3.5,3.5,0),Quaternion(1,0,0,0)), Pose(Vector3(0,0,0),Quaternion(1,0,0,0)), ] path = Path() #go from goal to goal for goal in goals: robot.goto_reactive(goal) while robot.navigation_goal is not None: #sample the current odometry if robot.odometry_ is not None: path.poses.append(robot.odometry_.pose) #wait time.sleep(0.1) #check the robot distance to goal odom = robot.odometry_.pose #compensate for the height of the robot as we are interested only in achieved planar distance odom.position.z = 0 #calculate the distance dist = goal.dist(odom) print("[t1b_eval] distance to goal: %.2f" % dist) robot.stop_navigation() robot.turn_off() #plot the robot path fig, ax = plt.subplots() path.plot(ax, 30) plt.xlabel('x[m]') plt.ylabel('y[m]') plt.axis('equal') plt.show() The expected output is the print of the distance readings below or equal the ''DELTA_DISTANCE'' threshold and the plot of the robot path between the individual goals similar to the following figures. Note that the resulting path may be completely different, the main focus of this task is not to collide with the obstacles, hence, any collision-free trajectory is valid. {{:courses:uir:hw:t1b-path.jpg?500|}} {{:courses:uir:hw:t1b-path2.jpg?500|}}