From ee257c39be3c0645bacd183f289c4465d19666e2 Mon Sep 17 00:00:00 2001
From: Jaeyoung Lee <jaeyoung.lee@uwaterloo.ca>
Date: Thu, 31 Jan 2019 17:55:50 -0500
Subject: [PATCH] Further improve Follow and make KeepLane default (available
 any time).

---
 options/simple_intersection/maneuvers.py | 32 +++++++++++++-----------
 1 file changed, 18 insertions(+), 14 deletions(-)

diff --git a/options/simple_intersection/maneuvers.py b/options/simple_intersection/maneuvers.py
index cda94f8..3a99a45 100644
--- a/options/simple_intersection/maneuvers.py
+++ b/options/simple_intersection/maneuvers.py
@@ -52,6 +52,16 @@ class KeepLane(ManeuverBase):
     def _features_dim_reduction(features_tuple):
         return extract_ego_features(features_tuple, 'v', 'v_ref', 'e_y', 'psi', 'v tan(psi/L)', 'theta', 'lane', 'acc', 'psi_dot')
 
+    @property
+    def initiation_condition(self):
+        """a virtual function (property) from ManeuverBase.
+            As KeepLane is a default maneuver, it has to be activated to be chosen at any time, state, and condition
+            (refer to initiation_condition of ManeuverBase for the usual case.
+            :returns True.
+        """
+
+        return True
+
     @property
     def extra_termination_condition(self):
         if self._enable_low_level_training_properties:  # activated only for the low-level training.
@@ -383,6 +393,7 @@ class ChangeLane(ManeuverBase):
                                     'psi_dot')
 
 
+# TODO: In the low-level training of Follow, sometime the initial state doesn't satisfy its initiation condition.
 class Follow(ManeuverBase):
 
     _reward_in_goal = None
@@ -422,7 +433,6 @@ class Follow(ManeuverBase):
         self.env._terminate_in_goal = False
         self._penalty_for_out_of_range = 200
         self._penalty_for_change_lane = 200
-        self._reward_in_goal = 200
         self._enable_low_level_training_properties = True
         self._extra_action_weights_flag = True
 
@@ -436,7 +446,6 @@ class Follow(ManeuverBase):
         self._set_v_ref()
 
     def _set_v_ref(self):
-        #if self._enable_low_level_training_properties:
         self._target_veh_i, _ = self.env.get_V2V_distance()
 
         if self._target_veh_i is not None:
@@ -460,23 +469,18 @@ class Follow(ManeuverBase):
 
     @property
     def extra_termination_condition(self):
-        # APs = self.env.ego.APs
-
         if self._target_veh_i is None:
             return False
 
         if self._enable_low_level_training_properties:  # activated only for the low-level training.
-            if self.env.ego.APs['has_stopped_in_stop_region'] and \
-                    self.env.ego.APs['in_stop_region']:
-                self._extra_r_terminal = None
-                return True
-
-            elif (rd.speed_limit / 5 < self._v_ref) and \
+            APs = self.env.ego.APs
+            if (rd.speed_limit / 5 < self._v_ref) and \
                     (self.env.ego.v < self._v_ref / 2) and \
                     self.env.ego.acc < 0 and \
-                    not self.env.ego.APs['veh_ahead_stopped_now']:
-                self._extra_r_terminal = -100
-                return True
+                    not APs['veh_ahead_stopped_now'] and \
+                    not APs['in_stop_region']:
+                    self._extra_r_terminal = -100
+                    return True
 
             else:
                 self._extra_r_terminal = None
@@ -493,4 +497,4 @@ class Follow(ManeuverBase):
             return ego_features + extract_other_veh_features(
                 features_tuple, self._target_veh_i, 'rel_x', 'rel_y', 'v', 'acc')
         else:
-            return ego_features + (0.0, 0.0, 0.0, 0.0)
\ No newline at end of file
+            return ego_features + (0.0, 0.0, 0.0, 0.0)
-- 
GitLab