- 05 Feb, 2019 1 commit
-
-
Jae Young Lee authored
- Also added ManualWait class.
-
- 01 Feb, 2019 1 commit
-
-
Jae Young Lee authored
-
- 31 Jan, 2019 1 commit
-
-
Jae Young Lee authored
-
- 30 Jan, 2019 3 commits
-
-
Jae Young Lee authored
-
Jae Young Lee authored
-
Jae Young Lee authored
-
- 29 Jan, 2019 1 commit
-
-
Jae Young Lee authored
-
- 24 Jan, 2019 1 commit
-
-
Jae Young Lee authored
- Added RestrictedEpsGreedyPolicy and RestrictedGreedyPolicy and use them as policy and test_policy in DQNLearner. Now, the agent never chooses the action corresponding to -inf Q-value if there is at least one action with finite Q-value (if not, it chooses any action randomly, which is necessary for compatibility with keras-rl -- see the comments in select_action). - Now, generate_scenario in SimpleIntersectionEnv generates veh_ahead_scenario even when randomize_special_scenario = 1. - In EpisodicEnvBase, the terminal reward is by default determined by the minimum one; - Small change of initiation_condition of EpisodicEnvBase (simplified);
-
- 22 Jan, 2019 1 commit
-
-
Jae Young Lee authored
The high-level policy was trained without changelane maneuver but with immediatestop maneuver. Two problems remain: 1) the agent chooses changelane maneuver too frequently; 2) before the stop region, immediatestop maneuver works but was not chosen property after 2.5m the high-level policy training...
-
- 17 Jan, 2019 1 commit
-
-
Jae Young Lee authored
Each low level policy was retrained with better LTL conditions and rewards, some parts of which are also designed to encourage exploration (to prevent the vehicle from being stopped all the time).
-
- 19 Nov, 2018 2 commits
-
-
Ashish Gaurav authored
-
Ashish Gaurav authored
-
- 18 Nov, 2018 3 commits
-
-
Jae Young Lee authored
-
Jae Young Lee authored
-
Jae Young Lee authored
-
- 17 Nov, 2018 1 commit
-
-
Aravind Bk authored
-