Skip to content
Snippets Groups Projects
  1. Jan 30, 2019
  2. Jan 29, 2019
  3. Jan 24, 2019
    • Jae Young Lee's avatar
      Improve and Bug-fix DQNLearner and environments. · 4a9327bd
      Jae Young Lee authored
      - Added RestrictedEpsGreedyPolicy and RestrictedGreedyPolicy and use them as policy and test_policy in DQNLearner. Now, the agent never chooses the action corresponding to -inf Q-value if there is at least one action with finite Q-value (if not, it chooses any action randomly, which is necessary for compatibility with keras-rl --
       see the comments in select_action).
      
      - Now, generate_scenario in SimpleIntersectionEnv generates veh_ahead_scenario even when randomize_special_scenario = 1.
      
      - In EpisodicEnvBase, the terminal reward is by default determined by the minimum one;
      
      - Small change of initiation_condition of EpisodicEnvBase (simplified);
      4a9327bd
    • Jae Young Lee's avatar
  4. Jan 22, 2019
    • Jae Young Lee's avatar
      Add and train more low-level policies, train a high-level policy. · a90b4bc5
      Jae Young Lee authored
      The high-level policy was trained without changelane maneuver but with immediatestop maneuver. Two problems remain: 1) the agent chooses changelane maneuver too frequently; 2) before the stop region, immediatestop maneuver works but was not chosen property after 2.5m the high-level policy training...
      a90b4bc5
  5. Jan 17, 2019
  6. Nov 19, 2018
  7. Nov 18, 2018
  8. Nov 17, 2018
  9. Nov 16, 2018
Loading