Skip to content
Snippets Groups Projects
  1. Feb 06, 2019
  2. Feb 05, 2019
  3. Feb 04, 2019
  4. Feb 01, 2019
  5. Jan 31, 2019
  6. Jan 30, 2019
  7. Jan 29, 2019
  8. Jan 24, 2019
    • Jae Young Lee's avatar
      Improve and Bug-fix DQNLearner and environments. · 4a9327bd
      Jae Young Lee authored
      - Added RestrictedEpsGreedyPolicy and RestrictedGreedyPolicy and use them as policy and test_policy in DQNLearner. Now, the agent never chooses the action corresponding to -inf Q-value if there is at least one action with finite Q-value (if not, it chooses any action randomly, which is necessary for compatibility with keras-rl --
       see the comments in select_action).
      
      - Now, generate_scenario in SimpleIntersectionEnv generates veh_ahead_scenario even when randomize_special_scenario = 1.
      
      - In EpisodicEnvBase, the terminal reward is by default determined by the minimum one;
      
      - Small change of initiation_condition of EpisodicEnvBase (simplified);
      4a9327bd
    • Jae Young Lee's avatar
  9. Jan 22, 2019
    • Jae Young Lee's avatar
      Add and train more low-level policies, train a high-level policy. · a90b4bc5
      Jae Young Lee authored
      The high-level policy was trained without changelane maneuver but with immediatestop maneuver. Two problems remain: 1) the agent chooses changelane maneuver too frequently; 2) before the stop region, immediatestop maneuver works but was not chosen property after 2.5m the high-level policy training...
      a90b4bc5
  10. Jan 17, 2019
  11. Nov 19, 2018
  12. Nov 18, 2018
Loading