-
- Downloads
Improve and Bug-fix DQNLearner and environments.
- Added RestrictedEpsGreedyPolicy and RestrictedGreedyPolicy and use them as policy and test_policy in DQNLearner. Now, the agent never chooses the action corresponding to -inf Q-value if there is at least one action with finite Q-value (if not, it chooses any action randomly, which is necessary for compatibility with keras-rl -- see the comments in select_action). - Now, generate_scenario in SimpleIntersectionEnv generates veh_ahead_scenario even when randomize_special_scenario = 1. - In EpisodicEnvBase, the terminal reward is by default determined by the minimum one; - Small change of initiation_condition of EpisodicEnvBase (simplified);
Showing
- backends/kerasrl_learner.py 92 additions, 7 deletionsbackends/kerasrl_learner.py
- env/env_base.py 4 additions, 5 deletionsenv/env_base.py
- env/simple_intersection/simple_intersection_env.py 1 addition, 1 deletionenv/simple_intersection/simple_intersection_env.py
- options/simple_intersection/maneuver_base.py 1 addition, 2 deletionsoptions/simple_intersection/maneuver_base.py
Loading
Please register or sign in to comment