Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Peer review Lab 10 - Dimitri Masetta, s306130 #5

Open
dimi-it opened this issue Jan 5, 2024 · 0 comments
Open

Peer review Lab 10 - Dimitri Masetta, s306130 #5

dimi-it opened this issue Jan 5, 2024 · 0 comments

Comments

@dimi-it
Copy link

dimi-it commented Jan 5, 2024

Preface

There is a usefull readme, that combined with the well written code, although a lack of comments, enable an easy understanding of the code!

Reinforcement Learning algorithm

You implemented a Q-Learning RL that in general seems correct, but there is something that don't enable the training process to be effective, and the results prove it (agent vs random with 54 wins, 11 ties, 35 losses is almost as good as random vs random considering that the agent play always as first):

  • a dynamic tuning of the parameters during the training is missing, but this is not the main problem, as their influence on the performance is low
  • probably something in your "update_q_value" function don't correctly update the Qvalue, but I couldn't point to the source of the problem.

I really liked that you train your agent against different players and really admire your work to develop them but I think that you can improve your training process to greatly improve your agent performance.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant