Reinforcement Learning Adventure
In this little project, I aim at summarizing the Reinforcement Learning knowledge, step by step. The motivation is to organize what I have learned from Udacity Deep Reinforcement Learning Nanodegree and goes beyond. Here I will write blogs and codes for the basic knowledges and also for the important papers in this area, starting with the recommandation in the Nanodegree.
The main 'Adventure' play ground will be in OpenAI's gym environment.
Notice that many of my codes will be referencing the Nanodegree contents and other excellent implementations on Github.
-
the excellent algorithm implementations like ShangtongZhang and rlcode on Github.
-
Blogs writing (20% completed)
-
Reinforcement Learning Intro (blogs 30% completed)
-
Dynamic Progrmming (0% completed)
-
Classic
- Monte Carlo method (workable codes implemented, Blogs 0% completed)
- Temporal-Difference methods (workable codes implemented, Blogs 0% completed)
- RL in Continous Space (Blogs 0% completed)
-
Value-based methods
- Classic DQN with image inputs for learning(workable codes implemented, still tuning for request performance...)
- Classic DQN (workable codes implemented, blogs 0% completed)
- Double-DQN (workable codes implemented, blogs 0% completed)
- Prioritized Replay DDQN (workable codes implemented, blogs completed)
- Dueling-DQN (workable codes implemented, blogs completed)
-
Policy-based methods
- Policy-based and Hill climbing algorthm(blogs and workable codes completed) -Policy Gradients method(Blogs completed)
-Proximal Policy Optimization(working on blogs)