Simple Reinforcement learning tutorials, 莫烦Python 中文AI教学
-
Updated
Mar 31, 2024 - Python
Simple Reinforcement learning tutorials, 莫烦Python 中文AI教学
An elegant PyTorch deep reinforcement learning library.
High-quality single file implementation of Deep Reinforcement Learning algorithms with research-friendly features (PPO, DQN, C51, DDPG, TD3, SAC, PPG)
PyTorch implementation of DQN, AC, ACER, A2C, A3C, PG, DDPG, TRPO, PPO, SAC, TD3 and ....
Massively Parallel Deep Reinforcement Learning. 🔥
PyTorch implementation of Advantage Actor Critic (A2C), Proximal Policy Optimization (PPO), Scalable trust-region method for deep reinforcement learning using Kronecker-factored approximation (ACKTR) and Generative Adversarial Imitation Learning (GAIL).
Modularized Implementation of Deep RL Algorithms in PyTorch
Clean, Robust, and Unified PyTorch implementation of popular Deep Reinforcement Learning (DRL) algorithms (Q-learning, Duel DDQN, PER, C51, Noisy DQN, PPO, DDPG, TD3, SAC, ASL)
Implementations of basic RL algorithms with minimal lines of codes! (pytorch based)
For trading. Please star.
Minimal implementation of clipped objective Proximal Policy Optimization (PPO) in PyTorch
This is the official implementation of Multi-Agent PPO (MAPPO).
Modular Deep Reinforcement Learning framework in PyTorch. Companion library of the book "Foundations of Deep Reinforcement Learning".
PyTorch implementation of Deep Reinforcement Learning: Policy Gradient methods (TRPO, PPO, A2C) and Generative Adversarial Imitation Learning (GAIL). Fast Fisher vector product TRPO.
Proximal Policy Optimization (PPO) algorithm for Super Mario Bros
A simple and well styled PPO implementation. Based on my Medium series: https://medium.com/@eyyu/coding-ppo-from-scratch-with-pytorch-part-1-4-613dfc1b14c8.
XuanCe: A Comprehensive and Unified Deep Reinforcement Learning Library
Training a humanoid robot for locomotion using Reinforcement Learning
Really Fast End-to-End Jax RL Implementations
Add a description, image, and links to the ppo topic page so that developers can more easily learn about it.
To associate your repository with the ppo topic, visit your repo's landing page and select "manage topics."