【IS视角下的PPO】从Importance Sampling(IS)到PPO

less than 1 minute read

Published:

I recap the Importance Sampling (IS) and Proximal Policy Optimization (PPO) algorithms in the context of RL. Chech out the post for more details.