DesikRengarajan/LOGO
[ICLR 2022 Spotlight] Code for Reinforcement Learning with Sparse Rewards using Guidance from Offline Demonstration
Python
Issues
- 0
Whether LOGO also applies to PPO algorithm?
#5 opened by zhx0506 - 0
The sparse reward settings on mujoco
#4 opened by GaoHaoCN - 0
The sparse reward settings on mujoco
#3 opened by GaoHaoCN - 0
some issues about logo
#2 opened by Marioooooooooooooo - 4
You summarize the LOGO algorithm in formula (10) in your paper, but I don't know why the formula of minimize the upper bound is a minus sign. Why can the same function be used to update parameters in the code,
#1 opened by yongpan0715