V-MPO: On-Policy Maximum a Posteriori Policy Optimization for Discrete and Continuous ControlDownload PDF

25 Sept 2019, 19:16 (edited 10 Feb 2022)ICLR 2020 Conference Blind SubmissionReaders: Everyone
Original Pdf: pdf
Data:
Keywords:
TL;DR:
Abstract:
14 Replies

Loading