"Online Markov Decision Processes Under Bandit Feedback."

Gergely Neu et al. (2014)

Details and statistics

DOI: 10.1109/TAC.2013.2292137

access: closed

type: Journal Article

metadata version: 2020-10-26