July 14, 2020
In complex tasks, such as those with large combinatorial action spaces, random exploration may be too inefficient to achieve meaningful learning progress. In this work, we use a curriculum of progressively growing action spaces to accelerate learning. We assume the environment is out of our control, but that the agent may set an internal curriculum by initially restricting its action space. Our approach uses off-policy reinforcement learning to estimate optimal value functions for multiple action spaces simultaneously and efficiently transfers data, value estimates, and state representations from restricted action spaces to the full task. We show the efficacy of our approach in proof-of-concept control tasks and on challenging large-scale StarCraft micromanagement tasks with large, multi-agent action spaces.
Written by
Gregory Farquhar
Laura Gustafson
Randi Cabezas
Zeming Lin
Shimon Whiteson
Nicolas UsunierGabriel SynnaevePublisher
International Conference on Machine Learning (ICML)
Research Topics
Artifical Intelligence
June 02, 2019
This paper explores the problem of ranking short social media posts with respect to user queries using neural networks. Instead of starting with a complex architecture, we proceed from the bottom up and examine the effectiveness of a simple,…
Peng Shi, Jinfeng Rao, Jimmy Lin
June 02, 2019
June 09, 2019
Over the past few years, neural networks were proven vulnerable to adversarial images: targeted but imperceptible image perturbations lead to drastically different predictions. We show that adversarial vulnerability increases with the gradients…
Carl-Johann Simon-Gabriel, Yann Ollivier, Bernhard Scholkopf, Leon Bottou, David Lopez-Paz
June 09, 2019
May 31, 2019
Abuse on the Internet represents a significant societal problem of our time. Previous research on automated abusive language detection in Twitter has shown that community-based profiling of users is a promising technique for this task. However,…
Pushkar Mishra, Marco Del Tredici, Helen Yannakoudakis, Ekaterina Shutova
May 31, 2019
June 01, 2019
Reduced models are simplified versions of a given domain, designed to accelerate the planning process. Interest in reduced models has grown since the surprising success of determinization in the first international probabilistic planning…
Luis Pineda, Shlomo Zilberstein
June 01, 2019