Hierarchical Deep Multiagent Reinforcement Learning with Temporal Abstraction. (arXiv:1809.09332v2 [cs.LG] UPDATED)

Multiagent reinforcement learning (MARL) is commonly considered to suffer
from non-stationary environments and exponentially increasing policy space. It
would be even more challenging when rewards are sparse and delayed over long
trajectories. In this paper, we study hierarchical deep MARL in cooperative
multiagent problems with sparse and delayed reward. With temporal abstraction,
we decompose the problem into a hierarchy of different time scales and
investigate how agents can learn high-level coordination based on the
independent skills learned at the low level. Three hierarchical deep MARL
architectures are proposed to learn hierarchical policies under different MARL
paradigms. Besides, we propose a new experience replay mechanism to alleviate
the issue of the sparse transitions at the high level of abstraction and the
non-stationarity of multiagent learning. We empirically demonstrate the
effectiveness of our approaches in two domains with extremely sparse feedback:
(1) a variety of Multiagent Trash Collection tasks, and (2) a challenging
online mobile game, i.e., Fever Basketball Defense.

Source link

Related posts

Letter regarding the MIT Schwarzman College of Computing working groups and Idea Bank


Reasons why to integrate Facebook messenger bot for your business


Energy-based Graph Convolutional Networks for Scoring Protein Docking Models.


This website uses cookies to improve your experience. We'll assume you're ok with this, but you can opt-out if you wish. Accept Read More

Privacy & Cookies Policy


COVID-19 (Coronavirus) is a new illness that is having a major effect on all businesses globally LIVE COVID-19 STATISTICS FOR World