Social Imitation in Cooperative Multiarmed Bandits: Partition-Based Algorithms with Strictly Local Information

Peter Landgren, Vaibhav Srivastava, Naomi Ehrich Leonard

Research output: Chapter in Book/Report/Conference proceedingConference contribution

24 Scopus citations

Abstract

We study distributed cooperative decision-making in a multi-agent stochastic multi-armed bandit (MAB) problem in which agents are connected through an undirected graph and observe the actions and rewards of their neighbors. We develop a novel policy based on partitions of the communication graph and propose a distributed method for selecting an arbitrary number of leaders and partitions. We analyze this new policy and evaluate its performance using Monte-Carlo simulations.

Original languageEnglish (US)
Title of host publication2018 IEEE Conference on Decision and Control, CDC 2018
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages5239-5244
Number of pages6
ISBN (Electronic)9781538613955
DOIs
StatePublished - Jul 2 2018
Externally publishedYes
Event57th IEEE Conference on Decision and Control, CDC 2018 - Miami, United States
Duration: Dec 17 2018Dec 19 2018

Publication series

NameProceedings of the IEEE Conference on Decision and Control
Volume2018-December
ISSN (Print)0743-1546
ISSN (Electronic)2576-2370

Conference

Conference57th IEEE Conference on Decision and Control, CDC 2018
Country/TerritoryUnited States
CityMiami
Period12/17/1812/19/18

All Science Journal Classification (ASJC) codes

  • Control and Systems Engineering
  • Modeling and Simulation
  • Control and Optimization

Fingerprint

Dive into the research topics of 'Social Imitation in Cooperative Multiarmed Bandits: Partition-Based Algorithms with Strictly Local Information'. Together they form a unique fingerprint.

Cite this