Abstract
Modern multi-robot service robotics applications often rely on coordination capabilities at multiple levels, from global (system-wide) task allocation and selection, to local (nearby) spatial coordination to avoid collisions. Often, the global methods are considered to be the heart of the multi-robot system, while local methods are tacked on to overcome intermittent, spatially-limited hindrances. We tackle this general assumption. Utilizing the alphabet soup simulator (simulating order picking, made famous by Kiva Systems), we experiment with a set of myopic, local methods for obstacle avoidance. We report on a series of experiments with a reinforcement-learning approach, using the Effectiveness-Index intrinsic reward, to allow robots to learn to select between methods to use when avoiding collisions. We show that allowing the learner to explore the space of parameterized methods results in significant improvements, even compared to the original methods provided by the simulator.
Original language | English |
---|---|
Title of host publication | Springer Proceedings in Advanced Robotics |
Publisher | Springer Science and Business Media B.V. |
Pages | 299-311 |
Number of pages | 13 |
DOIs | |
State | Published - 2018 |
Publication series
Name | Springer Proceedings in Advanced Robotics |
---|---|
Volume | 6 |
ISSN (Print) | 2511-1256 |
ISSN (Electronic) | 2511-1264 |
Bibliographical note
Publisher Copyright:© 2018, Springer International Publishing AG.
Funding
Acknowledgements We gratefully acknowledge support by ISF grants #1511/12, and #1865/16, and good advice from Avi Seifert. As always, thanks to K. Ushi.
Funders | Funder number |
---|---|
Israel Science Foundation | 1511/12, 1865/16 |