Decision-theoretic planning under uncertainty with information rewards for active cooperative perception

Verfasser / Beitragende:
[Matthijs Spaan, Tiago Veiga, Pedro Lima]
Ort, Verlag, Jahr:
2015
Enthalten in:
Autonomous Agents and Multi-Agent Systems, 29/6(2015-11-01), 1157-1185
Format:
Artikel (online)
ID: 60551464X
LEADER caa a22 4500
001 60551464X
003 CHVBK
005 20210128100705.0
007 cr unu---uuuuu
008 210128e20151101xx s 000 0 eng
024 7 0 |a 10.1007/s10458-014-9279-8  |2 doi 
035 |a (NATIONALLICENCE)springer-10.1007/s10458-014-9279-8 
245 0 0 |a Decision-theoretic planning under uncertainty with information rewards for active cooperative perception  |h [Elektronische Daten]  |c [Matthijs Spaan, Tiago Veiga, Pedro Lima] 
520 3 |a Partially observable Markov decision processes (POMDPs) provide a principled framework for modeling an agent's decision-making problem when the agent needs to consider noisy state estimates. POMDP policies take into account an action's influence on the environment as well as the potential information gain. This is a crucial feature for robotic agents which generally have to consider the effect of actions on sensing. However, building POMDP models which reward information gain directly is not straightforward, but is important in domains such as robot-assisted surveillance in which the value of information is hard to quantify. Common techniques for uncertainty reduction such as expected entropy minimization lead to non-standard POMDPs that are hard to solve. We present the POMDP with Information Rewards (POMDP-IR) modeling framework, which rewards an agent for reaching a certain level of belief regarding a state feature. By remaining in the standard POMDP setting we can exploit many known results as well as successful approximate algorithms. We demonstrate our ideas in a toy problem as well as in real robot-assisted surveillance, showcasing their use for active cooperative perception scenarios. Finally, our experiments show that the POMDP-IR framework compares favorably with a related approach on benchmark domains. 
540 |a The Author(s), 2014 
690 7 |a Active cooperative perception  |2 nationallicence 
690 7 |a Planning under uncertainty for robots  |2 nationallicence 
690 7 |a Partially observable Markov decision processes  |2 nationallicence 
700 1 |a Spaan  |D Matthijs  |u Delft University of Technology, Delft, The Netherlands  |4 aut 
700 1 |a Veiga  |D Tiago  |u Institute for Systems and Robotics, Instituto Superior Técnico, Universidade de Lisboa, Lisbon, Portugal  |4 aut 
700 1 |a Lima  |D Pedro  |u Institute for Systems and Robotics, Instituto Superior Técnico, Universidade de Lisboa, Lisbon, Portugal  |4 aut 
773 0 |t Autonomous Agents and Multi-Agent Systems  |d Springer US; http://www.springer-ny.com  |g 29/6(2015-11-01), 1157-1185  |x 1387-2532  |q 29:6<1157  |1 2015  |2 29  |o 10458 
856 4 0 |u https://doi.org/10.1007/s10458-014-9279-8  |q text/html  |z Onlinezugriff via DOI 
898 |a BK010053  |b XK010053  |c XK010000 
900 7 |a Metadata rights reserved  |b Springer special CC-BY-NC licence  |2 nationallicence 
908 |D 1  |a research-article  |2 jats 
949 |B NATIONALLICENCE  |F NATIONALLICENCE  |b NL-springer 
950 |B NATIONALLICENCE  |P 856  |E 40  |u https://doi.org/10.1007/s10458-014-9279-8  |q text/html  |z Onlinezugriff via DOI 
950 |B NATIONALLICENCE  |P 700  |E 1-  |a Spaan  |D Matthijs  |u Delft University of Technology, Delft, The Netherlands  |4 aut 
950 |B NATIONALLICENCE  |P 700  |E 1-  |a Veiga  |D Tiago  |u Institute for Systems and Robotics, Instituto Superior Técnico, Universidade de Lisboa, Lisbon, Portugal  |4 aut 
950 |B NATIONALLICENCE  |P 700  |E 1-  |a Lima  |D Pedro  |u Institute for Systems and Robotics, Instituto Superior Técnico, Universidade de Lisboa, Lisbon, Portugal  |4 aut 
950 |B NATIONALLICENCE  |P 773  |E 0-  |t Autonomous Agents and Multi-Agent Systems  |d Springer US; http://www.springer-ny.com  |g 29/6(2015-11-01), 1157-1185  |x 1387-2532  |q 29:6<1157  |1 2015  |2 29  |o 10458