posted on 2024-02-09, 18:04authored byD. Koert, G. Maeda, J. Peters, Gerhard Neumann
<p>Challenging tasks in unstructured environments require robots to learn complex models. Given a large amount of information, learning multiple simple models can offer an efficient alternative to a monolithic complex network. Training multiple models—that is, learning their parameters and their responsibilities—has been shown to be prohibitively hard as optimization is prone to local minima. To efficiently learn multiple models for different contexts, we thus develop a new algorithm based on expectation maximization (EM). In contrast to comparable concepts, this algorithm trains multiple modules of paired forward-inverse models by using the prediction errors of both forward and inverse models simultaneously. In particular, we show that our method yields a substantial improvement over only considering the errors of the forward models on tasks where the inverse space contains multiple solutions</p>
History
School affiliated with
School of Computer Science (Research Outputs)
Date Submitted
2018-04-17
Date Accepted
2018-05-21
Date of First Publication
2018-05-21
Date of Final Publication
2018-05-21
Event Name
International Conference on Robotics and Automation (ICRA)