Keywords: task-parameterized imitation learning, interactive online learning, imitation learning
TL;DR: Interactive incremental imitation learning framework that simultaneously leverages local and global modulations of trajectory distributions to improve generalization and enabling online skill modulation through human feedback.
Abstract: The problem of generalization in learning from demonstration (LfD) has received considerable attention over the years, particularly within the context of movement primitives, where two prominent approaches have emerged. While one leverages via-points to adapt skills locally by modulating demonstrated trajectories, the other relies on so-called task-parameterized models that encode movements with respect to different coordinate systems, using a product of probabilities for generalization. While the former are well-suited to precise, local modulations, the latter aim at generalizing over large regions of the workspace and often involve multiple objects. Addressing the quality of generalization by leveraging both approaches simultaneously has received little attention. In this work, we propose an interactive imitation learning framework that simultaneously leverages local and global modulations of trajectory distributions. Building on the kernelized movement primitives (KMP) framework, we introduce novel mechanisms for skill modulation from direct human corrective feedback. Our approach particularly exploits the concept of via-points to incrementally and interactively 1) improve the model accuracy locally, 2) add new objects to the task during execution and 3) extend the skill into regions where demonstrations were not provided. We evaluate our method on a bearing ring-loading task using a torque-controlled, 7-DoF, robot.
Supplementary Material: zip
Submission Number: 2
Loading