- Keywords: Domain Adaptation, Basis Transfer, Transfer Learning, Low Rank Approximation, Nyström Approximation
- TL;DR: The paper describes a low-rank basis transfer algorithm using only a subset from the domains with outstanding results.
- Abstract: Domain adaptation focuses on the reuse of supervised learning models in a new context. Prominent applications can be found in robotics, image processing or web mining. In these areas, learning scenarios change by nature, but often remain related and motivate the reuse of existing supervised models. While the majority of symmetric and asymmetric domain adaptation algorithms utilize all available source and target domain data, we show that efficient domain adaptation requires only a substantially smaller subset from both domains. This makes it more suitable for real-world scenarios where target domain data is rare. The presented approach finds a target subspace representation for source and target data to address domain differences by orthogonal basis transfer. By employing a low-rank approximation, the approach remains low in computational time. The presented idea is evaluated in typical domain adaptation tasks with standard benchmark data.
- Code: https://github.com/iclr-nbt/nbt.git