MT-GBM: A Multi-Task Gradient Boosting Machine with Shared Decision TreesDownload PDF


Sep 29, 2021 (edited Oct 04, 2021)ICLR 2022 Conference Blind SubmissionReaders: Everyone
  • Keywords: Multi-Task learning, Classification
  • Abstract: Despite the success of deep learning in computer vision and natural language processing, Gradient Boosted Decision Tree (GBDT) is yet one of the most powerful tools for applications with tabular data such as e-commerce and FinTech. However, applying GBDT to multi-task learning is still a challenge. Unlike deep models that can jointly learn a shared latent representation across multiple tasks, GBDT can hardly learn a shared tree structure. In this paper, we propose Multi-Task Gradient Boosting Machine (MT-GBM), a GBDT-based method for multi-task learning. The MT-GBM can find the shared tree structures and split branches according to multi-task losses. First, it assigns multiple outputs to each leaf node. Next, it computes the gradient corresponding to each output (task). Then, we also propose an algorithm to combine the gradients of all tasks and update the tree. Finally, we apply MT-GBM to LightGBM. Experiments show that our MT-GBM improves the performance of the main task significantly, which means the proposed MT-GBM is efficient and effective.
  • One-sentence Summary: We propose an algorithm to combine gradients of all tasks to build the trees whose efficiency is on par with the original GBDT, to get that the performance of the main task significantly improved.
0 Replies