Heterogeneous Multi-task Learning with Expert Diversity
Autor: | Raquel, Aoki, Frederick, Tung, Gabriel L, Oliveira |
---|---|
Rok vydání: | 2021 |
Předmět: | |
DOI: | 10.48550/arxiv.2106.10595 |
Popis: | Predicting multiple heterogeneous biological and medical targets is a challenge for traditional deep learning models. In contrast to single-task learning, in which a separate model is trained for each target, multi-task learning (MTL) optimizes a single model to predict multiple related targets simultaneously. To address this challenge, we propose the Multi-gate Mixture-of-Experts with Exclusivity (MMoEEx). Our work aims to tackle the heterogeneous MTL setting, in which the same model optimizes multiple tasks with different characteristics. Such a scenario can overwhelm current MTL approaches due to the challenges in balancing shared and task-specific representations and the need to optimize tasks with competing optimization paths. Our method makes two key contributions: first, we introduce an approach to induce more diversity among experts, thus creating representations more suitable for highly imbalanced and heterogenous MTL learning; second, we adopt a two-step optimization [6, 11] approach to balancing the tasks at the gradient level. We validate our method on three MTL benchmark datasets, including Medical Information Mart for Intensive Care (MIMIC-III) and PubChem BioAssay (PCBA). Comment: 10 pages, 7 figures, BIOKDD, IEEE/ACM |
Databáze: | OpenAIRE |
Externí odkaz: |