Skip to main content

Learning Task Sampling Policy for Multitask Learning

Publication ,  Conference
Sundararaman, D; Tsai, H; Lee, KH; Turc, I; Carin, L
Published in: Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021
January 1, 2021

It has been shown that training multi-task models with auxiliary tasks can improve the target tasks quality through cross-task transfer. However, the importance of each auxiliary task to the primary task is likely not known a priori. While the importance weights of auxiliary tasks can be manually tuned, it becomes practically infeasible with the number of tasks scaling up. To address this, we propose a search method that automatically assigns importance weights. We formulate it as a reinforcement learning problem and learn a task sampling schedule based on evaluation accuracy of the multi-task model. Our empirical evaluation on XNLI and GLUE shows that our method outperforms uniform sampling and the corresponding single-task baseline.

Duke Scholars

Published In

Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021

DOI

Publication Date

January 1, 2021

Start / End Page

4410 / 4415
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Sundararaman, D., Tsai, H., Lee, K. H., Turc, I., & Carin, L. (2021). Learning Task Sampling Policy for Multitask Learning. In Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021 (pp. 4410–4415). https://doi.org/10.18653/v1/2021.findings-emnlp.375

Published In

Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021

DOI

Publication Date

January 1, 2021

Start / End Page

4410 / 4415