Tchebycheff Procedure for Multi-task Text Classification

Yuren Mao, Shuang Yun, Weiwei Liu, Bo Du

Abstract Paper Share

Machine Learning for NLP Long Paper

Session 7B: Jul 7 (09:00-10:00 GMT)
Session 8B: Jul 7 (13:00-14:00 GMT)
Abstract: Multi-task Learning methods have achieved great progress in text classification. However, existing methods assume that multi-task text classification problems are convex multiobjective optimization problems, which is unrealistic in real-world applications. To address this issue, this paper presents a novel Tchebycheff procedure to optimize the multi-task classification problems without convex assumption. The extensive experiments back up our theoretical analysis and validate the superiority of our proposals.
You can open the pre-recorded video in a separate window.
NOTE: The SlidesLive video may display a random order of the authors. The correct author list is shown at the top of this webpage.

Similar Papers

Towards Emotion-aided Multi-modal Dialogue Act Classification
Tulika Saha, Aditya Patra, Sriparna Saha, Pushpak Bhattacharyya,
A representative figure from paper main.402
A Novel Graph-based Multi-modal Fusion Encoder for Neural Machine Translation
Yongjing Yin, Fandong Meng, Jinsong Su, Chulun Zhou, Zhengyuan Yang, Jie Zhou, Jiebo Luo,
A representative figure from paper main.273
The Dialogue Dodecathlon: Open-Domain Knowledge and Image Grounded Conversational Agents
Kurt Shuster, Da JU, Stephen Roller, Emily Dinan, Y-Lan Boureau, Jason Weston,
A representative figure from paper main.222
Dynamic Sampling Strategies for Multi-Task Reading Comprehension
Ananth Gottumukkala, Dheeru Dua, Sameer Singh, Matt Gardner,
A representative figure from paper main.86