/Awesome-Multitask-Learning

This repository periodicly updates the MTL paper and resources

Awesome-Multitask-Learning

Our survey primarily aims to provide a comprehensive understanding of MTL, encompassing its definition, taxonomy, applications, and their connections and trends. We delve into the various aspects of MTL methods, including the loss function, network architecture, and optimization methods, offering explanations and insights from the perspective of technical details. For each method, we provide the corresponding paper link, as well as the code repository for the MTL methods discussed in the paper. We sincerely hope that this survey aids in your comprehension of MTL and its associated methods. If you have any questions or suggestions, please feel free to contact us.

@article{yu2024unleashing,
title={Unleashing the Power of Multi-Task Learning: A Comprehensive Survey Spanning Traditional, Deep, and Pre-Trained Foundation Model Eras},
author={Yu, Jun and Dai, Yutong and Liu, Xiaokang and Huang, Jin and Shen, Yishan and Zhang, Ke and Zhou, Rong and Adhikarla, Eashan and Ye, Wenxuan and Liu, Yixin and others},
journal={arXiv preprint arXiv:2404.18961},
year={2024}
}

Table of Contents:

Existing survey papers

  • A survey on multi-task learning
    Yu Zhang and Qiang Yang IEEE Transactions on Knowledge and Data Engineering 2022. [Paper]
    March 31, 2021
  • Multi-task learning for dense prediction tasks: A survey
    Simon Vandenhende, Stamatios Georgoulis, Marc Proesmans, Dengxin Dai and Luc Van Gool
    IEEE Transactions on Pattern Analysis and Machine Intelligence 2021. [Paper]
    Jan 26, 2021
  • A Brief Review of Deep Multi-task Learning and Auxiliary Task Learning
    Partoo Vafaeikia and Khashayar Namdar and Farzad Khalvati
    arXiv 2020. [Paper]
    Jul 02, 2020
  • Multi-Task Learning with Deep Neural Networks: A Survey
    Michael Crawshaw
    arXiv 2020. [Paper]
    Sep 10, 2020
  • A brief review on multi-task learning [Paper]
    Kim-Han Thung, Chong-Yaw Wee
    Multimedia Tools and Applications 2018.
    Aug 08, 2018
  • An overview of multi-task learning in deep neural networks
    Sebastian Ruder
    arXiv 2017. [Paper]
    Jun 15, 2017

Datasets

Regression task

  • Synthetic Data This dataset is often artificially defined by researchers, thus different from one another. The features are often generated via drawing random variables from a shared distribution and adding irrelevant variants from other distributions, and the corresponding responses are produced by a specific computational method.

  • School Data This dataset comes from the Inner London Education Authority (ILEA) and contains 15,362 records of student examination, which are described by 27 student- and school-specific features from 139 secondary schools. The goal is to predict exam scores from 27 features.

  • SARCOS Data This dataset is in humanoid robotics consists of 44,484 training examples and 4, 449 test examples. The goal of learning is to estimate inverse dynamics model of a 7 degrees-of-freedom (DOF) SARCOS anthropomorphic robot arm.

  • Computer Survey Data This dataset is from a survey on the likelihood (11 point scale from 0 to 10) of purchasing personal computers. There are 20 computer models as examples, each of which contains 13 computer descriptions (e.g., price, CPU speed, and screen size) and 6 subject-level covariates (e.g., gender, computer knowledge, and work experience) as features and ratings of 179 subjects as targets, i.e., tasks.

  • Climate Dataset This real-time dataset is collected from a sensor network (e.g., anemometer, thermistor, and pressure transducer) of four climate stations—Cambermet, Chimet, Sotonmet and Bramblemet—in the south on England, which can represent 4 tasks as needed. The archived data are reported in 5-minute intervals, including ∼ 10 climate signals (e.g., wind speed, wave period, barometric pressure, and water temperature).

Classification task

  • 20 Newsgroups This dataset is a collection of approximately 19, 000 netnews articles, organized into 20 hierarchical newsgroups according to the topic, such as root categories (e.g., comp,rec, sci, and talk) and sub-categories (e.g., comp.graphics, sci.electronics, and talk.politics.guns). Users can design different combinations as multiple text classifications tasks.

  • Reuters-21578 Collection This text collection contains 21578 documents from Reuters newswire dating back to 1987. These documents were assembled and indexed with more than 90 correlated categories—5 top categories (i.e., exchanges, orgs, people, place, topic), and each of them includes variable sub-categories。

  • MultiMNIST This dataset is a MTL version of MNIST dataset9. By overlaying multiple images together, traditional digit classification is converted to a MTL problem, where classifying the digits on the different positions are considered as distinctive tasks.

  • ImageCLEF-2014 This dataset is a benchmark for domain adaptation challenge, which contains 2, 400 images of 12 common categories selected from 4 domains: Caltech 256, ImageNet 2012, Pascal VOC 2012, and Bing.

  • Office-Caltech This dataset is a standard benchmark for domain adaption in computer vision, consisting of real-world images of 10 common categories from Office dataset and Caltech-256 dataset. There are 2,533 images from 4 distinct domains/tasks: Amazon, DSLR, Webcam, and Caltech.

  • Office-31 This dataset consists of 4,110 images from 31 object categories across 3 domains/tasks: Amazon, DSLR, and Webcam.

  • Office-Home Dataset. This dataset is collected for object recognition to validate domain adaptation models in the era of deep learning, which includes 15,588 images images in office and home settings (e.g., alarm clock, chair, eraser, keyboard, telephone, etc.) organized into 4 domains/tasks: Art (paintings, sketches and artistic depictions), Clipart (clipart images), Product (product images from www.amazon.com), and Real-World (real-world objects captured with a regular camera).

  • DomainNet This dataset is annotated for the purpose of multi-source unsupervised domain adaptation (UDA) research. It contains ∼ 0.6 million images from 345 categories across 6 distinct domains, e.g., sketch, infograph, quickdraw, real, etc.

  • EMMa This dataset comprises more than 2.8 million objects from Amazon product listings, each annotated with images, listing text, mass, price, product ratings, and its position in Amazon’s product-category taxonomy. It includes a comprehensive taxonomy of 182 physical materials, and objects are annotated with one or more materials from this taxonomy. EMMa offers a new benchmark for multi-task learning in computer vision and NLP, allowing for the addition of new tasks and object attributes at scale.

  • SYNTHIA This dataset is a synthetic dataset created to address the need for a large and diverse collection of images with pixel-level annotations for vision-based semantic segmentation in urban scenarios, particularly for autonomous driving applications. It consists of precise pixel-level semantic annotations for 13 classes, including sky, building, road, sidewalk, fence, vegetation, lane-marking, pole, car, traffic signs, pedestrians, cyclists, and miscellaneous objects.

Dense prediction task

  • CityScapes This dataset consists of 5,000 images with high quality annotations and 20,000 images with coarse annotations from 50 different cities, which contains 19 classes for semantic urban scene understanding.

  • NYU-Depth Dataset V2 This dataset is comprised of 1,449 images from 464 indoor scenes across 3 cities, which contains 35,064 distinct objects of 894 different classes. The dense per-pixel labels of class, instance, and depth are used in many computer vision tasks, e.g., semantic segmentation, depth prediction, and surface normal estimation.

  • PASCAL VOC Project This project provides standardized image datasets for object class recognition and also has run challenges evaluating performance on object class recognition from 2005 to 2012, where VOC07, VOC08, and VOC12 are commonly used for MTL research. The multiple tasks covers classification, detection (e.g., body part, saliency, semantic edge), segmentation, attribute prediction, surface normals prediction, etc.

  • Taskonomy This dataset is currently the most diverse product for computer vision in MTL, consisting of 4 million samples from 3D scans of ∼ 600 buildings. This product is a dictionary of 26 tasks (e.g., 2D, 2.5D, 3D, semantics, etc.) as a computational taxonomic map for task transfer learning.

Methods

Traditional era

Feature Selection

  • Adaptive multi-task sparse learning with an application to fMRI study [Paper]
    Xi Chen, Jinghui He, Rick Lawrence and Jaime G Carbonell
    Proceedings of the 2012 SIAM International Conference on Data Mining 2012.

  • Multi-stage multi-task feature learning [Paper]
    Pinghua Gong, Jieping Ye and Changshui Zhang
    Advances in neural information processing systems 2012.

  • Sparse Multi-Task Lasso
    Aurelie C Lozano and Grzegorz Swirszcz
    Proceedings of the 29th International Coference on International Conference on Machine Learning 2012. [Paper]

  • Modeling disease progression via fused sparse group lasso
    Jiayu Zhou, Jun Liu, Vaibhav A Narayan and Jieping Ye
    Proceedings of the 18th ACM SIGKDD international conference on Knowledge discovery and data mining 2012. [Paper]

  • A multi-task learning formulation for predicting disease progression
    Jiayu Zhou, Lei Yuan, Jun Liu and Jieping Ye
    Proceedings of the 17th ACM SIGKDD international conference on Knowledge discovery and data mining 2011. [Paper]

  • Dirty Block-Sparse Model
    Ali Jalali, Sujay Sanghavi, Chao Ruan and Pradeep Ravikumar
    Advances in neural information processing systems 2010. [Paper]

  • Adaptive Sparse Multi-Task Lasso
    Seunghak Lee, Jun Zhu and Eric Xing
    Advances in neural information processing systems 2010. [Paper]

  • Multi-Task Feature Selection
    Guillaume Obozinski, Ben Taskar and Michael Jordan
    researchgate 2006. [Paper]

  • A probabilistic framework for multi-task learning
    Jian Zhang
    Ph.D. Thesis 2006. [Paper]

Feature Transformation

  • Multi-task learning for multiple language translation [Paper]
    Daxiang Dong, Hua Wu, Wei He, Dianhai Yu and Haifeng Wang
    Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing 2015.

  • A convex formulation for learning shared structures from multiple tasks [Paper]
    Jianhui Chen, Lei Tang, Jun Liu and Jieping Ye
    Proceedings of the 26th annual international conference on machine learning 2009.

  • Multi-task feature learning [Paper]
    Andreas Argyriou, Theodoros Evgeniou and Massimiliano Pontil
    Advances in neural information processing systems 2006.

  • A framework for learning predictive structures from multiple tasks and unlabeled data [Paper]
    Rie Kubota Ando, Tong Zhang and Peter Bartlett
    Journal of Machine Learning Research 2005.

Low-Rank Factorization

  • Learning Linear and Nonlinear Low-Rank Structure in Multi-Task Learning [Paper]
    Yi Zhang, Yu Zhang and Wei Wang
    IEEE Transactions on Knowledge and Data Engineering 2023

  • Multi-stage multi-task learning with reduced rank [Paper]
    Lei Han and Yu Zhang
    Proceedings of the AAAI Conference on Artificial Intelligence 2016.

  • Multitask learning meets tensor factorization: task imputation via convex optimization [Paper]
    Kishan Wimalawarne, Masashi Sugiyama and Ryota Tomioka
    Advances in neural information processing systems 2014

  • Multilinear multitask learning [Paper]
    Bernardino Romera-Paredes, Hane Aung, Nadia Bianchi-Berthouze and Massimiliano Pontil
    Proceedings of the 30th International Conference on Machine Learning 2013

  • An accelerated gradient method for trace norm minimization [Paper]
    Shuiwang Ji and Jieping Ye
    Proceedings of the 26th annual international conference on machine learning 2009.

Decomposition

  • Learning incoherent sparse and low-rank patterns from multiple tasks [Paper]
    Jianhui Chen, Ji Liu and Jieping Ye
    ACM Transactions on Knowledge Discovery from Data 2012.

  • Multi-level lasso for sparse multi-task regression [Paper]
    Aurelie C Lozano and Grzegorz Swirszcz
    Proceedings of the 29th International Coference on International Conference on Machine Learning 2012.

  • Robust multi-task feature learning [Paper]
    Pinghua Gong, Jieping Ye and Changshui Zhang
    Proceedings of the 18th ACM SIGKDD international conference on Knowledge discovery and data mining 2012.

  • Integrating low-rank and group-sparse structures for robust multi-task learning [Paper]
    Jianhui Chen, Jiayu Zhou and Jieping Ye
    Proceedings of the 17th ACM SIGKDD international conference on Knowledge discovery and data mining 2011.

  • A dirty model for multi-task learning [Paper]
    Ali Jalali, Sujay Sanghavi, Chao Ruan and Pradeep Ravikumar
    Advances in neural information processing systems 2010.

  • A framework for learning predictive structures from multiple tasks and unlabeled data. [Paper]
    Rie Kubota Ando, Tong Zhang and Peter Bartlett
    Journal of Machine Learning Research 2005

Priori Sharing

  • A convex formulation for learning task relationships in multi-task learning. [Paper]
    Zhang, Yu and Yeung, Dit--Yan
    arXiv preprint arXiv:1203.3536 2012

  • Hierarchical multitask structured output learning for large-scale sequence segmentation. [Paper]
    G{"o}rnitz, Nico and Widmer, Christian and Zeller, Georg and Kahles, Andr{'e} and R{"a}tsch, Gunnar and Sonnenburg, S{"o}ren
    Advances in Neural Information Processing Systems 2011

  • Large margin multi-task metric learning. [Paper]
    Evgeniou, Theodoros and Pontil, Massimiliano
    Advances in neural information processing systems 2010

  • Multi-task learning using generalized t process. [Paper]
    Zhang, Yu and Yeung, Dit--Yan
    JMLR Workshop and Conference Proceedings 2010

  • Multi-task learning via conic programming. [Paper]
    Kato, Tsuyoshi and Kashima, Hisashi and Sugiyama, Masashi and Asai, Kiyoshi
    Advances in Neural Information Processing Systems 2007

  • Multi-task Gaussian process prediction. [Paper]
    Bonilla, Edwin V and Chai, Kian and Williams, Christopher
    Advances in Neural Information Processing Systems 2007

  • Learning multiple tasks with kernel methods. [Paper]
    Evgeniou, Theodoros and Micchelli, Charles A and Pontil, Massimiliano and Shawe-Taylor, John
    Journal of machine learning research 2005

  • Regularized multi--task learning. [Paper]
    Parameswaran, Shibin and Weinberger, Kilian Q
    Proceedings of the tenth ACM SIGKDD international conference on Knowledge discovery and data mining 2004

Task Clustering

  • Learning tree structure in multi-task learning. [Paper]
    Han, Lei and Zhang, Yu
    Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 2015

  • Clustered Multi-Task Learning Via Alternating Structure Optimization. [Paper]
    Zhou, Jiayu and Chen, Jianhui and Ye, Jieping
    Advances in Neural Information Processing Systems 2011

  • A Framework for Learning Predictive Structures from Multiple Tasks and Unlabeled Data. [Paper]
    Ando, Rie Kubota and Zhang, Tong
    The Journal of Machine Learning Research 2005

Deep Learning Era

Feature Fusion

Latent multi-task architecture learning [Paper]
Authors: Sebastian Ruder, Joachim Bingel, Isabelle Augenstein and Anders Sogaard
Publisher: Proceedings of the AAAI Conference on Artificial Intelligence
Year: 2019
Description of image
Nddr-cnn: Layerwise feature fusing in multi-task cnns by neural discriminative dimensionality reduction [Paper] [Code]
Authors: Yuan Gao, Jiayi Ma, Mingbo Zhao , Wei Liu and Alan L Yuille
Publisher: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition
Year: 2019
Description of image
Cross-stitch networks for multi-task learning [Paper] [Code]
Authors: Ishan Misra, Abhinav Shrivastava, Abhinav Gupta and Martial Hebert
Publisher: CVPR
Year: 2016
Description of image

Cascading

Deep cascade multi-task learning for slot filling in online shopping assistant [Paper] [Code]
Authors: Yu Gong, Xusheng Luo, Yu Zhu, Wenwu Ou, Zhao Li, Muhua Zhu, Kenny Q. Zhu, Lu Duan and Xi Chen
Publisher: Proceedings of the AAAI conference on artificial intelligence
Year: 2019
Description of image
A hierarchical multi-task approach for learning embeddings from semantic tasks [Paper] [Code]
Authors: Victor Sanh, Thomas Wolf and Sebastian Ruder
Publisher: Proceedings of the AAAI conference on artificial intelligence
Year: 2019
Description of image
Deep multi-task learning with low level tasks supervised at lower layers [Paper]
Authors: Anders S{\o}gaard and Yoav Goldberg
Publisher: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics Volume 2: Short Papers
Year: 2016
Description of image
Instance-aware semantic segmentation via multi-task network cascades [Paper] [Code]
Authors: Jifeng Dai, Kaiming He and Jian Sun
Publisher: Proceedings of the IEEE conference on computer vision and pattern recognition
Year: 2016
Description of image
A Joint Many-Task Model: Growing a Neural Network for Multiple NLP Tasks [Paper] [Code]
Authors: Kazuma Hashimoto, Caiming Xiong, Yoshimasa Tsuruoka, Richard Socher
Publisher: arXiv preprint arXiv:1611.01587
Year: 2016
Description of image

Knowledge Distilation

Online Knowledge Distillation for Multi-Task Learning [Paper]
Authors: Geethu Miriam Jacob and Vishal Agarwal
Publisher: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision
Year: 2023
Description of image
Cross-task knowledge distillation in multi-task recommendation [Paper]
Authors: Chenxiao Yang, Junwei Pan, Xiaofeng Gao, Tingyu Jiang, Dapeng Liu and Guihai Chen
Publisher: Proceedings of the AAAI Conference on Artificial Intelligence
Year: 2022
Description of image
Multi-task self-training for learning general representations [Paper]
Authors: Golnaz Ghiasi, Barret Zoph, Ekin D Cubuk and Quoc V Le and Tsung-Yi Lin
Publisher: Proceedings of the IEEE/CVF International Conference on Computer Vision
Year: 2021
Description of image
Knowledge distillation for multi-task learning [Paper] [Code]
Authors: WeiHong Li and Hakan Bilen
Publisher: Computer Vision--ECCV
Year: 2020
Description of image

Cross-Task Attention

Demt: Deformable mixer transformer for multi-task learning of dense prediction [Paper] [Code]
Authors: Yangyang Xu, Yibo Yang and Lefei Zhang
Publisher: arXiv e-prints
Year: 2023
Description of image
Exploring relational context for multi-task dense prediction [Paper] [Code]
Authors: David Bruggemann, Menelaos Kanakis, Anton Obukhov, Stamatios Georgoulis and Luc Van Gool
Publisher: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
Year: 2021
Description of image
Pattern-structure diffusion for multi-task learning [Paper]
Authors: Ling Zhou, Zhen Cui, Chunyan Xu, Zhenyu Zhang, Chaoqun Wang, Tong Zhang and Jian Yang
Publisher: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
Year: 2020
Description of image
Mti-net: Multi-scale task interaction networks for multi-task learning [Paper] [Code]
Authors: Simon Vandenhende, Stamatios Georgoulis and Luc Van Gool
Publisher: Computer Vision--ECCV 2020: 16th European Conference, Glasgow, UK, August 23--28, 2020, Proceedings, Part IV 16
Year: 2020
Description of image
End-to-end multi-task learning with attention [Paper] [Code]
Authors: Shikun Liu, Edward Johns and Andrew J Davison
Publisher: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
Year: 2019
Pattern-affinitive propagation across depth, surface normal and semantic segmentation [Paper]
Authors: Zhenyu Zhang, Zhen Cui, Chunyan Xu, Yan Yan, Nicu Sebe and Jian Yang
Publisher: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
Year: 2019
Description of image
Attentive single-tasking of multiple tasks [Paper] [Code]
Authors: Kevis-Kokitsi Maninis, Ilija Radosavovic and Iasonas Kokkinos
Publisher: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
Year: 2019
Description of image
Pad-net: Multi-tasks guided prediction-and-distillation network for simultaneous depth estimation and scene parsing [Paper]
Authors: Dan Xu, Wanli Ouyang, Xiaogang Wang and Nicu Sebe
Publisher: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
Year: 2018
Description of image

Scalarization

Recon: Reducing Conflicting Gradients From the Root For Multi-Task Learning [Paper] [Code]
Authors: Guangyuan Shi, Qimai Li, Wenlong Zhang, Jiaxin Chen, Xiao-Ming Wu
Publisher: The Eleventh International Conference on Learning Representations
Year: 2022
Towards impartial multi-task learning [Paper] [Code]
Authors: Liyang Liu, Yi Li, Zhanghui Kuang, Jing-Hao Xue, Yimin Chen, Wenming Yang, Qingmin Liao, Wayne Zhang
Publisher: ICLR
Year: 2021
Gradient surgery for multi-task learning [Paper] [Code]
Authors: Tianhe Yu, Saurabh Kumar, Abhishek Gupta, Sergey Levine, Karol Hausman, Chelsea Finn
Publisher: Advances in Neural Information Processing Systems
Year: 2020
Description of image
Multi-task learning using uncertainty to weigh losses for scene geometry and semantics [Paper] [Code]
Authors: Alex Kendall, Yarin Gal and Roberto Cipollae
Publisher: Proceedings of the IEEE conference on computer vision and pattern recognition
Year: 2018
Gradnorm: Gradient normalization for adaptive loss balancing in deep multitask networks [Paper] [Code]
Authors: Zhao Chen, Vijay Badrinarayanan, Chen-Yu Lee and Andrew Rabinovich
Publisher: International conference on machine learning
Year: 2018

Multi-objective Optimization (MOO)

Mitigating gradient bias in multi-objective learning: A provably convergent approach [Paper] [Code]
Authors:Heshan Devaka Fernando,Han Shen, Miao Liu,Subhajit Chaudhury, Keerthiram Murugesan and Tianyi Chen
Publisher:The Eleventh International Conference on Learning Representations
Year:2022
Multi-task learning as a bargaining game [Paper] [Code]
Authors:Aviv Navon, Aviv Shamsian, Idan Achituve, Haggai Maron, Kenji Kawaguchi, Gal Chechik and Ethan Fetaya
Publisher:arXiv preprint arXiv:2202.01017
Year:2022
Multi-task learning with user preferences: Gradient descent with controlled ascent in pareto optimization [Paper] [Code]
Authors:Debabrata Mahapatra and Vaibhav Rajan
Publisher:International Conference on Machine Learning
Year:2020
Pareto multi-task learning [Paper] [Code]
Authors:Xi Lin, Hui-Ling Zhen, Zhenhua Li, Qing-Fu Zhang and Sam Kwong
Publisher:Advances in neural information processing systems
Year:2019
Multi-task learning as multi-objective optimization [Paper] [Code]
Authors:Ozan Sener and Vladlen Koltun
Publisher:Advances in neural information processing systems
Year:2018
Multicriteria optimization [Paper]
Authors:Matthias Ehrgott
Publisher:Springer Science \& Business Media
Year:2005
Steepest descent methods for multicriteria optimization [Paper]
Authors:J{\"o}rg Fliege and Benar Fux Svaiter
Publisher:Springer
Year:2000

Adversarial Training

Representation disentanglement for multi-task learning with application to fetal ultrasound [Paper] [Code]
Authors:Qingjie Meng, Nick Pawlowski, Daniel Rueckert, Bernhard Kainz
Publisher: Springer
Year: 2019
Description of image
Multi-task adversarial network for disentangled feature learning [Paper]
Authors: Yang Liu, Zhaowen Wang, Hailin Jin, Ian Wassell
Publisher: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
Year: 2018
Description of image
Gradient adversarial training of neural networks [Paper]
Authors:Ayan Sinha, Zhao Chen, Vijay Badrinarayanan, Andrew Rabinovich
Publisher: arXiv preprint arXiv:1806.08028
Year: 2018
Description of image
Adversarial Multi-task Learning for Text Classification [Paper] [Code]
Authors: Pengfei Liu, Xipeng Qiu, Xuanjing Huang
Publisher: arXiv preprint arXiv:1704.05742
Year: 2017
Description of image

Mixture of Experts

Mod-Squad: Designing Mixtures of Experts As Modular Multi-Task Learners [Paper]
Authors: Zitian Chen, Yikang Shen, Mingyu Ding, Zhenfang Chen, Hengshuang Zhao, Erik Learned-Miller, Chuang Gan
Publisher: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
Year: 2023
AdaMV-MoE: Adaptive Multi-Task Vision Mixture-of-Experts [Paper] [Code]
Authors: Tianlong Chen, Xuxi Chen, Xianzhi Du, Abdullah Rashwan, Fan Yang, Huizhong Chen, Zhangyang Wang, Yeqing Li
Publisher: Proceedings of the IEEE/CVF International Conference on Computer Vision
Year: 2023
SSummaReranker: A multi-task mixture-of-experts re-ranking framework for abstractive summarization [Paper]
Authors: Mathieu Ravaut, Shafiq Joty, Nancy F. Chen
Publisher: arXiv preprint arXiv:2203.06569
Year: 2022
Multi-task learning with calibrated mixture of insightful experts [Paper]
Authors: Sinan Wang, Yumeng Li, Hongyan Li, Tanchao Zhu, Zhao Li, Wenwu Ou
Publisher: 2022 IEEE 38th International Conference on Data Engineering (ICDE)
Year: 2022
Eliciting transferability in multi-task learning with task-level mixture-of-experts [Paper] [Code]
Authors: Qinyuan Ye, Juan Zha, Xiang Ren
Publisher: arXiv preprint arXiv:2205.12701
Year: 2022
Description of image
Eliciting and Understanding Cross-Task Skills with Task-Level Mixture-of-Experts [Paper] [Code]
Authors: Qinyuan Ye, Juan Zha, Xiang Ren
Publisher: arXiv preprint arXiv:2205.12701
Year: 2022
M³ViT: Mixture-of-Experts Vision Transformer for Efficient Multi-task Learning with Model-Accelerator Co-design [Paper] [Code]
Authors: Hanxue Liang, Zhiwen Fan, Rishov Sarkar, Ziyu Jiang, Tianlong Chen, Kai Zou, Yu Cheng, Cong Hao, Zhangyang Wang
Publisher: Advances in Neural Information Processing Systems
Year: 2022
Dselect-k: Differentiable selection in the mixture of experts with applications to multi-task learning [Paper] [Code]
Authors: Hussein Hazimeh, Zhe Zhao, Aakanksha Chowdhery, Maheswaran Sathiamoorthy, Yihua Chen, Rahul Mazumder, Lichan Hong, Ed H. Chi
Publisher: Advances in Neural Information Processing Systems
Year: 2021
Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners [Paper]
Authors: Shashank Gupta, Subhabrata Mukherjee, Krishan Subudhi, Eduardo Gonzalez, Damien Jose, Ahmed H. Awadallah, Jianfeng Gao
Publisher: arXiv preprint arXiv:2204.07689
Year: 2021
Modeling task relationships in multi-task learning with multi-gate mixture-of-experts [Paper] [Code]
Authors: Jiaqi Ma, Zhe Zhao, Xinyang Yi, Jilin Chen, Lichan Hong and Ed H Chi
Publisher: Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery \& data mining
Year: 2018
Description of image
Outrageously large neural networks: The sparsely-gated mixture-of-experts layer [Paper]
Authors: Noam Shazeer, Azalia Mirhoseini, Krzysztof Maziarz, Andy Davis, Quoc Le,, Geoffrey Hinton and Jeff Dean
Publisher: arXiv preprint arXiv:1701.06538
Year: 2017
Description of image

Graph-based

Relational Multi-Task Learning: Modeling Relations between Data and Tasks [Paper] [Code]
Authors: Kaidi Cao, Jiaxuan You, Jure Leskovec
Publisher: ParXiv preprint arXiv:2303.07666
Year: 2023
Description of image
Multi-label image recognition with graph convolutional networks [Paper] [Code]
Authors: Zhao-Min Chen, Xiu-Shen Wei, Peng Wang, Yanwen Guo
Publisher: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition
Year: 2019
Description of image
Leveraging sequence classification by taxonomy-based multitask learning [Paper]
Authors: Christian Widmer, Jose Leiva, Yasemin Altun, Gunnar Rätsch
Publisher: Research in Computational Molecular Biology: 14th Annual International Conference, RECOMB 2010, Lisbon, Portugal, April 25-28, 2010. Proceedings 14
Year: 2010
-->

Visitor Count: Visitor Count