SD-MTCNN: Self-Distilled Multi-Task

Published in British Machine Vision Conference (BMVC), 2020

Recommended citation: Ankit Jha, Awanish Kumar, Biplab Banerjee and Vinay Namboodiri, “SD-MTCNN: Self-Distilled Multi-Task CNN”, Proceedings of British Machine Vision Conference (2020), https://www.bmvc2020-conference.com/conference/papers/paper_0448.html

Download paper here

Multi-task learning (MTL) using convolutional neural networks (CNN) deals with training the network for multiple correlated tasks in concert. For accuracy-critical applications, there are endeavors to boost the model performance by resorting to a deeper network, which also increases the model complexity. However, such burdensome models are difficult to be deployed on mobile or edge devices. To ensure a trade-off between performance and complexity of CNNs in the context of MTL, we introduce the novel paradigm of self-distillation within the network. Different from traditional knowledge distillation (KD), which trains the Student in accordance with a cumbersome Teacher, our self-distilled multi-task CNN model: SD-MTCNN aims at distilling knowledge from deeper CNN layers into the shallow layers. Precisely, we follow a hard-sharing based MTL setup where all the tasks share a generic feature-encoder on top of which separate task-specific decoders are enacted. Under this premise, SD-MTCNN distills the more abstract features from the decoders to the encoded feature space, which guarantees improved multi-task performance from different parts of the network. We validate SD-MTCNN on three benchmark datasets: CityScapes, NYUv2, and Mini-Taskonomy, and results confirm the improved generalization capability of self-distilled multi-task CNNs in comparison to the literature and baselines

Recommended citation: Ankit Jha, Awanish Kumar, Biplab Banerjee and Vinay Namboodiri, “SD-MTCNN: Self-Distilled Multi-Task CNN”, Proceedings of British Machine Vision Conference (2020),