TY - JOUR
T1 - Efficient hyperspectral image segmentation for biosecurity scanning using knowledge distillation from multi-head teacher
AU - Phan, Minh Hieu
AU - Phung, Son Lam
AU - Luu, Khoa
AU - Bouzerdoum, Abdesselam
N1 - Publisher Copyright:
© 2022 Elsevier B.V.
PY - 2022/9/14
Y1 - 2022/9/14
N2 - Foreign species can deteriorate the environment and the economy of a country. To automatically monitor biosecurity threats at country borders, this paper investigates compact deep networks for accurate and real-time object segmentation for hyperspectral images. To this end, knowledge distillation (KD) approaches compress the model by distilling the knowledge of a large teacher network to a compact student network. However, when the student is over-compressed, the performance of standard KD methods degrades significantly due to the large capacity gap between the teacher and the student. This gap can be addressed by adding medium-sized teacher assistants, but training them incurs significant computation and hence is impractical. To address this problem, this paper proposes a new framework called Knowledge Distillation from Multi-head Teacher (KDM), which distills the knowledge of a multi-head teacher to the student. By encapsulating multiple teachers in a single network, our proposed KDM assists the learning of a very compact student and significantly reduces the training time. We also introduce Bio-HSI, a new large benchmark hyperspectral image dataset of 3,125 high-resolution images with dense seg-mentation ground truth. This new, large dataset can be expected to advance research on deep models for hyperspectral image segmentation. Evaluated on this dataset, the student trained via our KDM has 762 times fewer parameters than the state-of-the-art segmentation model (i.e., HRNet), while achieving com-petitive accuracy. (c) 2022 Elsevier B.V. All rights reserved.
AB - Foreign species can deteriorate the environment and the economy of a country. To automatically monitor biosecurity threats at country borders, this paper investigates compact deep networks for accurate and real-time object segmentation for hyperspectral images. To this end, knowledge distillation (KD) approaches compress the model by distilling the knowledge of a large teacher network to a compact student network. However, when the student is over-compressed, the performance of standard KD methods degrades significantly due to the large capacity gap between the teacher and the student. This gap can be addressed by adding medium-sized teacher assistants, but training them incurs significant computation and hence is impractical. To address this problem, this paper proposes a new framework called Knowledge Distillation from Multi-head Teacher (KDM), which distills the knowledge of a multi-head teacher to the student. By encapsulating multiple teachers in a single network, our proposed KDM assists the learning of a very compact student and significantly reduces the training time. We also introduce Bio-HSI, a new large benchmark hyperspectral image dataset of 3,125 high-resolution images with dense seg-mentation ground truth. This new, large dataset can be expected to advance research on deep models for hyperspectral image segmentation. Evaluated on this dataset, the student trained via our KDM has 762 times fewer parameters than the state-of-the-art segmentation model (i.e., HRNet), while achieving com-petitive accuracy. (c) 2022 Elsevier B.V. All rights reserved.
KW - Deep learning
KW - Hyperspectral image segmentation
KW - Knowledge distillation
UR - https://www.scopus.com/pages/publications/85134627366
U2 - 10.1016/j.neucom.2022.06.095
DO - 10.1016/j.neucom.2022.06.095
M3 - Article
AN - SCOPUS:85134627366
SN - 0925-2312
VL - 504
SP - 189
EP - 203
JO - Neurocomputing
JF - Neurocomputing
ER -