A novel model compression method based on joint distillation for deepfake video detection

In recent years, deepfake videos have been abused to create fake news, which threaten the integrity of digital videos. Although existing detection methods leveraged cumbersome neural networks to achieve promising detection performance, they cannot be deployed in resource-constrained scenarios. To ov...

Full description

Bibliographic Details
Main Authors: Xiong Xu, Shuai Tang, Mingcheng Zhu, Peisong He, Sirui Li, Yun Cao
Format: Article
Language:English
Published: Elsevier 2023-10-01
Series:Journal of King Saud University: Computer and Information Sciences
Subjects:
Online Access:http://www.sciencedirect.com/science/article/pii/S1319157823003464
_version_ 1797629735854931968
author Xiong Xu
Shuai Tang
Mingcheng Zhu
Peisong He
Sirui Li
Yun Cao
author_facet Xiong Xu
Shuai Tang
Mingcheng Zhu
Peisong He
Sirui Li
Yun Cao
author_sort Xiong Xu
collection DOAJ
description In recent years, deepfake videos have been abused to create fake news, which threaten the integrity of digital videos. Although existing detection methods leveraged cumbersome neural networks to achieve promising detection performance, they cannot be deployed in resource-constrained scenarios. To overcome this limitation, we propose a novel model compression framework based on joint distillation for deepfake detection, which includes a pre-training stage and a knowledge transfer stage. In the pre-training stage, a teacher network is trained with sufficient labeled samples. Then, in the knowledge transfer stage, a lightweight student network is constructed by considering dimension alignment. To transfer forensics knowledge comprehensively, a joint distillation loss is designed, including cross-entropy loss, knowledge distillation loss, and gradient-guided feature distillation loss. For feature distillation, feature maps from both shallow and deep layers are utilized to calculate channel-wise mean square error weighted by gradient information for transferring knowledge of forensics features adaptively. Besides, a decayed teaching strategy is constructed to adjust the importance of feature distillation, which aims at mitigating the risk of negative transfer. Extensive experiments show that student networks obtained by our model compression method can achieve competitive detection performance and outstanding efficiency by distinctly reducing computational costs compared against state-of-the-art methods.
first_indexed 2024-03-11T10:58:11Z
format Article
id doaj.art-5eadbfe5428444c6a40e317de1e812ae
institution Directory Open Access Journal
issn 1319-1578
language English
last_indexed 2024-03-11T10:58:11Z
publishDate 2023-10-01
publisher Elsevier
record_format Article
series Journal of King Saud University: Computer and Information Sciences
spelling doaj.art-5eadbfe5428444c6a40e317de1e812ae2023-11-13T04:09:06ZengElsevierJournal of King Saud University: Computer and Information Sciences1319-15782023-10-01359101792A novel model compression method based on joint distillation for deepfake video detectionXiong Xu0Shuai Tang1Mingcheng Zhu2Peisong He3Sirui Li4Yun Cao5Southwest China Institute of Electronic Technology, Chengdu 610036, ChinaSchool of Cyber Science and Engineering, Sichuan University, Chengdu 610207, ChinaDepartment of Computing, Imperial College London, London SW72AZ, United KingdomSchool of Cyber Science and Engineering, Sichuan University, Chengdu 610207, China; Corresponding author.Pittsburgh Institute, Sichuan University, Chengdu 610207, ChinaState Key Laboratory of Information Security, Institute of Information Engineering, Chinese Academy of Sciences, Beijing 100085, ChinaIn recent years, deepfake videos have been abused to create fake news, which threaten the integrity of digital videos. Although existing detection methods leveraged cumbersome neural networks to achieve promising detection performance, they cannot be deployed in resource-constrained scenarios. To overcome this limitation, we propose a novel model compression framework based on joint distillation for deepfake detection, which includes a pre-training stage and a knowledge transfer stage. In the pre-training stage, a teacher network is trained with sufficient labeled samples. Then, in the knowledge transfer stage, a lightweight student network is constructed by considering dimension alignment. To transfer forensics knowledge comprehensively, a joint distillation loss is designed, including cross-entropy loss, knowledge distillation loss, and gradient-guided feature distillation loss. For feature distillation, feature maps from both shallow and deep layers are utilized to calculate channel-wise mean square error weighted by gradient information for transferring knowledge of forensics features adaptively. Besides, a decayed teaching strategy is constructed to adjust the importance of feature distillation, which aims at mitigating the risk of negative transfer. Extensive experiments show that student networks obtained by our model compression method can achieve competitive detection performance and outstanding efficiency by distinctly reducing computational costs compared against state-of-the-art methods.http://www.sciencedirect.com/science/article/pii/S1319157823003464Video forensicsDeepfake detectionModel compressionJoint distillationGradient-guided feature distillationDecayed teaching strategy
spellingShingle Xiong Xu
Shuai Tang
Mingcheng Zhu
Peisong He
Sirui Li
Yun Cao
A novel model compression method based on joint distillation for deepfake video detection
Journal of King Saud University: Computer and Information Sciences
Video forensics
Deepfake detection
Model compression
Joint distillation
Gradient-guided feature distillation
Decayed teaching strategy
title A novel model compression method based on joint distillation for deepfake video detection
title_full A novel model compression method based on joint distillation for deepfake video detection
title_fullStr A novel model compression method based on joint distillation for deepfake video detection
title_full_unstemmed A novel model compression method based on joint distillation for deepfake video detection
title_short A novel model compression method based on joint distillation for deepfake video detection
title_sort novel model compression method based on joint distillation for deepfake video detection
topic Video forensics
Deepfake detection
Model compression
Joint distillation
Gradient-guided feature distillation
Decayed teaching strategy
url http://www.sciencedirect.com/science/article/pii/S1319157823003464
work_keys_str_mv AT xiongxu anovelmodelcompressionmethodbasedonjointdistillationfordeepfakevideodetection
AT shuaitang anovelmodelcompressionmethodbasedonjointdistillationfordeepfakevideodetection
AT mingchengzhu anovelmodelcompressionmethodbasedonjointdistillationfordeepfakevideodetection
AT peisonghe anovelmodelcompressionmethodbasedonjointdistillationfordeepfakevideodetection
AT siruili anovelmodelcompressionmethodbasedonjointdistillationfordeepfakevideodetection
AT yuncao anovelmodelcompressionmethodbasedonjointdistillationfordeepfakevideodetection
AT xiongxu novelmodelcompressionmethodbasedonjointdistillationfordeepfakevideodetection
AT shuaitang novelmodelcompressionmethodbasedonjointdistillationfordeepfakevideodetection
AT mingchengzhu novelmodelcompressionmethodbasedonjointdistillationfordeepfakevideodetection
AT peisonghe novelmodelcompressionmethodbasedonjointdistillationfordeepfakevideodetection
AT siruili novelmodelcompressionmethodbasedonjointdistillationfordeepfakevideodetection
AT yuncao novelmodelcompressionmethodbasedonjointdistillationfordeepfakevideodetection