Abstract
Transfer learning is a popular software reuse technique in the deep learning community that enables developers to build custom mod-els (students) based on sophisticated pretrained models (teachers). However, like vulnerability inheritance in traditional software reuse, some defects in the teacher model may also be inherited by students, such as well-known adversarial vulnerabilities and backdoors. Re-ducing such defects is challenging since the student is unaware of how the teacher is trained and/or attacked. In this paper, we propose ReMoS, a relevant model slicing technique to reduce defect inheri-tance during transfer learning while retaining useful knowledge from the teacher model. Specifically, ReMoS computes a model slice (a subset of model weights) that is relevant to the student task based on the neuron coverage information obtained by profiling the teacher model on the student task. Only the relevant slice is used to fine-tune the student model, while the irrelevant weights are retrained from scratch to minimize the risk of inheriting defects. Our experi-ments on seven DNN defects, four DNN models, and eight datasets demonstrate that ReMoS can reduce inherited defects effectively (by 63% to 86% for CV tasks and by 40% to 61 % for NLP tasks) and efficiently with minimal sacrifice of accuracy (3% on average).
Author supplied keywords
Cite
CITATION STYLE
Zhang, Z., Li, Y., Wang, J., Liu, B., Li, D., Guo, Y., … Liu, Y. (2022). ReMoS: Reducing Defect Inheritance in Transfer Learning via Relevant Model Slicing. In Proceedings - International Conference on Software Engineering (Vol. 2022-May, pp. 1856–1868). IEEE Computer Society. https://doi.org/10.1145/3510003.3510191
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.