A survey of model compression strategies for object detection

53Citations
Citations of this article
62Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Deep neural networks (DNNs) have achieved great success in many object detection tasks. However, such DNNS-based large object detection models are generally computationally expensive and memory intensive. It is difficult to deploy them to devices with low memory resources or scenarios with high real-time requirements, which greatly limits their application and promotion. In recent years, many researchers have focused on compressing large object detection models without significantly degrading their performance, and have made great progress. Therefore, this paper presents a survey of object detection model compression techniques in recent years. Firstly, these compression techniques were divided into six categories: network pruning, lightweight network design, neural architecture search (NAS), low-rank decomposition, network quantization, and Knowledge distillation (KD) methods. For each category, we select some representative state-of-the-art methods and compare and analyze their performance on public datasets. After that, we discuss the application scenarios and future directions of model compression techniques. Finally, this paper is further concluded by analyzing the advantages and disadvantages of six types of model compression techniques.

Cite

CITATION STYLE

APA

Lyu, Z., Yu, T., Pan, F., Zhang, Y., Luo, J., Zhang, D., … Li, G. (2024). A survey of model compression strategies for object detection. Multimedia Tools and Applications, 83(16), 48165–48236. https://doi.org/10.1007/s11042-023-17192-x

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free