Abstract
Deep learning models find extensive applications across various domains. However, their large number of parameters, high storage requirements, and computational overhead pose challenges for deploying these models on resource-constrained embedded devices. This study focuses on addressing this issue by exploring techniques to optimize and deploy lightweight models on embedded devices. The approach involves optimization and adjustment of the model, followed by model conversion, quantization, and quantization calibration, aimed at reducing model size and improving inference speed. Notably, improvements are made to the quantization calibration algorithm to mitigate accuracy loss caused by model quantization. The experimental results demonstrate that light quantization significantly reduces model size, facilitating storage on embedded devices. Although there is a slight reduction in accuracy, the inference speed is substantially improved, enabling real-time human face recognition in video scenarios.
Original language | English |
---|---|
Title of host publication | 2023 28th International Conference on Automation and Computing |
Subtitle of host publication | ICAC 2023 |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
Number of pages | 6 |
ISBN (Electronic) | 9798350335859 |
ISBN (Print) | 9798350335866 |
DOIs | |
Publication status | Published - 16 Oct 2023 |
Event | 28th International Conference on Automation and Computing: Digitalisation for Smart Manufacturing and Systems - Aston University, Birmingham, United Kingdom Duration: 30 Aug 2023 → 1 Sep 2023 Conference number: 28 https://cacsuk.co.uk/icac/ |
Conference
Conference | 28th International Conference on Automation and Computing |
---|---|
Abbreviated title | ICAC 2023 |
Country/Territory | United Kingdom |
City | Birmingham |
Period | 30/08/23 → 1/09/23 |
Internet address |