Hierarchical Quantization Algorithm for Deep Learning Network Models

Mao Xiaoqi
Proceedings of 2024 International Conference on Machine Learning and Intelligent Computing, PMLR 245:265-270, 2024.

Abstract

Deep learning network models have achieved inspiring performances in various fields such as computer vision, natural language processing, and biomedicine. However, the high computational and storage costs of the models restrain their application in resource-limited situations. However, due to the increased com-plexity and computation of deep neural networks, there are still some challenges in deploying deep learning models into real-world applications for resource-constrained devices. To address this problem, researchers have proposed various quantization algorithms to decrease the expenditure of calculation and storage in deep learning models. This thesis addresses the problem of hierarchical quantization of deep learning models and proposes a simple hierarchical quantization algorithm that aims to effectively reduce the computation and storage requirements of deep learning network models and maintain the accuracy of the models. To demonstrate the effectiveness of the proposed hierarchical quantization method, we conducted experiments on several classical deep learning models. Our experiments prove our approach can better maintain the models’ accuracy while reduc-ing the storage and computation requirements compared to the traditional quantization algorithms.

Cite this Paper


BibTeX
@InProceedings{pmlr-v245-xiaoqi24a, title = {Hierarchical Quantization Algorithm for Deep Learning Network Models}, author = {Xiaoqi, Mao}, booktitle = {Proceedings of 2024 International Conference on Machine Learning and Intelligent Computing}, pages = {265--270}, year = {2024}, editor = {Nianyin, Zeng and Pachori, Ram Bilas}, volume = {245}, series = {Proceedings of Machine Learning Research}, month = {26--28 Apr}, publisher = {PMLR}, pdf = {https://raw.githubusercontent.com/mlresearch/v245/main/assets/xiaoqi24a/xiaoqi24a.pdf}, url = {https://proceedings.mlr.press/v245/xiaoqi24a.html}, abstract = {Deep learning network models have achieved inspiring performances in various fields such as computer vision, natural language processing, and biomedicine. However, the high computational and storage costs of the models restrain their application in resource-limited situations. However, due to the increased com-plexity and computation of deep neural networks, there are still some challenges in deploying deep learning models into real-world applications for resource-constrained devices. To address this problem, researchers have proposed various quantization algorithms to decrease the expenditure of calculation and storage in deep learning models. This thesis addresses the problem of hierarchical quantization of deep learning models and proposes a simple hierarchical quantization algorithm that aims to effectively reduce the computation and storage requirements of deep learning network models and maintain the accuracy of the models. To demonstrate the effectiveness of the proposed hierarchical quantization method, we conducted experiments on several classical deep learning models. Our experiments prove our approach can better maintain the models’ accuracy while reduc-ing the storage and computation requirements compared to the traditional quantization algorithms.} }
Endnote
%0 Conference Paper %T Hierarchical Quantization Algorithm for Deep Learning Network Models %A Mao Xiaoqi %B Proceedings of 2024 International Conference on Machine Learning and Intelligent Computing %C Proceedings of Machine Learning Research %D 2024 %E Zeng Nianyin %E Ram Bilas Pachori %F pmlr-v245-xiaoqi24a %I PMLR %P 265--270 %U https://proceedings.mlr.press/v245/xiaoqi24a.html %V 245 %X Deep learning network models have achieved inspiring performances in various fields such as computer vision, natural language processing, and biomedicine. However, the high computational and storage costs of the models restrain their application in resource-limited situations. However, due to the increased com-plexity and computation of deep neural networks, there are still some challenges in deploying deep learning models into real-world applications for resource-constrained devices. To address this problem, researchers have proposed various quantization algorithms to decrease the expenditure of calculation and storage in deep learning models. This thesis addresses the problem of hierarchical quantization of deep learning models and proposes a simple hierarchical quantization algorithm that aims to effectively reduce the computation and storage requirements of deep learning network models and maintain the accuracy of the models. To demonstrate the effectiveness of the proposed hierarchical quantization method, we conducted experiments on several classical deep learning models. Our experiments prove our approach can better maintain the models’ accuracy while reduc-ing the storage and computation requirements compared to the traditional quantization algorithms.
APA
Xiaoqi, M.. (2024). Hierarchical Quantization Algorithm for Deep Learning Network Models. Proceedings of 2024 International Conference on Machine Learning and Intelligent Computing, in Proceedings of Machine Learning Research 245:265-270 Available from https://proceedings.mlr.press/v245/xiaoqi24a.html.

Related Material