Research on Features Extraction and Classification for Images based on Transformer Learning

Wang Chao
Proceedings of 2024 International Conference on Machine Learning and Intelligent Computing, PMLR 245:67-75, 2024.

Abstract

Image processing and analysis have become an essential method in many areas including medical impact, facial recognition, and social media analysis. With the rapid development of big data and artificial intelligence technology, especially the emergence of Transformer learning models, new methods have been brought to image feature extraction and classification. However, the existing transformer model limits the ability to handle variable-length sequences and understand complex sequence relationships. In this work, we propose a novel transformer-based framework that combines a self-attention mechanism and a multi-head attention technique to efficiently extract features from complex image data. In addition, we introduce an improved classifier that enables efficient image classification using extracted features. Our method takes into account not only the local features of the image but also the global relationships between different regions to achieve a more accurate representation of the features. We simulate our model with existing convolutional neural networks and other traditional machine learning methods in the public datasets including CIFAR-10 and MNIST. From our experimental results, we can observe that our transformer-learning-based framework shows significant performance improvement in image feature extraction and classifica-tion tasks.

Cite this Paper


BibTeX
@InProceedings{pmlr-v245-chao24a, title = {Research on Features Extraction and Classification for Images based on Transformer Learning}, author = {Chao, Wang}, booktitle = {Proceedings of 2024 International Conference on Machine Learning and Intelligent Computing}, pages = {67--75}, year = {2024}, editor = {Nianyin, Zeng and Pachori, Ram Bilas}, volume = {245}, series = {Proceedings of Machine Learning Research}, month = {26--28 Apr}, publisher = {PMLR}, pdf = {https://raw.githubusercontent.com/mlresearch/v245/main/assets/chao24a/chao24a.pdf}, url = {https://proceedings.mlr.press/v245/chao24a.html}, abstract = {Image processing and analysis have become an essential method in many areas including medical impact, facial recognition, and social media analysis. With the rapid development of big data and artificial intelligence technology, especially the emergence of Transformer learning models, new methods have been brought to image feature extraction and classification. However, the existing transformer model limits the ability to handle variable-length sequences and understand complex sequence relationships. In this work, we propose a novel transformer-based framework that combines a self-attention mechanism and a multi-head attention technique to efficiently extract features from complex image data. In addition, we introduce an improved classifier that enables efficient image classification using extracted features. Our method takes into account not only the local features of the image but also the global relationships between different regions to achieve a more accurate representation of the features. We simulate our model with existing convolutional neural networks and other traditional machine learning methods in the public datasets including CIFAR-10 and MNIST. From our experimental results, we can observe that our transformer-learning-based framework shows significant performance improvement in image feature extraction and classifica-tion tasks. } }
Endnote
%0 Conference Paper %T Research on Features Extraction and Classification for Images based on Transformer Learning %A Wang Chao %B Proceedings of 2024 International Conference on Machine Learning and Intelligent Computing %C Proceedings of Machine Learning Research %D 2024 %E Zeng Nianyin %E Ram Bilas Pachori %F pmlr-v245-chao24a %I PMLR %P 67--75 %U https://proceedings.mlr.press/v245/chao24a.html %V 245 %X Image processing and analysis have become an essential method in many areas including medical impact, facial recognition, and social media analysis. With the rapid development of big data and artificial intelligence technology, especially the emergence of Transformer learning models, new methods have been brought to image feature extraction and classification. However, the existing transformer model limits the ability to handle variable-length sequences and understand complex sequence relationships. In this work, we propose a novel transformer-based framework that combines a self-attention mechanism and a multi-head attention technique to efficiently extract features from complex image data. In addition, we introduce an improved classifier that enables efficient image classification using extracted features. Our method takes into account not only the local features of the image but also the global relationships between different regions to achieve a more accurate representation of the features. We simulate our model with existing convolutional neural networks and other traditional machine learning methods in the public datasets including CIFAR-10 and MNIST. From our experimental results, we can observe that our transformer-learning-based framework shows significant performance improvement in image feature extraction and classifica-tion tasks.
APA
Chao, W.. (2024). Research on Features Extraction and Classification for Images based on Transformer Learning. Proceedings of 2024 International Conference on Machine Learning and Intelligent Computing, in Proceedings of Machine Learning Research 245:67-75 Available from https://proceedings.mlr.press/v245/chao24a.html.

Related Material