Handwriting Image Classification for Automated Diagnosis of Learning Disabilities: A Review on Deep Learning Models and Future Directions
This study reviews deep learning models used in handwriting image classification for the automated diagnosis of learning disabilities. By addressing handwriting diversity and misclassification challenges, two models were highlighted: Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs...
Published in: | 19th International Joint Symposium on Artificial Intelligence and Natural Language Processing, iSAI-NLP 2024 |
---|---|
Main Author: | |
Format: | Conference paper |
Language: | English |
Published: |
Institute of Electrical and Electronics Engineers Inc.
2024
|
Online Access: | https://www.scopus.com/inward/record.uri?eid=2-s2.0-85216562911&doi=10.1109%2fiSAI-NLP64410.2024.10799245&partnerID=40&md5=94ad7c5e822a95e47c2061ced489ffc5 |
id |
2-s2.0-85216562911 |
---|---|
spelling |
2-s2.0-85216562911 Sukiman S.A.; Husin N.A.; Hamdan H.; Murad M.A.A. Handwriting Image Classification for Automated Diagnosis of Learning Disabilities: A Review on Deep Learning Models and Future Directions 2024 19th International Joint Symposium on Artificial Intelligence and Natural Language Processing, iSAI-NLP 2024 10.1109/iSAI-NLP64410.2024.10799245 https://www.scopus.com/inward/record.uri?eid=2-s2.0-85216562911&doi=10.1109%2fiSAI-NLP64410.2024.10799245&partnerID=40&md5=94ad7c5e822a95e47c2061ced489ffc5 This study reviews deep learning models used in handwriting image classification for the automated diagnosis of learning disabilities. By addressing handwriting diversity and misclassification challenges, two models were highlighted: Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs). Literature was retrieved from major databases including IEEE Xplore, Scopus, Web of Science (WoS), and Google Scholar, with studies on Parkinson's disease, tremor patients, and machine learning excluded. CNNs represent a more mature architecture focusing on convolutions, pooling, and activation function. Meanwhile, ViTs emerges as a promising alternative via its multi-head attention architecture. This review also compares the accuracy of both models, specifying the sources of handwriting images, as well as providing future directions relevant to the research field. © 2024 IEEE. Institute of Electrical and Electronics Engineers Inc. English Conference paper |
author |
Sukiman S.A.; Husin N.A.; Hamdan H.; Murad M.A.A. |
spellingShingle |
Sukiman S.A.; Husin N.A.; Hamdan H.; Murad M.A.A. Handwriting Image Classification for Automated Diagnosis of Learning Disabilities: A Review on Deep Learning Models and Future Directions |
author_facet |
Sukiman S.A.; Husin N.A.; Hamdan H.; Murad M.A.A. |
author_sort |
Sukiman S.A.; Husin N.A.; Hamdan H.; Murad M.A.A. |
title |
Handwriting Image Classification for Automated Diagnosis of Learning Disabilities: A Review on Deep Learning Models and Future Directions |
title_short |
Handwriting Image Classification for Automated Diagnosis of Learning Disabilities: A Review on Deep Learning Models and Future Directions |
title_full |
Handwriting Image Classification for Automated Diagnosis of Learning Disabilities: A Review on Deep Learning Models and Future Directions |
title_fullStr |
Handwriting Image Classification for Automated Diagnosis of Learning Disabilities: A Review on Deep Learning Models and Future Directions |
title_full_unstemmed |
Handwriting Image Classification for Automated Diagnosis of Learning Disabilities: A Review on Deep Learning Models and Future Directions |
title_sort |
Handwriting Image Classification for Automated Diagnosis of Learning Disabilities: A Review on Deep Learning Models and Future Directions |
publishDate |
2024 |
container_title |
19th International Joint Symposium on Artificial Intelligence and Natural Language Processing, iSAI-NLP 2024 |
container_volume |
|
container_issue |
|
doi_str_mv |
10.1109/iSAI-NLP64410.2024.10799245 |
url |
https://www.scopus.com/inward/record.uri?eid=2-s2.0-85216562911&doi=10.1109%2fiSAI-NLP64410.2024.10799245&partnerID=40&md5=94ad7c5e822a95e47c2061ced489ffc5 |
description |
This study reviews deep learning models used in handwriting image classification for the automated diagnosis of learning disabilities. By addressing handwriting diversity and misclassification challenges, two models were highlighted: Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs). Literature was retrieved from major databases including IEEE Xplore, Scopus, Web of Science (WoS), and Google Scholar, with studies on Parkinson's disease, tremor patients, and machine learning excluded. CNNs represent a more mature architecture focusing on convolutions, pooling, and activation function. Meanwhile, ViTs emerges as a promising alternative via its multi-head attention architecture. This review also compares the accuracy of both models, specifying the sources of handwriting images, as well as providing future directions relevant to the research field. © 2024 IEEE. |
publisher |
Institute of Electrical and Electronics Engineers Inc. |
issn |
|
language |
English |
format |
Conference paper |
accesstype |
|
record_format |
scopus |
collection |
Scopus |
_version_ |
1825722578778456064 |