Enhancing Handwritten Text Recognition Performance with Encoder Transformer Models

Authors

  • Dipesh Bhattarai Texas College of Management and IT, Kathmandu, Nepal
  • Pawan Kumar Sharma Texas College of Management and IT, Kathmandu, Nepal

DOI:

https://doi.org/10.3126/joeis.v4i1.81606

Keywords:

Handwritten Text Recognition, Transformers, BERT, CNN, Deep Learning, OCR

Abstract

Handwritten Text Recognition (HTR) is a critical area in computer vision and natural language processing, aiming to convert handwritten content into machine-readable text. The task poses significant challenges due to the inherent variability in handwriting styles, stroke patterns, character spacing, and writing instruments. Traditional HTR techniques, often based on statistical models or shallow neural networks, frequently struggle to generalize across diverse handwriting samples, leading to suboptimal performance in real-world applications.

This improvement demonstrates a relative gain of approximately 10-15% over traditional RNN and LSTM-based models on the same dataset.

Downloads

Download data is not yet available.
Abstract
123
pdf
45

Author Biographies

Dipesh Bhattarai, Texas College of Management and IT, Kathmandu, Nepal

Texas College of Management and IT, Kathmandu, Nepal

Pawan Kumar Sharma, Texas College of Management and IT, Kathmandu, Nepal

Texas College of Management and IT, Kathmandu, Nepal

Downloads

Published

2025-07-21

How to Cite

Bhattarai, D., & Sharma, P. K. (2025). Enhancing Handwritten Text Recognition Performance with Encoder Transformer Models. Journal of Engineering Issues and Solutions, 4(1), 456–459. https://doi.org/10.3126/joeis.v4i1.81606

Issue

Section

Research Articles