Enhancing Ottoman word recognition via self-supervised pretraining using a siamese swin transformer

Warning The system is temporarily closed to updates for reporting purpose.

Qara, Sadiq and Sharifi, Kourosh and Kuru, Mehmet and Toprak, Sultan and Aptoula, Erchan (2025) Enhancing Ottoman word recognition via self-supervised pretraining using a siamese swin transformer. In: 33rd Signal Processing and Communications Applications Conference (SIU), Istanbul, Turkiye

Full text not available from this repository. (Request a copy)

Abstract

The Ottoman archive contains millions of handwritten documents, and their efficient management and retrieval is of paramount significance to historians. In this paper, a semi-supervised method is presented for Ottoman word recognition from handwritten historical documents. The approach involves a two-stage process: self-supervised pretraining followed by super-vised training with triplet loss. A Siamese network architecture with a Shifted Window Transformer backbone is used to learn robust feature representations. It is shown that pretraining significantly enhances the model's ability to discriminate between similar Ottoman words. The method is evaluated on a historical dataset of labeled Ottoman word crops. To foster reproducibility and future research, we release the full code and dataset at https://github.com/Sadiq04/Ottoman-KWS-IEEE-SIU
Item Type: Papers in Conference Proceedings
Uncontrolled Keywords: Shifted Window Transformer; Siamese Network; Visual Keyword Spotting
Divisions: Faculty of Engineering and Natural Sciences
Depositing User: Erchan Aptoula
Date Deposited: 26 Sep 2025 14:52
Last Modified: 26 Sep 2025 14:52
URI: https://research.sabanciuniv.edu/id/eprint/52564

Actions (login required)

View Item
View Item