Recent advancements in Deep Learning-based Handwritten Text Recognition (HTR) have led to models with remarkable performance on both modern and historical manuscripts in large benchmark datasets. Nonetheless, those models struggle to obtain the same performance when applied to manuscripts with peculiar characteristics, such as language, paper support, ink, and author handwriting. This issue is very relevant for valuable but small collections of documents preserved in historical archives, for which obtaining sufficient annotated training data is costly or, in some cases, unfeasible. To overcome this challenge, a possible solution is to pretrain HTR models on large datasets and then fine-tune them on small single-author collections. In this paper, we take into account large, real benchmark datasets and synthetic ones obtained with a styled Handwritten Text Generation model. Through extensive experimental analysis, also considering the amount of fine-tuning lines, we give a quantitative indication of the most relevant characteristics of such data for obtaining an HTR model able to effectively transcribe manuscripts in small collections with as little as five real fine-tuning lines.

How to Choose Pretrained Handwriting Recognition Models for Single Writer Fine-Tuning / Pippi, V.; Cascianelli, S.; Kermorvant, C.; Cucchiara, R.. - 14188:(2023), pp. 330-347. (Intervento presentato al convegno 17th International Conference on Document Analysis and Recognition, ICDAR 2023 tenutosi a usa nel 2023) [10.1007/978-3-031-41679-8_19].

How to Choose Pretrained Handwriting Recognition Models for Single Writer Fine-Tuning

Pippi V.
;
Cascianelli S.;Cucchiara R.
2023

Abstract

Recent advancements in Deep Learning-based Handwritten Text Recognition (HTR) have led to models with remarkable performance on both modern and historical manuscripts in large benchmark datasets. Nonetheless, those models struggle to obtain the same performance when applied to manuscripts with peculiar characteristics, such as language, paper support, ink, and author handwriting. This issue is very relevant for valuable but small collections of documents preserved in historical archives, for which obtaining sufficient annotated training data is costly or, in some cases, unfeasible. To overcome this challenge, a possible solution is to pretrain HTR models on large datasets and then fine-tune them on small single-author collections. In this paper, we take into account large, real benchmark datasets and synthetic ones obtained with a styled Handwritten Text Generation model. Through extensive experimental analysis, also considering the amount of fine-tuning lines, we give a quantitative indication of the most relevant characteristics of such data for obtaining an HTR model able to effectively transcribe manuscripts in small collections with as little as five real fine-tuning lines.
2023
17th International Conference on Document Analysis and Recognition, ICDAR 2023
usa
2023
14188
330
347
Pippi, V.; Cascianelli, S.; Kermorvant, C.; Cucchiara, R.
How to Choose Pretrained Handwriting Recognition Models for Single Writer Fine-Tuning / Pippi, V.; Cascianelli, S.; Kermorvant, C.; Cucchiara, R.. - 14188:(2023), pp. 330-347. (Intervento presentato al convegno 17th International Conference on Document Analysis and Recognition, ICDAR 2023 tenutosi a usa nel 2023) [10.1007/978-3-031-41679-8_19].
File in questo prodotto:
Non ci sono file associati a questo prodotto.
Pubblicazioni consigliate

Licenza Creative Commons
I metadati presenti in IRIS UNIMORE sono rilasciati con licenza Creative Commons CC0 1.0 Universal, mentre i file delle pubblicazioni sono rilasciati con licenza Attribuzione 4.0 Internazionale (CC BY 4.0), salvo diversa indicazione.
In caso di violazione di copyright, contattare Supporto Iris

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11380/1363931
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 1
  • ???jsp.display-item.citation.isi??? ND
social impact