Back to articles
Work Presented at Archiving 2025 FastTrack
Volume: 0 | Article ID: 020401
Image
A Building-Block Approach to Character-Level Writer Verification on the Great Isaiah Scrolls
Abstract
Abstract

This study presents a novel character-level writer verification framework for ancient manuscripts, employing a building-block approach that integrates decision strategies across multiple token levels, including characters, words, and sentences. The proposed system utilized edge-directional and hinge features along with machine learning techniques to verify the hands that wrote the Great Isaiah Scroll. A custom dataset containing over 12,000 samples of handwritten characters from the associated scribes was used for training and testing. The framework incorporated character-specific parameter tuning, resulting in 22 separate models and demonstrated that each character has distinct features that enhance system performance. Evaluation was conducted through soft voting, comparing probability scores across different token levels, and contrasting the results with majority voting. This approach provides a detailed method for multi-scribe verification, bridging computational and paleographic methods for historical manuscript studies.

Subject Areas :
Views 0
Downloads 0
 articleview.views 0
 articleview.downloads 0
  Cite this article 

T. Lumban Tobing, P. Bours, "A Building-Block Approach to Character-Level Writer Verification on the Great Isaiah Scrollsin Journal of Imaging Science and Technology,  2025,  pp 1 - 10,  https://doi.org/10.2352/J.ImagingSci.Technol.2025.69.2.020401

 Copy citation
  Copyright statement 
Copyright © Society for Imaging Science and Technology 2025
 Open access
  Article timeline 
  • received December 2024
  • accepted March 2025
jist
JIMTE6
Journal of Imaging Science and Technology
J. Imaging Sci. Technol.
J. Imaging Sci. Technol.
1062-3701
1943-3522
Society for Imaging Science and Technology
1.
Introduction
The period of ancient history extends approximately from the fifth millennium BCE to the fifth century CE. During this period, people transcribed mostly on thick paper made from plants and animal skins; those that survived named are referred to as ancient manuscripts. Considering the writing medium used and the thousands of years they have survived, physical quality degradation is inevitable and is a major obstacle for accurate document analysis. Digital image production of ancient manuscripts has been made possible through the vast and sophisticated technology of document imaging, aiming to preserve them at a specific point in time. Furthermore, advancements in digital imaging have enabled the restoration of degraded manuscripts, addressing challenges such as bleed-through removal [1, 2], alignment correction [3], and multispectral text extraction [4], making historical documents more accessible and amenable for computational analysis.
Paleography, a study of ancient writing systems is one of the auxiliary sciences of history. In general, paleography investigates the type of script and ductus (distinctive features of the strokes of particular hands) inscribed on the manuscript to determine whether additional manuscripts studied were written by the same person. Such investigations are needed to identify additional manuscripts written by the same person, as these manuscripts may provide additional content for deeper understanding of the contextual knowledge gleaned from the precedent manuscript [5]. Thus, paleography plays a central role in analyzing ancient manuscripts and verifying their authorship, and offers a foundational approach to the understanding of historical documents.
Simultaneously, the field of computer science has adapted techniques from image processing and pattern recognition to address similar problems through the tasks of writer identification and verification [6]. As summarized by Bensefia et al. [7], writer identification task deals with the retrieval of handwritten samples from a database depending on the graphical analysis of the handwritten samples under study, while writer verification task aims to determine whether two samples of documents were written by the same writer. Based on the description, writer identification and verification tasks are equivalent to the objectives that paleography aimed to achieve. This notion then forms the basis of research in computer-aided writer identification and verification based on digital images of various handwritten scripts in modern and historical documents.
This study integrates the knowledge from paleography with advancements in computer-aided writer recognition to verify the authorship of the Great Isaiah Scrolls, an ancient manuscript believed to have been written by two different scribes. The primary objective of this study is the proposal of a writer verification system that can verify the dual-scribe hypothesis using a novel building-block approach. This approach is designed to improve verification accuracy by iteratively analyzing character-specific data, which is crucial for the detailed and precise identification of the scribes. Additionally, this study critically evaluates the data and methods used in the writer verification system, examining the integration of paleographic expertise with modern computational techniques.
The rest of the article is organized as follows: Section 2 reviews research on paleography-based writer recognition and computer-aided writer recognition. Section 3 focuses on the description of the proposed framework and the methods used in this study. Section 4 describes the experiment setting and Section 5 provides the results and its interpretation. Section 6 summarizes and concludes the study.
2.
Related Work
To perform writer verification on historical manuscripts is to bridge a profound gap between computer science and historical studies (disciplines focused on the historical significance of documents and relics). This bridging process requires alignment of expertise from both fields, where each discipline must temper its assumptions and expectations to find common ground. Not all advanced techniques in computer science can be applied directly to historical manuscripts, nor can traditional historical methods fully address the computational challenges. Therefore, this section explores the background and contributions of both disciplines, highlighting gaps in existing research and the drivers of the proposed framework.
2.1
1QIsa-a Scrolls and Paleographic Approaches to Identifying Scribes
1QIsa-a, the Great Isaiah Scroll, discovered in 1947 in Cave 1 at Qumran, is one of the most important biblical manuscripts found in the Dead Sea Scrolls collection. It is notable for its length (734 cm) and preservation, containing the full ancient square script Hebrew text of the Book of Isaiah from ca. 125 BCE, making it one of the oldest surviving biblical texts (additional information and images can be accessed at http://dss.collections.imj.org.il/isaiah) [8]. The study of ancient manuscripts, such as the 1QIsa-a, combines historical research with various methodologies to uncover both the textual content and the scribal practices behind these texts. While the primary focus often lies in extracting historical facts from the content itself, the application of contextualization, sourcing, and corroboration is essential in the scholarly approach to understanding these manuscripts. As Van Drie et al. describe, corroboration involves comparing documents to address historical questions or confirm claims, with the identification of comparable documents—such as those written by the same scribe— providing deeper insights into their creation and context [9]. For ancient manuscripts, this requires identifying comparable documents, often determined by whether they were written by the same scribe. Documents attributed to the same scribe can provide additional comparative material for deeper contextual insights [5].
Paleography, the study of ancient handwriting, is a vital tool in this process. This methodology, as defined by Wakelin, involves analyzing features such as handwriting size, letterforms, and corrections to reveal individual scribal practices and styles [10]. Tov’s work emphasized that examining these characteristics in Hebrew manuscripts can help identify distinct scribes, allowing scholars to trace the evolution and transmission of texts [11]. The consensus surrounding the authorship of the 1QIsa-a Scroll is not uniform. Traditionally, paleographic methods suggested that the entire manuscript was copied by a single scribe, with subtle variations in handwriting attributed to personal idiosyncrasies or occasional inconsistencies. However, some scholars also proposed that the manuscript was actually the work of two distinct scribes—one responsible for Columns I–XXVII and another for Columns XXVIII–LIV. Recent advancements, particularly through AI techniques that detect micro-level handwriting variations, have revealed the involvement of at least two scribes in its creation [12]. This breakthrough demonstrates the power of combining traditional paleographic methods with modern technology to enhance our understanding of ancient texts, as well as the scribes who worked on them.
2.2
Computational Writer Verification
The study of writer verification has grown significantly in recent decades, driven by advances in machine learning and its applications to handwriting analysis. At least three key components differentiate computational approaches: the token level of textual input (e.g., character, word, sentence), the feature extraction methods, and the classification techniques (see Table I).
Table I.
Research on writer recognition for Latin and non-Latin scripts.
Dataset nameScriptInput levelFeature extraction methodClassifier
PSLWC
IAM, CVL,Dutch,Refs.Refs.Refs.Refs.LBP, LTP, LPQ,SVM,
FiremakerEnglish,[1317][18][19] and[20] and FAST, SIFT,Nearest
German [21][22] HKD, CNNNeighbor,
feature maps,CNN, RNN
SRS-LBP
ICFHR2012,Arabic,Refs.Refs.Refs.Refs.CNN featureSVM
IFN/ENIT,Chinese,[13, 19],[23][19] and[18, 24, 25]maps, SURFNearest
QUWIKannada,[2629] [23][30]SIFT, CANeighbor,
HIT-MW, IHP,Devanagari, LDCF, GITF,Distance
HWDB1.1,Japanese CLGP, HOG,Calculation,
JEITA-HP, GLRL, ZoningCNN
KHATT,
AHTID/MW,
Custom dataset
Across diverse studies, research on Latin and non-Latin scripts often employs similar feature extraction techniques and classifiers, demonstrating the adaptability of machine learning frameworks across scripts. However, the choice of input level significantly influences the system’s focus and accuracy. For example, character-level analysis can capture fine-grained handwriting traits, while word- or sentence-level inputs provide broader context. Despite these advances, integrating decision scores across token levels in a systematic manner remains a major challenge, as highlighted in Table I.
For historical manuscripts like the 1QIsa-a Scrolls, it is essential to select features and classifiers that align with paleographic practices. In this study, we employ edge-directional and hinge features, as these methods are intuitive to scholars studying historical manuscripts. These features have been successfully applied in handwriting analysis from their introduction to recent studies in capturing stroke directionality and curvature patterns (Refs. [31] and [32]). Additionally, we use an SVM classifier, which has demonstrated robust performance in tasks involving historical and modern handwriting (Refs. [30] and [33]).
2.3
Gaps in Existing Research
Despite significant advancements, computational writer verification faces critical limitations when applied to historical manuscripts:
Multi-Scribe Authorship: Most systems assume a single writer for a document, overlooking the possibility of collaborative manuscripts where multiple scribes contribute.
Unit-Level Integration: While character-level analysis provides precision, there is a lack of robust frameworks for aggregating information hierarchically across characters, words, and sentences.
Probabilistic Decision Frameworks: Few studies adopt probabilistic approaches to integrate evidence across granularities, which is essential for handling complex documents and transitions between scribes.
These gaps are particularly relevant for the 1QIsa-a Scrolls, where evidence suggests multi-scribe authorship. The inability to address such challenges limits the scope and accuracy of existing verification methods. Furthermore, due to the varying consensus on the authorship of the 1QIsa-a Scrolls, no standardized protocol exists for verifying proposed authorship claims through a writer verification system. Current systems lack the capability to assess and validate the theories surrounding the manuscript’s scribal origins.
2.4
Motivation for a Building-Block Approach
This study introduces a building-block approach to address the gaps identified. At the character level, this framework-based approach assigns probability scores to individual characters based on their likelihood of having been written by the same scribe. These scores are then aggregated hierarchically at word and sentence levels, allowing for a comprehensive verification process. This modular approach is well-suited for manuscripts like the 1QIsa-a Scrolls, where the complexity of multi-scribe authorship requires a flexible and scalable framework. By integrating paleographic expertise with computational methods, the proposed system offers an intuitive yet powerful solution for historical manuscript studies. Furthermore, this framework provides scholars with a deeper understanding of the methods employed, bridging the gap between disciplines and enhancing collaboration.
3.
Proposed Method
This section describes the proposed character-level writer verification data flow and additional methods used as validation baseline in the experiments.
3.1
Overall Architecture
The writer recognition framework mentioned in Section 2.2 mainly adopts an absolute decision strategy in identifying, verifying, or authenticating a specific token. For example, in a page-level writer recognition system, the recognition rate is usually computed using global features, and the decision process stops at this level. This method typically assesses whether the entire page can be attributed to a particular writer, relying on features that represent the overall style of the text. However, in the context of historical manuscripts, a more granular analysis is often required. In such cases, it is important to consider the recognition rate at smaller units, such as characters, words, or sentences. This is because, in historical manuscript studies, the goal is not only to determine whether an entire page was written by the same hand but also to investigate whether different hands may have written paragraphs within the same text, sentences within a paragraph, or even individual words within a sentence.
Unlike modern datasets, where the identity of the writer is typically labeled (often in controlled lab environments), historical manuscripts lack such clear labeling and may involve multiple hands contributing to the same document. In these cases, a detailed analysis at finer levels is necessary to assess the possibility of multiple authors contributing to a single page, paragraph, or even a specific word. This approach provides a more nuanced understanding of the document’s authorship, which is especially important when studying manuscripts where the attribution of authorship is uncertain or disputed. Thus, in this study, we integrated the building-block approach in the decision strategy stage with a machine learning- based writer verification system, as illustrated in Figure 1.
Figure 1.
Character-Level Writer Verification Framework: from machine learning processes (data collection, feature extraction, model training, and testing) to the construction of a hierarchical decision strategy for writer verification.
The building-block approach uses a layered decision-making strategy built on the outcomes of a machine learning model. At the lowest level, raw probability scores result from testing and evaluation within the machine learning framework. These scores serve as foundational inputs, which are further used to compute the intermediate-level and highest-level scores, allowing for a hierarchical assessment of writer verification. This decision strategy consists of three hierarchical layers. At the highest level (Layer 2), S2 represents the sentence- level probability score, indicating the likelihood that the entire sentence is written by a specific scribe. The intermediate level (Layer 1), where S1, i represents the word-level probability score for the ith word in the sentence, indicating the likelihood that the word is associated with a specific writer. At the lowest level (Layer 0), p0, i, j represents the raw probability scores of each individual character in word i, with j indexing the character within the word. This layered structure permits a detailed analysis, ranging from characters to words and sentences, facilitating writer verification at multiple levels.
3.2
Validation Baseline
To implement the proposed framework, we built an adaptation system to investigate the hands that wrote an ancient manuscript written in square script Hebrew, known as the Great Isaiah Scrolls. Our baseline used edge-directional and hinge feature extraction methods proposed by Bulacu et al. [30]. Edge-directional and hinge methods extract angular information differently—Edge-directional features compute the angle between handwriting strokes and a horizontal reference line, while hinge features measure angles formed between pairs of adjacent strokes. In both methods, these angles are categorized into structured bins based on their frequency of occurrence, with each bin assigned an empirical probability. These binned values represent the feature vectors, which are then used as input for training a machine learning model. Training-wise, we used an RBF kernel Support Vector Machine (SVM) to train and predict data [33]. SVM is known as a robust classifier that is suitable for datasets with high-dimensional and non-linearly separable features, which aligns with the characteristics of the data used in this study. To assess the contribution of the building-block approach, we leveraged SVM’s probabilistic outputs by opting for class probability scores as the output of the testing stage. These scores were then processed for soft voting and majority voting systems.
3.2.1
Decision Strategy
The machine learning-based verification system generates class probability scores for each token at the lowest level. At the character level (Layer 0), the model produces p0, i, j representing the probability that this character j in word i belongs to the reference scribe. This probability is constrained within: p0, i, j ∈ [0,1]. Since this is a binary classification task (two scribes), the probability of the second scribe is implicitly given by: 1 − p0, i, j.
To compute higher-level scores, soft voting and majority voting were applied separately at the word level (Layer 1) and the sentence level (Layer 2). At the word level (Layer 1), the aggregated score is defined as: S1, k = {S1, ksoft, S1, kmajority}, where k is the word index, S1, ksoft is the soft voting score for word k, and S1, kmajority is the majority voting score for word k. At the sentence level (Layer 2), the final aggregated score is expressed as: S2 = {S2soft, S2majority}, where S2soft is the sentence-level soft voting score and S2majority is the sentence-level majority voting score. These scores are further explained in the following section.
(a) Soft Voting. The soft voting score for a word, denoted as S1, isoft, is computed as the average of the probability scores of its characters. Assuming a word i consists of n characters, each with a probability score p0, i, j for j = 1,2, …, n, the soft voting score is expressed as:
(1)
S1,isoft=1nj=1np0,i,j,
where p0, i, j is the probability score of the jth character in word i, and n is the total number of characters in the word.
For example, consider word 1 (i = 1) consisting of four characters with probability scores: p0,1,1 = 0.5, p0,1,2 = 0.55, p0,1,3 = 0.5, and p0,1,4 = 0.6. The soft voting score for this word is:
(2)
S1,1soft=14(0.50+0.55+0.50+0.60)=0.5375.
At the sentence level (Layer 2), the soft voting score is computed by averaging the soft voting scores of the words in the sentence:
(3)
S2soft=1mi=1mS1,isoft,
where m is the number of words in the sentence.
(b) Majority Voting. In majority voting, the decision score for a word, denoted as S1, imajority, is computed by counting the number of characters with a probability score greater than 0.5. The indicator function I (p0, i, j > 0.5) is used, returning 1 if p0, i, j > 0.5 or 0 otherwise.
The majority voting score is expressed as:
(4)
S1,imajority=1nj=1n(p0,i,j>0.5),
where n is the number of characters in the word.
For example, consider word 1 (i = 1) consisting of four characters with probability scores: p0,1,1 = 0.5, p0,1,2 = 0.55, p0,1,3 = 0.50, and p0,1,4 = 0.60. Applying the indicator function, the majority voting score for this word is:
(5)
S1,1majority=14 ((0.50>0.5)+(0.55>0.5)+(0.50>0.5)+(0.60>0.5))=14(0+1+0+1)=0.50.
At the sentence level (Layer 2), the majority voting score is computed in a similar way by averaging the decision scores obtained from each word in the sentence:
(6)
S2majority=1mi=1mS1,imajority,
where m is the number of words in the sentence.
4.
Experiment Setting
This section explains the rationale behind the selection of data, methods, and approaches of this study and its description.
4.1
Data Collection
The Great Isaiah scroll consists of 54 pages, commonly called columns, of which it was assumed (by humanities [20]) that Scribe A wrote Col. I–XXVII and Scribe B Col. XXVIII–LIV. Research done by Popovic et al. [29] proved that assumption by implementing unsupervised learning for all columns. Based on these findings, the adaptation system aims to build a dataset and verify the data based on the claimed columns belonging to each scribe. We present an illustrated version of the letter samples, traced from the original images (see Figure 2). To consider all tokens that are available in large quantities, we chose tokens to be investigated from the lowest level, i.e., characters, words, and one sentence. We constructed the character-specific model based on the non-final form of the 22 letters of ancient square script Hebrew. We excluded the five final form letters since their frequency of occurrence was so limited that the machine learning-based system was unlikely to benefit from the small dataset.
Figure 2.
Illustration of the 22 isolated letters with their corresponding names.
For data collection, we expanded our custom dataset using the same approach described in our earlier study [34]. A total of 283 samples representing each letter (except for the letter tet, with only 145 samples, and the letter samekh with 232 samples, due to limited availability) were taken from each scribe’s corresponding columns. The training-validation set referring to Scribe A comprised single characters extracted from Col. I–XXVI, while the set referring to Scribe B included single characters extracted from Col. XXVIII–LIII. The test set consisted of one sentence from Col. XXVII (referring to Scribe A’s handwriting) and one from Col. LIV (referring to Scribe B’s handwriting).
This separation of data into Scribe A and Scribe B’s columns follows the presumed authorship as well as AI-based writer identification methods discussed in Section 2. The goal of this work was to verify these assumptions using the proposed system, leveraging the computational approach to confirm the authorship of the individual columns attributed to each scribe.
4.2
Validation Baseline
To validate the proposed framework, we utilized edge-directional and hinge feature extraction methods and an SVM-based classification technique. A major advantage of implementing handcrafted features, specifically edge-directional and hinge features, is that unique information based on the slant angle and curvatures of handwriting projection can be extracted. The collaborative nature of SVM and handcrafted features also brings an advantage for robust implementation, especially with our relatively small dataset, which could perform poorly during model training with a deep learning network.
4.2.1
Edge-directional and Hinge Feature Extraction
In this study, four pixel-length variations were applied: 2 px, 3 px, 4 px, and 5 px, resulting in different numbers of feature elements, i.e., 9, 13, 17, and 21 elements, respectively. In addition, unlike the Sobel edge detection method used in the original article, we implement the skeleton transformation method. This transformation aims to reduce the pixel width of binary objects to a 1-pixel-wide representation. Our earlier study [35] demonstrated that using edge-directional features for writer verification of the Great Isaiah Scrolls yields better and more consistent accuracy scores when combined with skeleton transformation. Based on this study, each individual character image must be skeletonized before feature extraction. Once the skeleton transformation was applied to each character, these instances were then fed into the feature extraction stage.
For hinge feature extraction, we employed a similar skeletonization process as a prerequisite. Hinge features capture the angular relationships between pairs of edge directions, providing a detailed representation of how strokes curve and connect within a character. For every pixel on the skeletonized character, pairs of edge directions within a predefined radius were identified. The computed angles were grouped into predefined bins, creating a histogram that represented the distribution of angular relationships for the given character. In this study, four pixel-length variations were applied: 2 px, 3 px, 4 px, and 5 px, which resulted in a different number of feature elements, i.e., 104, 252, 464, and 740 feature elements, respectively.
4.2.2
SVM Model Training
For classification optimization, we used 81 combinations of parameter C and γ with C = 2−3, 2, 23, 25, 27, 29, 211, 213, 215, and γ = 2−15, 2−13, 2−11, 2−9, 2−7, 2−5, 2−3, 2,23. We used the exponentially growing sequences of the parameters as recommended in a practical guide [36] to SVM classification. In the model training stage, we distributed data for training and validation with an 80:20 ratio. The feature extraction from each character with four different pixel-length options was accomplished. Assuming that we choose features that belong to the letter alef and were extracted using the 2 px pixel-length setting, we then need to split the training and validation data to ensure an equal representation of samples from each scribe. This is illustrated by the following explanation (note: this does not apply to the letters tet and samekh). Since the letter alef has a set of 283 samples belonging to Scribe A and another set from Scribe B, the 112 validation samples (20% of total samples) should consist of 56 samples from Scribe A and 56 samples from Scribe B. For training, each scribe should be represented by 227 training samples. To ensure an equally distributed result, we employed a stratified cross-validation method. Furthermore, we implemented 5-fold cross- validation to create five different sets of training and validation data to avert the proneness of underfitting or overfitting. Next, we recorded the average of the training and validation accuracy scores obtained from 5-fold cross-validation. Finally, we obtained 22 combinations of parameters and a pixel-length type specific to each letter with the best training accuracy scores with low training-validation scores difference, to avoid overfitting. These combinations were used to test the new data derived from the testing set of Col. XXVII and Col.LIV.
5.
Result and discussion
5.1
Character-Specific Model
Table II presents the best results of parameter tuning obtained from the training and validation section of the proposed writer verification system. To determine which letters are more representative or less representative of their respective features (edge- directional and hinge), we can analyze the training, validation, and gap scores provided for each letter. Representative letters typically have higher scores (indicating better alignment with the feature set), while less representative letters present with lower scores, suggesting that the features struggle to capture the stylistic traits of those letters. Letters like dalet, tet, and qof are the most representative, as evidenced by their consistently high scores in both training and validation datasets. Their distinct stylistic traits make it easier to classify them with edge-directional and hinge features. Letters like zayin, vav, and samekh are less representative due to lower training and validation scores. This indicates that these letters either lack strong distinguishing features or that the current feature extraction methods struggle to model their traits effectively.
Table II.
Best average scores of training and validation of the 22 trained character-specific models.
Edge-directionalHinge
LetterμTraining (%)μValidation (%)μTraining (%)μValidation (%)
alef70.458.281.863.0
ayin68.058.081.160.3
bet67.056.979.363.7
dalet72.461.181.764.7
gimel70.758.983.061.0
he69.858.081.158.7
het65.260.776.762.4
kaf64.153.080.458.2
lamed65.352.874.857.2
mem70.759.479.763.2
nun66.057.580.659.3
pe63.656.479.760.2
qof72.160.583.363.5
resh69.961.180.165.3
samekh63.957.271.559.6
shin64.857.381.760.2
tav66.262.475.066.9
tet70.163.582.262.3
tsadi62.756.278.358.5
vav67.358.471.460.2
yod66.459.179.458.8
zayin61.958.972.761.6
To assess the effectiveness of the hinge and edge-directional features, we analyzed the mean differences in training scores, validation scores, and the gaps between these scores. Figure 3 intuitively depicts the trends in training and validation scores for each feature extraction method, complementing the results summarized in Table III.
Figure 3.
Best average scores of training and validation of the 22 trained character-specific models.
Table III.
Mean scores for training, validation, and gaps.
FeatureTraining mean (%)Validation mean (%)Gap mean (%)
Hinge features78.961.317.6
Edge-directional features67.258.48.8
Mean difference11.72.98.8
The mean training score for hinge method was 78.9%, while edge-directional achieved a mean score of 67.2%, resulting in a mean difference of 11.7%. This indicates that hinge consistently outperformed edge-directional in capturing patterns within the training data. The higher training scores for hinge suggest that it is more effective at modeling the stylistic differences in the handwriting captured during training, which is crucial for distinguishing between writers. For validation scores, hinge achieved a mean of 61.3%, whereas edge- directional attained a mean score of 58.4%, resulting in a smaller mean difference of 2.9%. While hinge maintains an advantage on unseen data, the reduced gap between the validation scores suggests that edge-directional performs closer to hinge when generalizing to new data. This indicates that edge-directional might generalize more consistently but is overall less accurate. The gap between training and validation scores for hinge was 17.6%, while edge- directional exhibited a smaller gap of 8.8%, with a mean difference of 8.8%. The larger gap for hinge highlights a potential overfitting issue, as its performance on training data was significantly higher than on validation data. In contrast, edge-directional had a smaller gap, suggesting better generalization to new data despite its lower overall accuracy. While hinge demonstrated higher accuracy on both training and validation data, its larger gap between training and validation scores suggests it may be more prone to overfitting. Edge-directional, with its smaller gap, appears to generalize better but sacrifices some level of accuracy. Based on these results, hinge is recommended if achieving the highest possible accuracy is the primary goal, and overfitting can be mitigated through regularization techniques or additional data augmentation. However, if generalization is more critical, edge-directional may be a more robust choice.
5.2
Decision Strategy Results: Scribe A versus Scribe B
Following the interpretation of character-specific training and validation scores, it is imperative to analyze the results of the decision strategy on the building-block approach for determining the probabilities at word and sentence levels. This approach provides a practical perspective on how the trained features translate into the testing phase and helps identify characteristics associated with each scribe’s writing. The results are detailed in Table IV (showing probabilities for Scribe A, Col. XXVII) and Table V (showing probabilities for Scribe B, Col. LIV).
Table IV.
The probability scores at word- and sentence-levels, a unit of analysis perceived to be written by Scribe A (Col. XXVII).
Word #Edge-directionalHinge
S1, isoft (%)S1, imajority (%)S1, isoft (%)S1, imajority (%)
151.575.054.650.0
233.225.056.050.0
351.457.155.042.9
449.740.049.960.0
566.480.065.860.0
652.450.046.950.0
760.750.029.50
845.833.351.633.3
959.360.049.620.0
1055.975.064.475.0
1135.825.055.250.0
1273.610055.825.0
1371.810065.350.0
Sentence #S2soft (%)S2majority (%)S2soft (%)S2majority (%)
154.453.853.823.1
Table V.
The probability scores on word- and sentence-level, a unit of analysis perceived to be written by Scribe B (Col. LIV).
Edge-directionalHinge
Word #S1, isoft (%)S1, imajority (%)S1, isoft (%)S1, imajority (%)
149.366.762.666.7
251.442.934.114.3
351.140.037.930.0
453.750.053.850.0
539.0033.133.3
660.762.556.862.5
740.1040.766.7
840.125.064.2100
955.657.154,342.9
Sentence #S2soft (%)S2majority (%)S2soft(%)S2majority(%)
149.033.348.644.4
As shown in Table IV, for Scribe A, edge-directional outperformed hinge at sentence- level probabilities under both soft voting and majority voting strategies. Under soft voting, edge-directional achieved a sentence-level probability of 54.4%, which is slightly higher than hinge’s 53.8%. Word-level probabilities for edge-directional under this strategy ranged from 33.2% to 73.6%, indicating a slightly wider spread but with consistently fewer extremely low values compared to hinge. Under majority voting, edge-directional also demonstrated stronger performance with a sentence-level probability of 53.8%, significantly higher than hinge’s 23.1%. This marked difference indicates that edge-directional, despite variability in word-level probabilities, aggregates better in majority voting.
It can be inferred from Table V that Scribe B presented a different trend. Under soft voting, edge-directional again slightly outperformed hinge, with a sentence-level probability of 49.0% compared to 48.6%. However, the margin of difference was narrower for Scribe B than for Scribe A. Word-level probabilities for edge- directional ranged from 39.0% to 60.7%, showing less spread compared to Scribe A’s data and a more consistent performance. For majority voting, hinge surprisingly outperformed edge- directional, with a sentence-level probability of 44.4% compared to 33.3%. This indicates that hinge may be more robust when handling shorter sentence structures, as Scribe B’s dataset consists of only 9 words per sentence compared to Scribe A’s 13 words.
Analyzing the results further, several key observations emerge when comparing the impact of sentence length and voting strategies on edge-directional and hinge features:
Sentence Length Impact: Scribe B’s shorter sentences (9 words) amplify the impact of low word-level probabilities on the aggregated sentence-level result. This effect is particularly evident in majority voting, where the edge-directional feature’s performance drops significantly for Scribe B (33.3%) compared to Scribe A (53.8%).
Edge-directional’s Stability: Edge-directional consistently outperformed hinge under soft voting for both scribes, indicating that edge-directional is more stable when aggregating probabilities using this strategy. However, its performance under majority voting varies more, particularly for shorter sentences (Scribe B).
Hinge’s Robustness in Majority Voting: While hinge performance was lower than edge-directional, it showed better sentence-level probabilities for Scribe B under majority voting. This suggests that hinge may handle extreme variability at the word level better than edge- directional in certain conditions.
The results emphasize the importance of selecting an appropriate voting strategy and feature representation. Soft voting appears more stable across scribes and features, making it a preferred approach when building sentence-level probabilities. Edge-directional emerges as the stronger candidate overall due to its higher sentence-level probabilities and more consistent word-level predictions under soft voting. However, the lower performance of both features for Scribe B highlights challenges in handling shorter sentences. Shorter units of analysis amplify inconsistencies in word-level predictions, especially under majority voting, where extreme values (e.g., 0%) can disproportionately affect the aggregated result.
5.3
Evaluation of Probability Scores
The relatively low probability scores observed in the analysis can be attributed to several underlying factors related to the data, feature extraction, and the nature of the scribes’ handwriting. Two key possibilities contributing to these results are discussed below.
Limitations in Feature Representation and Overlapping Handwriting Styles A significant factor influencing the low probability scores is the limited ability of the features— specifically, edge-directional and hinge—to capture the distinct differences between Scribe A and Scribe B. Both scribes may share overlapping stylistic traits, which make it difficult for the system to differentiate their handwriting accurately. Edge-directional and hinge features, which are used to represent characteristics such as stroke angles and curvatures, may not be sensitive enough to the subtle differences in each scribe’s unique writing style. When the handwriting styles between two scribes exhibit significant overlap, such as similar slant angles and letter shapes, these features may fail to highlight the necessary distinctions, resulting in lower confidence in the classification and relatively low probability scores.
Human-Generated Assumptions and Predefined Labels The system’s training process is based on human-generated assumptions and predefined labels from AI-driven research mentioned in Section 2.1, particularly involving unsupervised writer identification of ancient texts such as the 1QIsa-a scrolls. In this study, data augmentation techniques were used to artificially expand the dataset and increase the robustness of the model. However, the assumptions regarding the scribes’ styles and the division of the columns (such as the plane separation between Columns 1-54) might not fully capture the actual complexity of the scribes’ handwriting. These assumptions—while valuable for training purposes—may not fully account for the nuanced and dynamic nature of handwriting. The predefined labels that were assigned to the scribes based on these assumptions may not be entirely accurate, leading to mismatches in the system’s classification and, ultimately, to low or inconsistent probability scores.
6.
Conclusion
While notable advancements have been made in computational writer verification, significant challenges persist, particularly in the analysis of multi-scribe manuscripts such as the 1QIsa-a Scroll. The building-block approach presented in this study, which aggregates character-level probability scores to derive word- and sentence-level outcomes, offers a versatile and scalable solution to these challenges. By integrating computational techniques with paleographic analysis, this method provides a promising tool for historical manuscript studies and encourages interdisciplinary collaboration.
Future research should focus on refining feature extraction techniques, particularly enhancing edge-directional and hinge features to better capture the distinguishing characteristics of handwriting. Further evaluation using diverse writer identification datasets may contribute to refining and optimizing these features, ensuring robust performance across different script styles and manuscript conditions.
References
1JohnstonR. H.EastonR.KnoxK.EschbachR.TusinskiJ.ZapanM. C.1995Digital image restoration technology as applied to ancient degraded textual material using color imaging systemsColor Imaging Conf.3191194191–410.2352/CIC.1995.3.1.art00050
2DuboisE.DanoP.2005Joint compression and restoration of documents with bleed-throughArchiving2170174170–410.2352/issn.2168-3204.2005.2.1.art00037
3WangJ.TanC. L.2011Non-rigid registration and restoration of double-sided historical manuscripts2011 Int’l. Conf. on Document Analysis and Recognition137413781374–8IEEEBeijing, China10.1109/ICDAR.2011.276
4HedjamR.CherietM.2011Novel data representation for text extraction from multispectral historical document images2011 Int’l. Conf. on Document Analysis and Recognition172176172–6IEEEBeijing, China10.1109/ICDAR.2011.43
5LongacreD.A Contextualized Approach to the Hebrew Dead Sea Scrolls Containing Exodus,” Ph.D. thesis (University of Birmingham, Birmingham, UK, 2015)
6KlementV.SimonJ. C.HaralickR. M.1981Forensic writer recognitionDigital Image Processing519524519–24Springer NetherlandsDordrecht
7BensefiaA.PaquetT.HeutteL.2005A writer identification and verification systemPattern Recognit. Lett.26208020922080–9210.1016/j.patrec.2005.03.024
8Israel Museum“The great isaiah scroll,” (2024), accessed: March 29, 2025
9van DrieJ.van BoxtelC.2008Historical reasoning: towards a framework for analyzing students’ reasoning about the pastEduc. Psychol. Rev.208711087–11010.1007/s10648-007-9056-1
10WakelinD.2017PaleographyThe Encyclopedia of Medieval Literature in Britain161–6John Wiley Sons, LtdChichester, UK
11TovE.Scribal Practices and Approaches Reflected in the Texts Found in the Judean Desert2018BrillLeiden, The Netherlands
12PopovićM.DhaliM. A.SchomakerL.2021Artificial intelligence based writer identification generates new evidence for the unknown scribes of the dead sea scrolls exemplified by the great isaiah scroll (1QIsaa)PLoS One161281–28
13KumarP.SharmaA.2020Segmentation-free writer identification based on convolutional neural networkComput. Electr. Eng.85106707
14SrivastavaA.ChandaS.PalU.WallravenC.LiuQ.NagaharaH.2022Exploiting multi-scale fusion, spatial attention and patch interaction techniques for text-independent writer identificationPattern Recognit.203217203–17Springer International PublishingCham10.1007/978-3-031-02444-3_15
15NicolaouA.BagdanovA. D.LiwickiM.KaratzasD.2015Sparse radial sampling LBP for writer identification2015 13th Int’l. Conf. on Document Analysis and Recognition (ICDAR)716720716–20IEEETunis, Tunisia10.1109/ICDAR.2015.7333855
16BennourA.DjeddiC.GattalA.SiddiqiI.MekhazniaT.2019Handwriting based writer recognition using implicit shape codebookForens. Sci. Intl.3019110091–10010.1016/j.forsciint.2019.05.014
17HeS.SchomakerL.2021Gr-rnn: global-context residual recurrent neural networks for writer identificationPattern Recognit.11710797510.1016/j.patcog.2021.107975
18XingL.QiaoY.2016Deepwriter: a multi-stream deep CNN for text-independent writer identification2016 15th Int’l. Conf. on Frontiers in Handwriting Recognition (ICFHR)584589584–9IEEEShenzhen, China10.1109/ICFHR.2016.0112
19ChahiA.El merabetY.RuichekY.TouahniR.2020Cross multi-scale locally encoded gradient patterns for off-line text-independent writer identificationEng. Appl. Artif. Intell.8910345910.1016/j.engappai.2019.103459
20ShengH.SchomakerL.2020Fragnet: writer identification using deep fragment networksIEEE Trans. Inform. Forens. Security15301330223013–2210.1109/TIFS.2020.2981236
21ChenS.WangY.LinC.-T.DingW.CaoZ.2019Semi-supervised feature learning for improving writer identificationInform. Sci.482156170156–7010.1016/j.ins.2019.01.024
22KumarV.SundaramS.2024Siamese-based offline word level writer identification in a reduced subspaceEng. Appl. Artif. Intell.13010772010.1016/j.engappai.2023.107720
23HannadY.SiddiqiI.DjeddiC.El-KettaniM. E.-Y.2019Improving arabic writer identification using score-level fusion of textural descriptorsIET Biometr.8221229221–910.1049/iet-bmt.2018.5009
24NasunoR.AraiS.2017Writer identification for offline Japanese handwritten character using convolutional neural networkThe 5th IIAE Int’l. Conf. on Intelligent Systems and Image Processing 2017 (ICISIP2017)949794–7Institute of Industrial Applications Engineers (IIAE)Hawaii, USA
25NguyenH. T.NguyenC. T.InoT.IndurkhyaB.NakagawaM.2019Text-independent writer identification using convolutional neural networkPattern Recognit. Lett.121104112104–1210.1016/j.patrec.2018.07.022
26DurouA.Al-MaadeedS.ArefI.BouridaneA.ElbendakM.2019A comparative study of machine learning approaches for handwriter identification2019 IEEE 12th Int’l. Conf. on Global Security, Safety and Sustainability (ICGS3)206212206–12IEEELondon, UK10.1109/ICGS3.2019.8688032
27KumarP.SharmaA.2019DCWI: distribution descriptive curve and cellular automata based writer identificationExpert Syst. Appl.128187200187–20010.1016/j.eswa.2019.03.037
28RehmanA.NazS.RazzakM. I.HameedI. A.2019Automatic visual features for writer identification: a deep learning approachIEEE Access7171491715717149–5710.1109/ACCESS.2018.2890810
29DurouA.ArefI.EratebS.El-mihoubT.GhalutT.EmhemmedA.2022Offline writer identification using deep convolution neural network2022 IEEE 2nd Int’l. Maghreb Meeting of the Conf. on Sciences and Techniques of Automatic Control and Computer Engineering (MI-STA)434743–7IEEESabratha, Libya10.1109/MI-STA54861.2022.9837764
30DarganS.KumarM.GargA.ThakurK.2020Writer identification system for pre-segmented offline handwritten devanagari characters using k-nn and svmSoft Comput.24101111012210111–2210.1007/s00500-019-04525-y
31BulacuM.SchomakerL.PetkovN.WestenbergM. A.2003Writer style from oriented edge fragmentsComputer Analysis of Images and Patterns460469460–9SpringerBerlin, Heidelberg10.1007/978-3-540-45179-2_57
32DiamantatosP.KavallieratouE.GritzalisS.2022Directional hinge features for writer identification: the importance of the skeleton and the effects of character size and pixel intensitySN Comput. Sci.31181–1810.1007/s42979-021-00950-9
33BreretonR. G.LloydG. R.2010Support vector machines for classification and regressionAnalyst135230267230–6710.1039/B918972F
34TobingT.Yildirim YayilganS.GeorgeS.ElgvinT.2022Isolated handwritten character recognition of ancient hebrew manuscriptsArchiv. Conf.19353935–910.2352/issn.2168-3204.2022.19.1.8
35TobingT.ŠkrabánekP.Yildirim YayilganS.GeorgeS.ElgvinT.2023Character-based writer verification of ancient hebrew square-script manuscripts: on edge-direction featureArchiv. Conf.20155158155–810.2352/issn.2168-3204.2023.20.1.32
36HsuC.-w.ChangC.-c.LinC.-J.“A practical guide to support vector classification,” (2003). https://www.csie.ntu.edu.tw/ cjlin/papers/guide/guide.pdf. Accessed: March 29, 2025