Next Article in Journal
Transformative Effect of COVID-19 Pandemic on Magnetic Resonance Imaging Services in One Tertiary Cardiovascular Center
Next Article in Special Issue
A Computational Approach to Hand Pose Recognition in Early Modern Paintings
Previous Article in Journal
Manipulating Pixels in Computer Graphics by Converting Raster Elements to Vector Shapes as a Function of Hue
 
 
Correction published on 26 February 2024, see J. Imaging 2024, 10(3), 57.
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Siamese Transformer Network for Zero-Shot Ancient Coin Classification

1
School of Computer Science, University of St Andrews, Scotland KY16 9AJ, UK
2
Max Planck Institute for the History of Science, Boltzmannstraße 22, 14195 Berlin, Germany
*
Authors to whom correspondence should be addressed.
J. Imaging 2023, 9(6), 107; https://doi.org/10.3390/jimaging9060107
Submission received: 2 May 2023 / Revised: 10 May 2023 / Accepted: 19 May 2023 / Published: 25 May 2023 / Corrected: 26 February 2024
(This article belongs to the Special Issue Pattern Recognition Systems for Cultural Heritage)

Abstract

:
Ancient numismatics, the study of ancient coins, has in recent years become an attractive domain for the application of computer vision and machine learning. Though rich in research problems, the predominant focus in this area to date has been on the task of attributing a coin from an image, that is of identifying its issue. This may be considered the cardinal problem in the field and it continues to challenge automatic methods. In the present paper, we address a number of limitations of previous work. Firstly, the existing methods approach the problem as a classification task. As such, they are unable to deal with classes with no or few exemplars (which would be most, given over 50,000 issues of Roman Imperial coins alone), and require retraining when exemplars of a new class become available. Hence, rather than seeking to learn a representation that distinguishes a particular class from all the others, herein we seek a representation that is overall best at distinguishing classes from one another, thus relinquishing the demand for exemplars of any specific class. This leads to our adoption of the paradigm of pairwise coin matching by issue, rather than the usual classification paradigm, and the specific solution we propose in the form of a Siamese neural network. Furthermore, while adopting deep learning, motivated by its successes in the field and its unchallenged superiority over classical computer vision approaches, we also seek to leverage the advantages that transformers have over the previously employed convolutional neural networks, and in particular their non-local attention mechanisms, which ought to be particularly useful in ancient coin analysis by associating semantically but not visually related distal elements of a coin’s design. Evaluated on a large data corpus of 14,820 images and 7605 issues, using transfer learning and only a small training set of 542 images of 24 issues, our Double Siamese ViT model is shown to surpass the state of the art by a large margin, achieving an overall accuracy of 81%. Moreover, our further investigation of the results shows that the majority of the method’s errors are unrelated to the intrinsic aspects of the algorithm itself, but are rather a consequence of unclean data, which is a problem that can be easily addressed in practice by simple pre-processing and quality checking.

1. Introduction

Among the many application domains in which the rapidly advancing fields of computer vision and machine learning have found their use is that of numismatics and ancient numismatics in particular. The term “numismatics” refers both to the academic study of coins, paper currency and tokens, as well as the hobby of collecting these items. Ancient numismatics concerns ancient coins in particular, that is, the coins of Ancient Greece, Rome, Celtic tribes, etc.
Considering the inherently interdisciplinary focus of the present article, for the sake of clarity it is useful right at the start to introduce and define a few specialist terms from the vernacular of ancient numismatics, lest there be any confusion over their meaning due to their different use in everyday language. When referring to a “coin”, the reference is made to a specific and unique physical specimen. It is important to distinguish it from an “issue”, a more abstract notion that engenders all possible coins with the semantically identical design motifs. For example, two Roman Imperial coins may be said to correspond to the same issue if the same emperor, clothed in and oriented in a particular way, etc., is depicted on the obverse, and when the same inscriptions, and, say, deities in the identical poses and engaged in identical acts, etc., are shown on the reverse. Different issues of coins are uniquely referred by their identifiers from a variety of standard references, such as the Roman Imperial Coinage (RIC), as illustrated in Figure 1. A thorough summary of the relevant terminology for the non-specialist can be found in the recent review of computer vision challenges and problems in ancient numismatics by Arandjelović and Zachariou [1].
The determination of the issue that a particular coin corresponds to, that is its identification, is a task of foremost importance and the focus of the present article. In simple terms, it seeks to answer the question: “What coin is this?”. In most cases, this is a reasonably straightforward task for an expert, though there are exceptions, especially when the specimen in question is damaged and its issue is a rare one. For amateurs and especially beginner hobbyists, the challenge can be fiendish. For automatic, computer-based methods, the task has also proven to be a difficult one, both for reasons inherent in the problem as well as those that emerge from a variety of practical issues.
Given the aforementioned cardinality of coin identification, it is unsurprising that most of the work on the use of computer vision for ancient coin analysis has focused on solving precisely this problem. No more surprising is the overall approach that dominates the related literature. In particular, the structure of the coin identification problem is naturally seen through the lens of classification, with each issue seen as a class, thereby recasting the problem as that of assigning the correct class to an input image showing an unknown specimen [2,3,4,5]. Nevertheless, despite this superficial appeal of a classification-based approach to tackling the problem, it has become increasingly clear that its practical utility is highly limited by real world constraints pertaining to data availability. This becomes readily apparent as soon as the number of different coin issues is considered: there are over 50,000 for Roman Imperial coinage alone, and the number becomes far greater when Roman Republican, Roman Provincial, Ancient Greek, Celtic, etc., issues are included. It is practically impossible to obtain images of more than a small fraction of these, to say nothing of the need for multiple examples of each issue demanded by the present-day learning methods. There was some recognition of this major weakness of vision-based classification approaches already in the early years of work in this field [6], which has only become increasingly apparent since [7].
In the present paper, we propose a radically different approach whereby we learn to quantify the degree of belief that two specific specimens, e.g., a query and a gallery one, correspond to the same issue. Specifically, by learning what features should be extracted from an image of a coin for the purpose of its comparison with another coin and answering the question of whether they correspond to the same issue, in a manner independent of a specific comparison, we learn a fundamental representation of a coin that facilitates comparisons with new coins, that is coins that are not present in the training data set. By doing so, we learn a representation that does not rely on any a priori class structure. This means that if new images of coins are added to the reference gallery, be they additional examples of the already known classes or examples of an entirely new class, our model does not need to be retrained. It also means that our algorithm does not rely on multiple examples from the same class, as well as that its performance on underrepresented classes, which is a major issue for previous work, is not disadvantaged. The proposed approach provides a powerful way of making the most of the available information by facilitating different kinds of feedback to the user. Most obviously, if the best pairwise match is sufficiently high, the query coin can be attributed to the same issue as the corresponding match. On the other hand, if no such match is found, the gallery exemplars could be ordered by similarity, in a ranked retrieval fashion.
In terms of its technical underpinning, the present work pioneers two novelties. Firstly, this is the first work to describe the use of a Siamese architecture in this context. Secondly, as the baseline architectural component of each arm of the proposed Siamese network, we employ transformers, rather than convolutional neural networks that were featured in previous work [8,9,10].

2. Related Work

2.1. Automatic Ancient Coins Analysis Using Computer Vision and Machine Learning

Ancient coin analysis is a relatively new research domain for the application of computer vision and machine learning. The first forays into the territory were made a decade and a half ago by Zaharieva et al. [5]. The research effort in the field has since increased rapidly and dramatically [6,7,9,11,12,13,14,15,16], with an evermore varied range of specific tasks being targeted [1,6,17,18] and of modeling approaches [4,7,9,11,16,19].
Owing to the novelty of the problem, that is its unfamiliarity and the consequent lack of data, the earliest work centered its attention to what is arguably the simplest useful problem in ancient coin analysis which is specimen classification [5,20]. Thereafter, the focus in the field has quickly shifted towards the more challenging task issue classification. The reason for this lies in its broader practical interest as well as the greater technical challenge to automatic methods. Indeed, at present, the attention of nearly all existing computer vision work on ancient coin analysis is on issue classification [3,9,15,19,21], with a small number of notable exceptions [7,16,22].
In terms of technical methodology, the research on computer vision-based issue classification has largely mirrored the developments in computer vision more broadly. Thus, the initial attempts are addressing the challenge employed classical [2,4,5], that is non-learning, manually crafted features, e.g., SIFT [11] or wavelet transform [4] based descriptors, compared in a pairwise manner or aggregated using bagging [15]. Unlike in many natural image understanding applications, lacking in non-local geometric information, such representations quickly showed themselves to be insufficiently expressive for the task at hand. Hence, a number of follow-up methods sought to remedy this, for example by crafting geometric context aware features [19,21] or by aggregating local features in a spatially sensitive manner [15,23]. While effecting an improvement, such attempts have still proven insufficiently effective in producing a viable real-world solution; neither type of approach achieved sufficient expressive power nor robustness to the common challenges present in the data [23], with the latter kind of algorithms also suffering from sensitivity to the precise orientation of the coin, its centering, and variations across dies of the same issue [18].
Reflecting the trends in computer vision more broadly, a major leap in performance came with the adoption of deep learning [8,24]. Since then, a series of authors have demonstrated the power of deep learning, convolutional neural networks (CNNs) in particular, to address the key challenges that were thitherto insurmountable by classical computer vision approaches, namely intra-class variability caused by damage, the minting process, and different dies; and illumination and other photometric changes [3]. Complementing the CNN-based work, Zachariou et al. [16] recently demonstrated how a generative adversarial network can be used to synthetically reconstruct images of undamaged coins from original images of damaged specimens, thereby further directly addressing the major challenge that wear poses to automatic methods [23].
Notwithstanding the noted methodological improvements in the technical aspects of the methods proposed as a means of addressing the problem of ancient coin attribution, as recently pointed out by Cooper and Arandjelović [7], what has remained all but unchallenged in the 15 years of work is the fundamental manner of approach to the problem. In particular, the published work thus far frames attribution as a classification problem: given a known set of classes, each with images of exemplars, the correct class of an image of a novel specimen is sought. This is a reasonable framing if the classification is very coarse, e.g., by the denomination of a coin [22], when the premises of the setting are easily satisfied. Arguably, the strategy can also be defended when the classification is semi-coarse, e.g., when a class corresponds to the issuing authority on the coins’ obverses [8]; examples of coin images of only the rarest of issuers, which are few in number, may be problematic. However, when fine attribution is of interest, which is something that numismatists, be they hobbyists or professionals, are interested in first before any further analysis is conducted, then it can be readily seen that the classification paradigm is no longer viable. The reason lies in the very large number of emergent classes and the difficulty—or rather, impossibility in practice—of obtaining exemplars of but a small fraction of their total number. As noted by Arandjelović and Zachariou [1], Online Coins of the Roman Empire (OCRE; see http://numismatics.org/ocre/ accessed on 1 May 2023), a joint project of the American Numismatic Society and the Institute for the Study of the Ancient World at New York University, lists 43,000 published issues, and the true count is likely to be even greater. The only work to date that has tackled this challenge directly is that of Cooper and Arandjelović [7]. More precisely, what Cooper and Arandjelović propose is the first step towards overcoming the aforementioned problem, introducing a text mining and CNN-based method to learn to recognize the semantics of different elements depicted on coins, thereby transferring the representation from the image domain to the text one, the latter being far more abundant in data, easier to interpret, and simpler to match or otherwise analyze. Although their approach has demonstrated promising performance on a small number of frequently encountered concepts, at present there still remains a large gap between the method’s currently demonstrated capabilities and those needed to make the technology practically useful for the task of exact issue identification.
The method we introduce in the present paper emerges from the nexus of the described weaknesses of the previous work, while also drawing strength from ideas that have previous been showed to yield promising results. In particular, in order to overcome the difficulties associated with an extremely large number of classes (that is, coin issues), instead of seeking to learn a representation that distinguishes a particular class from all the others (classification), herein we seek a representation that is overall best at distinguishing classes one from another, thus relinquishing the demand for exemplars of any specific class. This leads to our adoption of the paradigm of pairwise coin matching by issue, rather than the usual classification paradigm, and the specific solution in the form of a Siamese neural network [25]. Furthermore, while adopting deep learning, motivated by its successes in the field and its unchallenged superiority over classical computer vision approaches, we also seek to leverage the advantages which transformers have over the previously employed convolutional neural networks [26], and in particular their non-local attention mechanisms, which ought to be particularly useful in ancient coin analysis by associating semantically but not visually related distal elements of a coin’s design (i.e., in the legend and in the pictorial motif).

2.2. Siamese Neural Networks

A Siamese neural network (SNN) [25], illustrated in Figure 2, is a kind of a coupling architecture. Comprising two mutually mirroring processing streams, it is based on two identical neural networks with shared hyperparameters. When fed two inputs from the same input space (images of coins in our case), it learns to produce their discriminative representations in a high-dimensional space. A comparison of these representations is also learned, ultimately producing a similarity score between them and thereby of the two original inputs they correspond to as well [25,27].
Siamese neural networks were first proposed in the context of signature verification, which is for the determination of signature forgeries [25]. Subsequently, they have been adopted and proven successful in a wide variety of other matching-based tasks, such as gait recognition [28], sporting activity recognition [29], natural language processing [30], object reconstruction [31] and others. Appropriately applied, Siamese neural-network-based algorithms have been shown to improve classification accuracy and enhance rejection quality compared with traditional convolutional neural networks [32]. Moreover, SNNs can significantly reduce the number of hyperparameters during model training and improve operational speed while maintaining their superior accuracy performance [33].

2.3. Transformers

The transformer [26] is a deep learning architecture originally proposed for natural language processing (NLP) applications, which revolutionized the field and led to new state-of-the-art models while also reducing training times for large data sets. Google’s Bidirectional Encoder Representations from Transformers (BERT) model [34] has been used to improve the search functionality for more complex queries. OpenAI’s Generative Pre-trained Transformer 3 (GPT-3) [35] became the largest neural network ever constructed, making headlines with its impressive ability to generate text that appeared to have been written by humans.
The transformer follows a similar encoder–decoder architecture to previous models, in which one sequence of tokens, representing words in a sentence, is used to generate another sequence (e.g., a translation of the sentence). What is special about the transformer architecture is that, unlike its predecessors, it does not use convolutional layers or recurrent connections, but instead largely relies on self-attention [26], a mechanism for focusing on information relevant to the current task. An attention unit’s role is to map equal-length sequences of query, key and value vectors to a sequence of context vectors, each of which is a weighted mean of the value vectors, weighted towards those that are most relevant to the corresponding position in the sequence for the given task. Attention weights are computed using three matrices that are learned during training, Q R n × d k , K R n × d k , and K R n × d v  [26], with their rows, respectively, being query, key, and value vectors, where n is the maximum sequence length, d k is the dimensionality of the query and key vectors, and d v is the dimensionality of the value vectors. In a translation context, the query vectors correspond to words in the target language, whereas the key and value vectors would correspond to words in the source language. Let the words of the input sentence be represented by the rows of X R n × d then the learnable embeddings W K R n × d k and W V R n × d v project the input X to the key matrix K = X W K and the value matrix V = X W V  [26]. Let the words of the translated output up to the current token be represented by Y R n × d , then a learnable embedding W Q R d × d k projects Y to the query matrix Q = Y W Q  [26].
The transformer architecture uses scaled dot-product attention whereby how much the j-th value vector contributes to the i-th context vector is determined by the dot-product of the corresponding query and key vectors. The dot-products are scaled by 1 / d k  [26], lest they become too large, resulting in problematically small gradients. Softmax is applied to the dot products to obtain positive weights that sum to one. The attention weights are then multiplied by the value vectors to obtain the context vectors.
Previous language models that used recurrent neural networks (RNNs) struggled to learn long-range dependencies between words in long sequences, because it processed tokens sequentially, meaning that any state passed forward in the network had to encode the entire sequence up to the current token, which became less effective the longer the sequence was. In the transformer model, the self-attention mechanism operates over the entire sequence of input symbols, so it is equally able to handle dependencies over any range. Another issue with sequential processing is that training could not be parallelized as effectively. The transformer employs multi-head self-attention (MSA), running multiple identical, but separately parameterized, self-attention units (“heads”) in parallel. This allows it to attend to different regions for different representations concurrently, which would not be possible with a single head, as the weighted mean of many many points of interest would result in a lack of focus on anything in particular. The output of a MSA unit is the concatenation of the output vectors from the individual heads, projected by a matrix back to vectors of dimensionality d m o d e l , which is constant throughout the transformer.
The encoder in the transformer consists of N identical layers. The input to the decoder is an embedding of the input sequence and the output feeds into the decoder. Each layer in the encoder has two sub-layers: the first is an MSA unit; the second is a fully connected feed-forward network (FFN), consisting of two linear transformations with a ReLU activation in between. Skip connections around each sub-layer are used. These are a widely used feature in deep learning that help with vanishing gradients. Furthermore, skip connections have been shown to improve the ability to learn by flattening the loss landscape [36]. As well as skip connections, layer normalization is applied to each sub-layer, as this has been show to improve training times [37].
The decoder also consists of N identical layers with further sub-layers. As for the encoder, a MSA unit and a FFN are two of the sub-layers, and skip connections and layer normalization are used. The extra sub-layer is a multi-head attention (not self-attention) unit, for which the key and value matrices come from the output of the encoder and the query matrix comes from the output of the preceding MSA unit. This is how information flows from the encoder into the decoder. The input to the decoder comprises the tokens of the output sequence.
Unlike the structure of an RNN or a CNN, the transformer architecture does not implicitly contain any notion of position for the input data. Instead, positional encodings for each token are added to the input embeddings that are passed into the encoder. Whereas an RNN or a CNN has a strong inductive bias towards locality, a transformer has few inductive biases, so it must learn the significance of positional relationships during training. This lack of a strong inductive bias makes transformers rather generic and able to model long-range dependencies, at a cost of worse performance for small training sets for which sensible inductive biases can be beneficial. Since transformers need a large amount of training data, transfer learning is typically used: a model that has been pre-trained on a large but more generic data set is fine-tuned with training data for a specific task, enabling it to make use of previously learned generalizations and thus avoiding the need for task-specific training from scratch.

Vision Transformer

The vision transformer (ViT) architecture [38] is a direct descendant of the transformer architecture. It follows the original transformer architecture closely, enabling existing efficient transformer implementations to be used with ease. Whereas transformer was designed for sequence-to-sequence language tasks and therefore had an encoder and a decoder, ViT is used for image classification tasks and so it only has an encoder, to which tokens representing an image are provided as input. One of the main design decisions for ViT was how to embed the image. A naive implementation of self-attention would allow each pixel to attend to every other pixel, resulting in O ( n 2 ) time and space complexity for images of n pixels. This would be prohibitively expensive, so a simplification has to be made. In the case of ViT, the simplification is to use image patch embeddings as the input tokens rather than pixels. Each image of width W and height H is divided into patches of P × P pixels, resulting in N = W H / P 2 patches, which is a small enough number to make self-attention across patches feasible. The square patches are flattened to vectors, which are projected by a learnable embedding to vectors of dimensionality d m o d e l , which is the size of the vectors used throughout the layers of the encoder.
As ViT is used for classification, an additional learnable class token embedding is passed in to the encoder as the zeroth “patch embedding”. After the final layer of the encoder, an additional FFN is added, which maps the context vectors from the zeroth position in the last layer to the image classes. During training, the network learns to encode in these output context vectors a representation of the image that is then used for classification purposes. During fine-tuning, a different FFN is used to project this image representation to the classes specific to the problem domain. In the FFNs of the encoder, GELU is used as the activation function, whereas the Transformer uses ReLU, but the authors offered no explanation for this modification. As for the Transformer, ViT includes positional information in the data passed to the encoder, otherwise spatial relationships between patches could not be learned. A 2D-aware positional embedding offered no significant improvement over a 1D positional embedding, so a 1D positional embedding is used instead, leaving it to the network to learn how the patch positions were spatially related to one another.
ViT was compared with CNNs [38], specifically ResNets [39], and hybrids of ViT and CNNs, for which the input sequence to the ViT was formed from the feature maps of a trained CNN, rather than image patches. While the hybrids outperformed ViT for smaller data sets, (presumably because the features already encoded at least local structure within the data), this performance difference vanished for larger data sets, demonstrating the ability of ViT to learn complex features without a strong inductive bias towards local features. As the size of the training dataset was scaled up to 300 million images, the performance of ViT continued to increase without reaching saturation, showing that more data are better when it comes to training ViT models.
For CNNs, the size of dependencies that can be represented by a feature at a given layer is limited by the receptive field for the feature. The size of the receptive field increases with depth. In contrast, ViTs can model long-range dependencies in their lowest layers. By visualizing the mean distance in the image space over which information was integrated for a given layer of ViT, it was found that some heads even in the lowest layers of ViT modeled long-range dependencies, whereas others were highly localized [15]. For the hybrid models tested, highly localized attention was less pronounced, suggesting that the role played by the highly localized attention heads was similar to that played by early convolutional layers in a CNN.

3. Proposed Methodology

To contextualize the design choices introduced in this section, remember the key practical problems that the present work seeks to address. The foremost one of these is the challenge of the input belonging to a class (coin issue) that was not present during training, which is something that no existing work has recognized fully or attempted to tackle. The second challenge is that of dramatic class imbalance, which has been noted in the relevant literature [7], but which has been left wanting in terms of a practicable and effective solution. The models we introduce here, all based on a Siamese architecture underlain by visual transformers, address both of the aforementioned challenges in a principled manner.

3.1. Proposed Network Architectures

In the present work, we propose and compare two different SNN-ViT-based architectures for ancient coin matching. The first one, hereafter referred to as the Single Siamese ViT, performs matching of obverses and reverses independently. The second architecture, hereafter referred to as the Double Siamese ViT, compares both the obverses and reverses, and integrates the obtained side-based scores into a single coin based similarity. In both cases, we employ a base ViT model pre-trained on imagenet-1k.

3.1.1. Single Siamese ViT

For the architecture of our obverse of reverse matching Single Siamese ViT, we adopted and adapted a generic Siamese Network as follows. Firstly, the backbone network of the network was replaced with a pre-trained ViT. Next, the semantic layer outputs of the two ViT models comprising the network and corresponding to the two streams processing the two inputs (obverses or reverses being matched) were flattened, and the absolute distance between them computed. Then, three linear layers and one batch normalization layer were used to reduce the dimension and produce the provisional output. Lastly, this output was passed through the sigmoid function to obtain a quasi-probability match measure, i.e., a number between 0 and 1. The architecture of this network can be seen in Figure 3.

3.1.2. Double Siamese ViT

Our Double Siamese ViT, which processes both the obverses and the reverses of two coins that are matched, is for the most part based on the already described Single Siamese ViT model, with changes and additions to the final layers of the network. The two Single Siamese ViT networks remain identical up to and including the computation of the absolute difference of their semantic layers. Following this stage, their outputs are concatenated and layer normalization applied to the concatenated result. This is followed by two fully connected layers. As before, the provisional output of the second layer is passed through the sigmoid function, thereby obtaining a quasi-probability matching score on the coin level. The architecture of this network can be seen in Figure 4.

3.2. Training Methodology and the Organization of Training Data

No less important than the architectures of the learning models to the success of our overall approach is the manner in which the training of these models is performed, that is, the methodology employed to make the best use of the available training data. In the present work, we had a total of 20,000 training images of coins (specimens) available, spanning 7000 different issues. With the conventional, classification approach pursued by the previous work, this would lead to 7000 classes. On the one hand, this is an imposing number of classes. Yet, it is vastly smaller than the number of potential classes, that is, the different Roman Imperial coin issues. What is more, there would have been a major challenge posed by high imbalance and few exemplars even for a large proportion of issues included.
In contrast is the approach we advocate herein, whereby the machine learning model learns coin characteristics, which allows for the discrimination of same issues vs. different issues on a pairwise basis. The challenge of class imbalance is inherently avoided (with a caveat upon we will elaborate shortly), as is that of a large number of classes. However, a new practical choice emerges, that of designing the training process in a manner that is feasible. In particular, the space of possible training inputs (coin pairs) is enormous, totaling C 20000 2 = 199,990,000 combinations. Even if only a single sample of each coin issue is considered, there are over C 7000 2 = 24,496,500 combinations, which is clearly impractical. However, the inherent non-reliance of our approach on the presence of any specific issue allowed us a straightforward way of dramatically downsizing the actual training set. In particular, for training we considered only those issues containing over 20 samples. Out of these, we randomly chose 14 for training our model, 3 for validating it, and the remaining 3 for its final testing, these being entirely unseen during the training-validation process. Doing this resulted in a training set containing 542 images representing 24 issues, and the test data set for the evaluation of the final model consisting of a total of 7605 issues over 196 individuals (emperors, empresses, etc.) depicted on their obverses. Figure 5 shows the distribution of the training set.
The process just described adequately addresses the challenges of a large number of classes, the consequent need for a vast amount of training data, and, partially, that of class imbalance. The latter challenge is at this stage overcome only partially because it is still the case that the number of all same-issue pairs still outnumbers the number of all different-issue pairs, risking the over-weighting of correct decisions when the input coins do belong to the same issue relative to the decisions when they do not. However, considering that the exemplar count of both is large, this remnant imbalance is resolved rather effortlessly. In particular, all that needs to be done and indeed what we did in this work, was to perform balanced sampling of same-issue and different-issue pairs. The flow chart of this process is summarized diagrammatically in Figure 6.

4. Results and Evaluation

Having described the technical specifics of our models in the previous section, we now turn to the empirical evaluation of the same. We start by presenting the results obtained using our Single Siamese ViT, separately matching coin obverses and reverses, and then follow up with an assessment of our Double Siamese ViT, which matches coins holistically, that is, both obverses and reverses jointly, thereby de facto matching the corresponding issues themselves.

4.1. Data

In this work, we made use of the large data set of ancient coin images provided by the Ancient Coins Search Engine (https://www.acsearch.info/ accessed on 1 May 2023) for research purposes, which has been used in a number of previous research efforts [10,16]. This corpus consists of high-quality images obtained in rather controlled environments, usually with a uniform background, favorable lighting, natural coin alignment, etc. Whilst including a variety of non-Roman coins (Greek, Celtic, and Byzantine, among others), as well as Roman non-Imperial ones (namely Provincial and Republican), the Roman Imperial coins included span the entire time period of the Empire and cover most of the obverse figures depicted on them and listed in the Appendix A in Table A1.
The acsearch data set in its raw form comprises images with the associated free-form textual descriptions as provided by auction houses. In other words, there is no semantically organized meta-information that would allow us to identify the entries that are of interest herein, namely Roman Imperial coins with the corresponding RIC identifiers. Considering the large size of the corpus and hence the impracticality of this being done by a human, we achieved the desired extraction automatically. In the processing of a single candidate entry, we first searched for the presence of the names listed in Table A3 in the associated text file. If none were found, or there were multiple different names found, the entry was not included in our experiments. The absence of a find suggests a coin other than Roman Imperial, whereas multiple matches meant that the entry was not a single item but a coin lot, or simply that there was ambiguity, which would have required a much more semantically nuanced data extraction method than was necessary for the extraction of a sufficient number of entries for the purposes of the present work. For entries that contained a single matching name, we next searched the text file for the RIC identifier using the regular expression “RIC.*?∖d”. Any entries without a match were also discarded; this would happen when another standard reference other than RIC was used (e.g., Roman Silver Coins (RSC)), or when a non-standard format for RIC was used. Finally, the images of the qualifying entries were split into two images, the obverse and the reverse, by dividing the image horizontally half way. No further efforts to register the resulting images were made, leaving any variation due to translation to be learned by our transformer-based, and hence patch-ordering-independent, model.

4.2. Single Siamese ViT

Recall that the proposed Single Siamese ViT is designed to match only a single side of a pair of coins, that is, either their obverses or reverses, and is accordingly trained with the corresponding sides only. Understanding the performance of this network, considering that it forms the basis of our more complex model, the Double Siamese ViT, evaluated subsequently, is crucial for understanding and contextualizing the performance of the latter. Further to providing insight into the power of the architecture itself and the manner in which we approach training, the findings presented here are also key to understanding how the network deals with the challenges presented by obverse and reverse motifs, which differ substantially. In particular, while obverses almost without an exception depict the head or the bust of a person (emperor, empress, heir, etc.) surrounded by a circularly arranged legend (text), the range of motifs on reverses is far more varied and complex, showing scenes (e.g., funeral pyres, bridges and building, rivers and forests, deities, etc.).

4.2.1. Obverse Matching

We turn our attention to the task of obverse matching first. As the plot in Figure 7a shows, save for stochastic oscillations, we observed a decrease in the training loss throughout the training process, that is, with additional training epochs. Nevertheless, the rate of loss decrease slows down significantly by epoch 100, which gives reassurance that not much further benefit would be conferred by longer training. The concurrent and mirroring behavior of the validation loss indicates successful learning and a well-fitted ultimate model. Indeed, evaluated on the test set, the model achieves the accuracy of 95.73%, which matches that of the final validation accuracy and is expectedly somewhat lower than the final training accuracy (see the accompanying plots in Figure 7b); the impressive corresponding ROC curve is shown in Figure 7c. Our test set accuracy significantly exceeds that achieved by previous work on the obverse matching task, e.g., that reported by a CNN-based approach of Schlag and Arandjelović [8]. Still, our result is even more astounding given that the exact problem addressed by Schlag and Arandjelović is weaker than ours: whereas they merely seek to match the depicted obverse persons’ identities, we tackle the more specific matching of the precise obverse issues, which requires not just the matching of the corresponding persons’ identities, but also of their dress and adornments, as well as obverse legends.

4.2.2. Reverse Matching

We next turn our attention to the task of reverse matching. As in the previously described experiments on coin obverses, in training we observe a declining loss, both on training and validation data, throughout the training process, with the decline slowing down considerably by the epoch 100; see Figure 8a. However, the differences between the two training processes are noteworthy and highlight a few insightful points, which we expected from the theory as explained previously. Firstly, notice that model improvement slows down earlier in the case of reverses, suggesting an inherent limitation in the model to learn further semantic nuance. This is important when one also observes that the final model loss, both on training and on validation data, ends up being significantly higher in the case of reverse matching than obverse matching, offering substantiation to our expectation that the greater complexity of reverse motifs is inherently more difficult to learn. These interpretations are additionally corroborated by the accuracy plots shown in Figure 8b. In particular, while the reverse training accuracy is almost insignificantly lower than the obverse training accuracy, the equivalent discrepancy between the validation accuracies is somewhat larger (while still small), and the final accuracy on the test data set even more so. The final accuracy achieved is 91.03% (compare this with 95.73% for obverses). The corresponding ROC curve is still impressive, though also not quite as close to the ideal as that achieved on the obverse matching task.

4.3. Double Siamese ViT

Empowered with an understanding of the strengths and weaknesses of our Single Siamese ViT, we finally evaluate the main model introduced in the present paper, namely our Double Siamese ViT, which uses Single Siamese ViT networks as its core building blocks. To overcome the computational challenge of training such a large network from scratch, and the problems associated with issues such as those of vanishing gradients and overfitting, we adopt the trained Single Siamese ViT networks of the previous section (one for the matching of images of obverses and one for the matching of images of reverses), freeze their weights, and train only the remainder of the architecture. Owing to this training design choice, we now observed rather different behavior of losses during training, as shown in Figure 9a. In particular, unlike during the training of the Single Siamese ViT on obverses and reverses, respectively, in Figure 7a and Figure 8a, here we note an initial increase in losses, which start to decline only following a peaking around the epoch 100. Thereafter, the behavior becomes much more familiar, the losses steadily declining following the peak, and settling by the epoch 500 (note the five-fold greater number of epochs needed as compared to the Single Siamese ViT). The greater challenge addressed by the Double Siamese ViT is also apparent from the accuracy plots in Figure 9b, with the training accuracy steadily and rather rapidly improving throughout the training process, reaching close to 100% performance by the epoch 500, contrasting the lack of validation accuracy improvement from as early as the epoch 100. The accuracy of the final, trained model was found to be 86.36%, which is impressive and far greater than that achieved by previous work on much simpler tasks, though understandably lower than the accuracy of the Single Siamese ViT on either of the sub-tasks of obverse or reverse only matching. Similar observations apply to the obtained ROC curve shown in Figure 9c.

Further Model Probing

While our Double Siamese ViT model achieved outstanding results, vastly outperforming the existing state-of-the-art, it expectedly did not perform perfectly, i.e., it could not match a human expert on the task of coin issue matching. Hence, we sought to understand the model’s performance with more nuance and gain insight into its strengths and weaknesses, both being important for future work and any potential improvements to it. As the first step towards this goal, we performed an additional set of experiments. In these experiments, we sought to match issues using (i) obverses only and (ii) reverses only, using our Single Siamese ViT model, and compared the results on an emperor-by-emperor basis with the joint matching performed by the proposed Double Siamese ViT model. Note that the single-side-based matching done here was different than that described in the previous section. In particular, while in the previous section we also used the Single Siamese ViT model to perform single side matching (obverse or reverse), a match was considered correct if it matched that side correctly. In contrast, here we take the match to extend to the entire issue. Clearly, in general, the information from only one side of the coin is insufficient to fully specify an issue, though in some cases it is (some issues feature obverse or reverse motifs or details not found on other issues), which is why a human would always examine the coin in its entirety when performing attribution. That is precisely the value of the approach taken in this experiment. Specifically, by making an emperor-by-emperor comparison, our findings illuminate both the magnitude of the value added of a joint consideration of both coin sides, as well as give insight into when this is most helpful. For example, we expected that the greatest gain would be seen when an issuing authority on the obverse is featured on many different issues, as well as when a particular motif recurs over long stretches of time (this would be the case, for example, for generic propaganda about prosperity and the virtues of the Empire, but not with one-of-a-kind events such as military victories).
The full numerical results over matching accuracies averaged across issuing authorities are presented in Table A2, Table A3 and Table A4; a graphical summary is shown in Figure 10. The immediately apparent finding is, as hypothesized, that the Double Siamese ViT model, i.e., issue matching using both coin sides, significantly outperforms both Single Siamese ViT models, i.e., issue matching using either side in isolation. The improvement is observed both on average as well as in the case of nearly every issuing authority; we shall return to the the unusual exceptions shortly. Observe that even when both single-side-based predictions perform poorly, their complementary role in the unique determination of an issue is reflected in the virtually universally highly accurate prediction when a coin is handled in a holistic manner. Indeed, the advantage of the Dual Siamese ViT model is particularly apparent when at least one of the two single side predictions is poor, e.g., because there are numerous issues under the same issuing authority (demonstrated by the poor predictive performance of obverses) or when a reverse motif is repeated across many issuing authorities (demonstrated by the poor predictive performance of reverses).

4.4. Analysis of Problematic Issuing Authorities

We noted previously that while the issuing authority averaged matching performance of the Dual Siamese ViT model is nearly universally high, there are some exceptions. In order to gain insight into this finding and discover a potential weakness of the proposed method, we identified the 15 most problematic issuing authorities, judged by the lowest average matching scores as per Table A4 and Figure 10, and manually examined the corresponding coin images. We readily identified a number of reasons for the aforementioned poorer than expected performance, most of which have to do with the quality of the available data rather than with any inherent, technical aspect of the proposed model itself. This is further elaborated next.

4.4.1. Physically Incomplete Specimens

Although our data set on the whole generally comprises good quality coin samples, a number of images show significantly damaged specimens, that is specimens which are either physically chipped or even cut in half. This is particularly important and noticeable as such specimens are of interest only in case of rare issues and rare issuing authorities, which are for this reason also least abundant in samples, their negative effect on the average performance being amplified by this fact. Examples of such specimens are shown in Figure 11a and Figure 12a, which feature significant semantic information loss as compared with well-preserved samples shown in, respectively, Figure 11b and Figure 12b. The obverse of the coin in Figure 11a is missing the head of Augustus, and the lettering in the top field is barely present. The reverse of the coin shows only the head of the crocodile, with the tree behind it entirely missing.

4.4.2. Worn and Environmentally Affected Coins

As a kind of currency, coins were continuously circulating in ancient times, resulting in surface wear and hence the loss of salient semantic detail crucial for their identification. Exposure to elements, e.g., due to being buried underground, can also effect wear, as well as surface appearance changes in the form of discoloration or patination. All of these factors confound the issue-based matching tasks. At the same time, there are statistical differences in how the coins of different issuers were affected. For example, heavy yet at the time lesser value coins such as sestertii, but which were gradually phased out over time, are more affected by physical wear, see Figure 13b; debased silver coins associated with the period of economic hardship of the Empire in the 3rd century AD are more easily affected by corrosion than good quality silver coins of the early empire, see Figure 13a; and so on.

4.4.3. Data Irregularities

Recall from Section 4.1 that a normal entry in our data set comprises an image that shows a single coin specimen, its obverse on the left hand side and its reverse on the right hand side, in the natural canonical orientation. However, our examination of problematic text exemplars revealed that a small but not negligible number of the entries in the corpus do not conform with the aforementioned assumption and were not filtered out by our data pre-processing also described in Section 4.1. Examples are shown in Figure 14.

4.4.4. High Similarity between Issues

Lastly, a number of erroneous matches made by our method can be attributed to the inherent difficulty in distinguishing between certain issues that differ in minute detail only. An example is shown in Figure 15, which shows issues RIC 158 and RIC 160 of Augustus. These have identical reverses, with the legend COL NEM and the motif showing a crocodile chained to a palm-shoot with long vertical fronds and tip left, and a wreath with long ties above on the left. Their obverses are virtually identical too, with the legend IMP DIVI F and the heads of Agrippa (left) and Augustus (right) back to back (Agrippa wearing a combined rostral crown and laurel wreath, and Augustus laureate), the sole difference being the lettering P P in the field of RIC 160. We identified such only subtly different pairs of issues for Domitia, Saloninus, Macrianus, Fausta, Britannicus, Vabalathus, Julia Paula, Valentinian III, and Octavia.

5. Conclusions and Future Work

In this work, our attention was on the problem of image-based ancient coin attribution, which has been at the focus of research on the use of computer vision in ancient numismatics since the nascence of the field. We commenced the article by contextualizing and motivating our key technical contribution, discussing the key limitations of the existing work in the field, both methodological and practical ones. Among the latter, we highlighted the hitherto almost entirely overlooked problem that emerges from the dominant type of approach to ancient coin attribution (namely that in the form of classification), which is the extremely large number of classes (10 s of thousands) for most of which training exemplars are unavailable. This makes the existing algorithms unable to deal with coins of unseen issues, requires a retraining of models when new class exemplars become available, and presents a major class imbalance challenge. Hence, we argued against the classification paradigm and in favor of an alternative. In particular, rather than trying to learn a class specific representation that distinguishes a particular class from all the others, we presented a case for seeking a representation that is overall best at distinguishing classes one from another, thus relinquishing the demand for exemplars of all classes or indeed of any specific class. This led to our adoption of the paradigm of pairwise coin matching by issue, and the specific technical approach in the form of a purpose-crafted Siamese neural network. Furthermore, while adopting deep learning, motivated by its successes in the field and its unchallenged superiority over classical computer vision approaches, we also sought to leverage the advantages that transformers have over the previously employed convolutional neural networks, and in particular their non-local attention mechanisms which ought to be particularly useful in ancient coin analysis by associating semantically but not visually related distal elements of a coin’s design. Finally, we presented a comprehensive and detailed evaluation of the proposed method using a large data corpus of 14,820 images and 7605 issues, and an in-depth analysis of its strengths and weaknesses. Using transfer learning and only a small training set of 542 images of 24 issues, our Double Siamese ViT model was shown to surpass the state of the art by a large margin, achieving an overall accuracy of 81%. Our further investigation of the results showed that the majority of the method’s errors are unrelated to the intrinsic aspects of the algorithm itself, but are rather a consequence of unclean data, which is a problem that can be easily addressed in practice by simple pre-processing and quality checking.
The success of the proposed method and the presented experimental results suggest a number of avenues for further research, which we are currently exploring. Firstly, we expect that an improvement in performance can be effected by training separate Double Siamese ViT models for different kinds of coins: e.g., most coarsely for Roman, Greek, Byzantine, Celtic, etc.; on a finer basis for, e.g., Roman Republican, Roman Imperial preceding the Crisis of the Third Century (which resulted in major changes in both the material and style of coinage), and late Roman Imperial coins; or even for different denominations that exhibit differences both in style and content due to the their different flan sizes and materials used. Secondly, we aim to explore if further informative inference could be made for unknown issues, i.e., issues which are not matched to any gallery ones. The idea here would be to make inferences based on the most similar issues, though not sufficiently similar to produce a match, in a manner conceptually similar to that which has demonstrated success in the context of face recognition, among others [40,41].

Author Contributions

Conceptualization, J.B. (Siamese structure), Z.G. and O.A. (transformer components); methodology, Z.G., O.A. and Y.L.; software, Z.G. and Y.L.; investigation, Z.G. and D.R.; resources, Z.G. and O.A.; data curation, O.A.; writing—original draft preparation, Z.G., O.A., D.R. and Y.L.; writing—review and editing, Z.G. and O.A.; visualization, Z.G.; supervision, O.A.; project administration, O.A.; after initial publication, J.B. has agreed to be added as a co-author. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data set used in the present article can be obtained freely for research purposes by contacting the authors.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. List of possible issuing authorities, including non-conformal entries such as City Commemoratives.
Table A1. List of possible issuing authorities, including non-conformal entries such as City Commemoratives.
Aelia AriadneFaustina IINero Claudius Drusus
Aelia FlacillaFlavia TitianaNerva
Aelia VerinaFlavius VictorNigrinian
AeliusFlorianNumerian
AemilianGaius and LuciusOctavia
AgrippaGalbaOrbiana
Agrippa PostumusGaleria ValeriaOtacilia Severa
Agrippina IGaleriusOtho
Agrippina IIGalerius AntoninusPacatian
AlexanderGalla PlacidiaPaulina
AllectusGallienusPertinax
Annia FaustinaGemellusPescennius Niger
Annius VerusGermanicusPetronius Maximus
AnonymousGetaPhilip I
AnthemiusGlyceriusPhilip II
AntinousGordian IPlautilla
AntoniaGordian IIPlotina
Antoninus PiusGordian IIIPoppaea
Aquilia SeveraGratianPostumus
ArcadiusHadrianPriscus Attalus
Asinius GallusHanniballianusProbus
AugustusHelenaProcopius
AurelianHerennia EtruscillaProculus
AureolusHerennius EtruscusPulcheria
AvitusHonoriaPupienus
BalbinusHonoriusQuietus
BasiliscusHostilianQuintillus
BonosusJohannesRegalianus
BritannicusJotapianRomulus
CaesoniaJovianRomulus Augustus
Caius and LuciusJovinusSabina
CaligulaJuliaSalonina
CaracallaJulia DomnaSaloninus
CarausiusJulia MaesaSebastianus
CarinusJulia MamaeaSejanus
CarusJulia PaulaSeptimius Severus
City CommemorativesJulia SoaemiasSeverina
Civil WarsJulia TitiSeverus II
ClaudiusJulian ISeverus III
Claudius II (Gothicus)Julian IISeverus Alexander
Clodius AlbinusJulius MarinusStatilia Messalina
Clodius MacerJulius NeposTacitus
CommodusLaelianusTesserae etc.
ConstansLeo ITetricus I
Constantine ILeo IITetricus II
Constantine IILiboTheodora
Constantine IIILicinia EudoxiaTheodosius I
Constantius ILicinius ITheodosius II
Constantius IILicinius IITiberius
Constantius IIILiviaTitus
Constantius GallusLivillaTrajan
Cornelia SuperaLucillaTrajan Decius
CrispinaLucius VerusTranquillina
CrispusMacrianusTrebonianus Gallus
DecentiusMacrinusUranius Antoninus
DelmatiusMagnentiusVabalathus
DiadumenianMagnia UrbicaValens
Didia ClaraMagnus MaximusValentinian I
Didius JulianusMajorianValentinian II
DiocletianManlia ScantillaValentinian III
DomitiaMarcianValeria Messalina
Domitia LucillaMarcianaValerian I
DomitianMarcus AureliusValerian II
DomitianusMarinianaValerius Valens
Domitilla IMariusVarbanov
Domitilla IIMartinianVarus
Domitius DomitianusMatidiaVespasian
DrususMaxentiusVespasian II
DryantillaMaximianusVetranio
ElagabalusMaximinus IVictorinus
EudociaMaximinus IIVindix
EudoxiaMaximusVitellius
EugeniusMaximus of SpainVolusian
Fabius MaximusNepotianZeno
FaustaNeroZenobia
Faustina INero and Drusus CaesarsZenonis
Table A2. Obverse matching performance averaged across each issuing authority.
Table A2. Obverse matching performance averaged across each issuing authority.
Issuing AuthorityCorrectTotalRate
Aelius24041358.1%
Aemilian3130810.1%
Agrippa51631.3%
Agrippina II51145.5%
Alexander447856.4%
Allectus228226.8%
Anthemius304566.7%
Antoninus Pius19,83242,55046.6%
Aquilia Severa41330.8%
Arcadius22941655.0%
Augustus68,606113,23660.6%
Aurelian829477817.4%
Balbinus248727.6%
Basiliscus425773.7%
Britannicus12621558.6%
Caligula58595861.1%
Caracalla698417,25940.5%
Carausius7644217.2%
Carinus13856524.4%
Carus3115420.1%
Claudius675710,06367.1%
Clodius Albinus4915930.8%
Commodus4280934445.8%
Constans1521466932.6%
Constantine I1340332740.3%
Constantine II4810645.3%
Constantius Gallus5915637.8%
Constantius I478133835.7%
Constantius II2360658135.9%
Constantius III31520.0%
Crispina8726133.3%
Crispus347143824.1%
Decentius2385.3%
Delmatius175729.8%
Diadumenian2314016.4%
Didius Julianus183158.1%
Diocletian3226716345.0%
Domitia61540.0%
Domitia Lucilla3310631.1%
Domitian701712,73755.1%
Domitianus3625634457.1%
Domitius Domitianus199220.7%
Elagabalus943309630.5%
Eudocia41040.0%
Eugenius123138.7%
Fausta154831.3%
Faustina I4423318.9%
Faustina II134528.9%
Flavius Victor61735.3%
Florian5796.3%
Gaius and Lucius162955.2%
Galba1038242442.8%
Galeria Valeria203754.1%
Galerius490118141.5%
Gallienus323911,20528.9%
Germanicus5448880461.9%
Geta293129222.7%
Gordian I2238.7%
Gordian III565209527.0%
Gratian3214422.2%
Hadrian25,63153,70247.7%
Helena2515416.2%
Herennia Etruscilla5711151.4%
Herennius Etruscus8723936.4%
Honorius858133664.2%
Hostilian176924.6%
Johannes91656.3%
Jovian116816.2%
Julia25650151.1%
Julia Domna471160729.3%
Julia Maesa228426.2%
Julia Mamaea4915930.8%
Julia Paula111957.9%
Julia Titi7012356.9%
Julian II12867219.0%
Julius Nepos51050.0%
Leo I10115565.2%
Licinius I417149427.9%
Licinius II802286628.0%
Livia8912670.6%
Lucilla9126134.9%
Lucius Verus1330279347.6%
Macrianus33010.0%
Macrinus251124020.2%
Magnentius21675228.7%
Magnia Urbica307042.9%
Magnus Maximus10628337.5%
Manlia Scantilla31127.3%
Marcian10713778.1%
Marciana165529.1%
Marcus Aurelius57,982112,71451.4%
Mariniana82532.0%
Marius52321.7%
Maxentius13144129.7%
Maximianus1209367732.9%
Maximinus I12842030.5%
Maximinus II5210748.6%
Maximus7912463.7%
Nero747614,40951.9%
Nero Claudius Drusus579560.0%
Nerva723167243.2%
Numerian6319232.8%
Octavia20550240.8%
Otacilia Severa5526420.8%
Otho327244.4%
Pacatian92832.1%
Pertinax11926145.6%
Pescennius Niger9528133.8%
Philip I337118828.4%
Philip II3514324.5%
Plautilla13830745.0%
Plotina61250.0%
Postumus591263922.4%
Probus201310,03320.1%
Procopius3411030.9%
Pupienus4210141.6%
Quietus2238.7%
Quintillus9223539.1%
Romulus25957045.4%
Sabina7515747.8%
Salonina3719019.5%
Saloninus83125.8%
Septimius Severus11,67625,25246.2%
Severina1712713.4%
Severus Alexander647712,93150.1%
Severus II11421353.5%
Severus III61060.0%
Tacitus13153724.4%
Tetricus I1214.8%
Theodosius I19763730.9%
Theodosius II865144260.0%
Tiberius1687230773.1%
Titus3494612457.1%
Trajan11,17725,25744.3%
Trajan Decius12735635.7%
Tranquillina225738.6%
Trebonianus Gallus466119339.1%
Vabalathus32213.6%
Valens5330917.2%
Valentinian I2814419.4%
Valentinian II2613918.7%
Valentinian III14538238.0%
Valerian I7432822.6%
Vespasian17,93531,93356.2%
Vetranio6538416.9%
Victorinus6823628.8%
Vitellius3089507060.9%
Volusian17567525.9%
Zeno10715668.6%
30,2974613,08749.4%
Table A3. Reverse matching performance averaged across each issuing authority.
Table A3. Reverse matching performance averaged across each issuing authority.
Issuing AuthorityCorrectTotalRate
Aelius29341370.9%
Aemilian13330943.0%
Agrippa121770.6%
Agrippina II61060.0%
Alexander397850.0%
Allectus548067.5%
Anthemius264656.5%
Antoninus Pius29,64142,54969.7%
Aquilia Severa101283.3%
Arcadius31041874.2%
Augustus79,227113,25370.0%
Aurelian2695477256.5%
Balbinus568565.9%
Basiliscus325756.1%
Britannicus7521235.4%
Caligula71196174.0%
Caracalla10,72817,26462.1%
Carausius33243975.6%
Carinus33956560.0%
Carus7115446.1%
Claudius759810,06675.5%
Clodius Albinus7516046.9%
Commodus6720933872.0%
Constans2193466947.0%
Constantine I2271332868.2%
Constantine II7410570.5%
Constantius Gallus10315367.3%
Constantius I749133955.9%
Constantius II4472657768.0%
Constantius III121580.0%
Crispina16825964.9%
Crispus943143765.6%
Decentius223857.9%
Delmatius235839.7%
Diadumenian8914362.2%
Didius Julianus183158.1%
Diocletian4950716869.1%
Domitia101566.7%
Domitia Lucilla2610724.3%
Domitian880312,73969.1%
Domitianus4419634269.7%
Domitius Domitianus529256.5%
Elagabalus1705309955.0%
Eudocia41136.4%
Eugenius163151.6%
Fausta204841.7%
Faustina I15623267.2%
Faustina II334573.3%
Flavius Victor81747.1%
Florian338041.3%
Gaius and Lucius192965.5%
Galba1565242864.5%
Galeria Valeria123732.4%
Galerius828118170.1%
Gallienus723811,19964.6%
Germanicus4576881151.9%
Geta688128953.4%
Gordian I152268.2%
Gordian III1176209456.2%
Gratian8414458.3%
Hadrian37,71253,70470.2%
Helena6015638.5%
Herennia Etruscilla8211074.5%
Herennius Etruscus16523669.9%
Honorius942133770.5%
Hostilian376854.4%
Johannes81650.0%
Jovian476968.1%
Julia37050373.6%
Julia Domna975160660.7%
Julia Maesa588469.0%
Julia Mamaea11815974.2%
Julia Paula151883.3%
Julia Titi10212482.3%
Julian II41267561.0%
Julius Nepos71070.0%
Leo I10215665.4%
Licinius I916149461.3%
Licinius II1227286942.8%
Livia9212673.0%
Lucilla18226369.2%
Lucius Verus2082279374.5%
Macrianus223171.0%
Macrinus660124253.1%
Magnentius53974772.2%
Magnia Urbica366852.9%
Magnus Maximus17628262.4%
Manlia Scantilla21118.2%
Marcian10513975.5%
Marciana385569.1%
Marcus Aurelius79,759112,71770.8%
Mariniana172568.0%
Marius102540.0%
Maxentius17543939.9%
Maximianus2509367568.3%
Maximinus I27841966.3%
Maximinus II6210956.9%
Maximus9412475.8%
Nero10,22714,41271.0%
Nero Claudius Drusus669470.2%
Nerva1143167368.3%
Numerian9219247.9%
Octavia28050555.4%
Otacilia Severa18526270.6%
Otho377350.7%
Pacatian222878.6%
Pertinax15226058.5%
Pescennius Niger16227858.3%
Philip I785118666.2%
Philip II9714168.8%
Plautilla16330952.8%
Plotina101283.3%
Postumus1776264567.1%
Probus586610,03058.5%
Procopius6010955.0%
Pupienus689869.4%
Quietus72330.4%
Quintillus12823454.7%
Romulus32456757.1%
Sabina9415759.9%
Salonina9719250.5%
Saloninus133043.3%
Septimius Severus17,82225,25270.6%
Severina8012962.0%
Severus Alexander847212,93365.5%
Severus II13921165.9%
Severus III71070.0%
Tacitus32553560.7%
Tetricus I152268.2%
Theodosius I43663668.6%
Theodosius II731144450.6%
Tiberius1848230480.2%
Titus4348612671.0%
Trajan18,42125,25872.9%
Trajan Decius26836074.4%
Tranquillina465780.7%
Trebonianus Gallus832118970.0%
Vabalathus152268.2%
Valens23731076.5%
Valentinian I9914170.2%
Valentinian II10513975.5%
Valentinian III23638261.8%
Valerian I17632853.7%
Vespasian23,50831,93473.6%
Vetranio19938651.6%
Victorinus16523769.6%
Vitellius3804507075.0%
Volusian46067767.9%
Zeno8915557.4%
421,686613,11168.8%
Table A4. Holistic matching performance averaged across each issuing authority.
Table A4. Holistic matching performance averaged across each issuing authority.
Issuing AuthorityCorrectTotalRate
Aelius35041185.2%
Aemilian24530879.5%
Agrippa101758.8%
Agrippina II71163.6%
Alexander717891.0%
Allectus608075.0%
Anthemius344673.9%
Antoninus Pius35,16942,55382.6%
Aquilia Severa121392.3%
Arcadius37641790.2%
Augustus90,666113,23680.1%
Aurelian3666477376.8%
Balbinus658675.6%
Basiliscus395768.4%
Britannicus13921564.7%
Caligula77296280.2%
Caracalla14,57017,26384.4%
Carausius32343973.6%
Carinus45456680.2%
Carus14215591.6%
Claudius807410,06780.2%
Clodius Albinus11916074.4%
Commodus7681934382.2%
Constans3283466770.3%
Constantine I2827332884.9%
Constantine II8910684.0%
Constantius Gallus12415779.0%
Constantius I1221133991.2%
Constantius II5616657985.4%
Constantius III1515100.0%
Crispina22126284.4%
Crispus1120143877.9%
Decentius323982.1%
Delmatius485784.2%
Diadumenian11414379.7%
Didius Julianus253278.1%
Diocletian5787716280.8%
Domitia61540.0%
Domitia Lucilla8910684.0%
Domitian10,41712,73981.8%
Domitianus5269634483.1%
Domitius Domitianus789284.8%
Elagabalus2598310183.8%
Eudocia91275.0%
Eugenius273187.1%
Fausta274757.4%
Faustina I21523292.7%
Faustina II364580.0%
Flavius Victor141782.4%
Florian748092.5%
Gaius and Lucius212875.0%
Galba1774242873.1%
Galeria Valeria363894.7%
Galerius1037118387.7%
Gallienus949611,20084.8%
Germanicus6819881577.4%
Geta1044129180.9%
Gordian I172373.9%
Gordian III1739209782.9%
Gratian12314485.4%
Hadrian44,30653,70982.5%
Helena13215784.1%
Herennia Etruscilla7910972.5%
Herennius Etruscus17123572.8%
Honorius1154133586.4%
Hostilian546978.3%
Johannes1515100.0%
Jovian616988.4%
Julia42050383.5%
Julia Domna1259160778.3%
Julia Maesa588369.9%
Julia Mamaea13216082.5%
Julia Paula121866.7%
Julia Titi11512294.3%
Julian II51267076.4%
Julius Nepos71070.0%
Leo I12615680.8%
Licinius I1123149475.2%
Licinius II2102286873.3%
Livia11012687.3%
Lucilla16426162.8%
Lucius Verus2252279280.7%
Macrianus183256.3%
Macrinus1044124084.2%
Magnentius60074880.2%
Magnia Urbica476869.1%
Magnus Maximus22228478.2%
Manlia Scantilla101283.3%
Marcian12313889.1%
Marciana475683.9%
Marcus Aurelius91,099112,70480.8%
Mariniana132356.5%
Marius192673.1%
Maxentius37943986.3%
Maximianus3111366984.8%
Maximinus I32142276.1%
Maximinus II9010982.6%
Maximus10712387.0%
Nero11,09514,40877.0%
Nero Claudius Drusus749478.7%
Nerva1343167480.2%
Numerian16019382.9%
Octavia34850369.2%
Otacilia Severa21026180.5%
Otho647190.1%
Pacatian232979.3%
Pertinax20226277.1%
Pescennius Niger22227979.6%
Philip I1013118885.3%
Philip II12414287.3%
Plautilla23431175.2%
Plotina91090.0%
Postumus2022264176.6%
Probus836610,03983.3%
Procopius8410977.1%
Pupienus729972.7%
Quietus172373.9%
Quintillus18523379.4%
Romulus48756586.2%
Sabina11315672.4%
Salonina15819182.7%
Saloninus132846.4%
Septimius Severus20,86925,25482.6%
Severina9912976.7%
Severus Alexander10,08112,92978.0%
Severus II18721089.0%
Severus III91090.0%
Tacitus41553577.6%
Tetricus I212295.5%
Theodosius I54863985.8%
Theodosius II1048144672.5%
Tiberius1912230483.0%
Titus4950611780.9%
Trajan20,73125,25982.1%
Trajan Decius31235687.6%
Tranquillina455680.4%
Trebonianus Gallus996119683.3%
Vabalathus152365.2%
Valens25631282.1%
Valentinian I12014185.1%
Valentinian II11313981.3%
Valentinian III26338169.0%
Valerian I23632872.0%
Vespasian25,73431,93380.6%
Vetranio30238877.8%
Victorinus18123676.7%
Vitellius3945507777.7%
Volusian58067685.8%
Zeno13115684.0%
496,882613,11081.0%

References

  1. Arandjelović, O.; Zachariou, M. Images of Roman imperial denarii: A curated data set for the evaluation of computer vision algorithms applied to ancient numismatics, and an overview of challenges in the field. Sci 2020, 2, 91. [Google Scholar] [CrossRef]
  2. Huber-Mörk, R.; Nölle, M.; Rubik, M.; Hödlmoser, M.; Kampel, M.; Zambanini, S. Automatic coin classification and identification. In Advances in Object Recognition Systems; Oxford University Press: Oxford, UK, 2012; Volume 127. [Google Scholar]
  3. Kiourt, C.; Evangelidis, V. AnCoins: Image-Based Automated Identification of Ancient Coins Through Transfer Learning Approaches. In Pattern Recognition, Proceedings of the ICPR International Workshops and Challenges, Virtual Event, 10–15 January 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 54–67. [Google Scholar]
  4. Wei, K.; He, B.; Wang, F.; Zhang, T.; Ding, Q. A novel method for classification of ancient coins based on image textures. In Proceedings of the Workshop on Digital Media and Its Application in Museum & Heritages, Chongqing, China, 10–12 December 2007; pp. 63–66. [Google Scholar]
  5. Zaharieva, M.; Kampel, M.; Zambanini, S. Image based recognition of ancient coins. In Computer Analysis of Images and Patterns, Proceedings of the 12th International Conference, CAIP 2007, Vienna, Austria, 27–29 August 2007; Springer: Berlin/Heidelberg, Germany, 2007; pp. 547–554. [Google Scholar]
  6. Arandjelović, O. Reading ancient coins: Automatically identifying denarii using obverse legend seeded retrieval. In Computer Vision–ECCV 2012, Proceedings of the 12th European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; Springer: Berlin/Heidelberg, Germany, 2012; pp. 317–330. [Google Scholar]
  7. Cooper, J.; Arandjelović, O. Understanding ancient coin images. In Recent Advances in Big Data and Deep Learning, Proceedings of the INNS Big Data and Deep Learning Conference INNSBDDL2019, held at Sestri Levante, Genova, Italy 16–18 April 2019; Springer: Berlin/Heidelberg, Germany, 2020; pp. 330–340. [Google Scholar]
  8. Schlag, I.; Arandjelovic, O. Ancient Roman coin recognition in the wild using deep learning based recognition of artistically depicted face profiles. In Proceedings of the International Conference on Computer Vision Workshops, Venice, Italy, 22–29 October 2017; pp. 2898–2906. [Google Scholar]
  9. Aslan, S.; Vascon, S.; Pelillo, M. Two sides of the same coin: Improved ancient coin classification using Graph Transduction Games. Pattern Recognit. Lett. 2020, 131, 158–165. [Google Scholar] [CrossRef]
  10. Cooper, J.; Arandjelović, O. Learning to Describe: A New Approach to Computer Vision Based Ancient Coin Analysis. Sci 2020, 2, 27. [Google Scholar] [CrossRef] [Green Version]
  11. Kampel, M.; Zaharieva, M. Recognizing ancient coins based on local features. In Advances in Visual Computing, Proceedings of the 4th International Symposium, ISVC 2008, Las Vegas, NV, USA, 1–3 December 2008; Springer: Berlin/Heidelberg, Germany, 2008; pp. 11–22. [Google Scholar]
  12. Kampel, M.; Huber-Mörk, R.; Zaharieva, M. Image-based retrieval and identification of ancient coins. IEEE Intell. Syst. 2009, 24, 26–34. [Google Scholar] [CrossRef]
  13. Zambanini, S.; Kampel, M. Robust Automatic Segmentation of Ancient Coins. In Proceedings of the International Conference on Computer Vision Theory and Applications, Lisboa, Portugal, 5–8 February 2009; pp. 273–276. [Google Scholar]
  14. Huber-Mörk, R.; Zambanini, S.; Zaharieva, M.; Kampel, M. Identification of ancient coins based on fusion of shape and local features. Mach. Vis. Appl. 2011, 22, 983–994. [Google Scholar] [CrossRef]
  15. Anwar, H.; Zambanini, S.; Kampel, M. Supporting ancient coin classification by image-based reverse side symbol recognition. In Computer Analysis of Images and Patterns, Proceedings of the 15th International Conference, CAIP 2013, York, UK, 27–29 August 2013; Springer: Berlin/Heidelberg, Germany, 2013; pp. 17–25. [Google Scholar]
  16. Zachariou, M.; Dimitriou, N.; Arandjelović, O. Visual reconstruction of ancient coins using cycle-consistent generative adversarial networks. Sci 2020, 2, 52. [Google Scholar] [CrossRef]
  17. Anwar, H.; Zambanini, S.; Kampel, M.; Vondrovec, K. Ancient coin classification using reverse motif recognition: Image-based classification of roman republican coins. IEEE Signal Process. Mag. 2015, 32, 64–74. [Google Scholar] [CrossRef]
  18. Conn, B.; Arandjelović, O. Towards computer vision based ancient coin recognition in the wild—Automatic reliable image preprocessing and normalization. In Proceedings of the International Joint Conference on Neural Networks, Anchorage, AK, USA, 14–19 May 2017; pp. 1457–1464. [Google Scholar]
  19. Arandjelović, O. Automatic attribution of ancient Roman imperial coins. In Proceedings of the Computer Vision and Pattern Recognition Conference, San Francisco, CA, USA, 13–18 June 2010; pp. 1728–1734. [Google Scholar]
  20. Zaharieva, M.; Huber-Mörk, R.; Nölle, M.; Kampel, M. On ancient coin classification. In Proceedings of the International Symposium on Virtual Reality, Archaeology and Intelligent Cultural Heritage, Brighton, UK, 26–30 November 2007; pp. 55–62. [Google Scholar]
  21. Anwar, H.; Zambanini, S.; Kampel, M. Encoding spatial arrangements of visual words for rotation-invariant image classification. In Pattern Recognition, Proceedings of the 36th German Conference, GCPR 2014, Münster, Germany, 2–5 September 2014; Springer: Berlin/Heidelberg, Germany, 2014; pp. 443–452. [Google Scholar]
  22. Ma, Y.; Arandjelović, O. Classification of ancient roman coins by denomination using colour, a forgotten feature in automatic ancient coin analysis. Sci 2020, 2, 37. [Google Scholar] [CrossRef]
  23. Fare, C.; Arandjelović, O. Ancient roman coin retrieval: A systematic examination of the effects of coin grade. In Advances in Information Retrieval, Proceedings of the 39th European Conference on IR Research, ECIR 2017, Aberdeen, UK, 8–13April 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 410–423. [Google Scholar]
  24. Kim, J.; Pavlovic, V. Discovering characteristic landmarks on ancient coins using convolutional networks. J. Electron. Imaging 2017, 26, 011018. [Google Scholar] [CrossRef] [Green Version]
  25. Bromley, J.; Guyon, I.; LeCun, Y.; Säckinger, E.; Shah, R. Signature verification using a “Siamese” time delay neural network. Adv. Neural Inf. Process. Syst. 1993, 6, 737–744. [Google Scholar] [CrossRef] [Green Version]
  26. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Transformer: Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 6000–6010. [Google Scholar]
  27. Chicco, D. Siamese neural networks: An overview. In Artificial Neural Networks; Springer: Berlin/Heidelberg, Germany, 2021; pp. 73–94. [Google Scholar] [CrossRef]
  28. Zhang, C.; Liu, W.; Ma, H.; Fu, H. Siamese neural network based gait recognition for human identification. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Shanghai, China, 20–25 March 2016; pp. 2832–2836. [Google Scholar]
  29. Long, T. Research on application of athlete gesture tracking algorithms based on deep learning. J. Ambient. Intell. Humaniz. Comput. 2020, 11, 3649–3657. [Google Scholar] [CrossRef]
  30. Ichida, A.Y.; Meneguzzi, F.; Ruiz, D.D. Measuring semantic similarity between sentences using a Siamese neural network. In Proceedings of the International Joint Conference on Neural Networks, Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–7. [Google Scholar]
  31. Ostertag, C.; Beurton-Aimar, M. Matching ostraca fragments using a Siamese neural network. Pattern Recognit. Lett. 2020, 131, 336–340. [Google Scholar] [CrossRef]
  32. Berlemont, S.; Lefebvre, G.; Duffner, S.; Garcia, C. Siamese neural network based similarity metric for inertial gesture classification and rejection. In Proceedings of the IEEE International Conference and Workshops on Automatic Face and Gesture Recognition, Ljubljana, Slovenia, 4–8 May 2015; Volume 1, pp. 1–6. [Google Scholar]
  33. Kim, M.; Alletto, S.; Rigazio, L. Similarity mapping with enhanced Siamese network for multi-object tracking. arXiv 2016, arXiv:1609.09156. [Google Scholar]
  34. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  35. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
  36. Li, H.; Xu, Z.; Taylor, G.; Studer, C.; Goldstein, T. Visualizing the loss landscape of neural nets. Adv. Neural Inf. Process. Syst. 2018, 31, 6391–6401. [Google Scholar]
  37. Ba, J.L.; Kiros, J.R.; Hinton, G.E. Layer normalization. arXiv 2016, arXiv:1607.06450. [Google Scholar]
  38. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  39. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the Computer Vision and Pattern Recognition Conference, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  40. Arandjelović, O. Reimagining the central challenge of face recognition: Turning a problem into an advantage. Pattern Recognit. 2018, 83, 388–400. [Google Scholar] [CrossRef] [Green Version]
  41. Arandjelovic, O. Learnt quasi-transitive similarity for retrieval from large collections of faces. In Proceedings of the Computer Vision and Pattern Recognition Conference, Las Vegas, NV, USA, 27–30 June 2016; pp. 4883–4892. [Google Scholar]
Figure 1. Examples of two different specimens of the same issue, namely of RIC 439 Aelius denarius.
Figure 1. Examples of two different specimens of the same issue, namely of RIC 439 Aelius denarius.
Jimaging 09 00107 g001
Figure 2. The architecture of a Siamese neural network comprises two mutually mirroring processing streams consisting of two identical neural networks with shared hyperparameters [25].
Figure 2. The architecture of a Siamese neural network comprises two mutually mirroring processing streams consisting of two identical neural networks with shared hyperparameters [25].
Jimaging 09 00107 g002
Figure 3. The architecture of Single Siamese ViT.
Figure 3. The architecture of Single Siamese ViT.
Jimaging 09 00107 g003
Figure 4. The architecture of Double Siamese ViT.
Figure 4. The architecture of Double Siamese ViT.
Jimaging 09 00107 g004
Figure 5. The data distribution of the train set.
Figure 5. The data distribution of the train set.
Jimaging 09 00107 g005
Figure 6. The flow chart for organizing data set.
Figure 6. The flow chart for organizing data set.
Jimaging 09 00107 g006
Figure 7. Performance characteristics of the proposed Single Siamese ViT on the obverse matching task.
Figure 7. Performance characteristics of the proposed Single Siamese ViT on the obverse matching task.
Jimaging 09 00107 g007
Figure 8. Performance characteristics of the proposed Single Siamese ViT on the reverse matching task.
Figure 8. Performance characteristics of the proposed Single Siamese ViT on the reverse matching task.
Jimaging 09 00107 g008
Figure 9. Training behavior of our Double Siamese ViT.
Figure 9. Training behavior of our Double Siamese ViT.
Jimaging 09 00107 g009
Figure 10. Summary of matching accuracy shown averaged over each issuing authority shown on the obverse.
Figure 10. Summary of matching accuracy shown averaged over each issuing authority shown on the obverse.
Jimaging 09 00107 g010
Figure 11. Examples of RIC 158 of Augustus. (a) An incomplete specimen of RIC 158 of Augustus. (b) A good condition specimen of RIC 158 of Augustus.
Figure 11. Examples of RIC 158 of Augustus. (a) An incomplete specimen of RIC 158 of Augustus. (b) A good condition specimen of RIC 158 of Augustus.
Jimaging 09 00107 g011
Figure 12. Examples of defective and complete specimens of Augustus in our data set. (a) An incomplete specimen of RIC 160 of Augustus. (b) A complete specimen of RIC 160 of Augustus.
Figure 12. Examples of defective and complete specimens of Augustus in our data set. (a) An incomplete specimen of RIC 160 of Augustus. (b) A complete specimen of RIC 160 of Augustus.
Jimaging 09 00107 g012
Figure 13. Examples of worn and discolored coins.
Figure 13. Examples of worn and discolored coins.
Jimaging 09 00107 g013
Figure 14. Examples of non-conforming data entries: (a) two specimens of Mariniana, also unusually shown reverse first then obverse, and (b) four diverse specimens, incorrectly matched as a whole with the issue corresponding to the specimen on the top left.
Figure 14. Examples of non-conforming data entries: (a) two specimens of Mariniana, also unusually shown reverse first then obverse, and (b) four diverse specimens, incorrectly matched as a whole with the issue corresponding to the specimen on the top left.
Jimaging 09 00107 g014
Figure 15. An example of two different issues which are virtually identical in their semantic content.
Figure 15. An example of two different issues which are virtually identical in their semantic content.
Jimaging 09 00107 g015
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guo, Z.; Arandjelović, O.; Reid, D.; Lei, Y.; Büttner, J. A Siamese Transformer Network for Zero-Shot Ancient Coin Classification. J. Imaging 2023, 9, 107. https://doi.org/10.3390/jimaging9060107

AMA Style

Guo Z, Arandjelović O, Reid D, Lei Y, Büttner J. A Siamese Transformer Network for Zero-Shot Ancient Coin Classification. Journal of Imaging. 2023; 9(6):107. https://doi.org/10.3390/jimaging9060107

Chicago/Turabian Style

Guo, Zhongliang, Ognjen Arandjelović, David Reid, Yaxiong Lei, and Jochen Büttner. 2023. "A Siamese Transformer Network for Zero-Shot Ancient Coin Classification" Journal of Imaging 9, no. 6: 107. https://doi.org/10.3390/jimaging9060107

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop