Author
Listed:
- Arseniy Mokin
(Smart Engines Service LLC, 117312 Moscow, Russia
The Faculty of Mechanics and Mathematics, Lomonosov Moscow State University, 119991 Moscow, Russia)
- Alexander Sheshkus
(Smart Engines Service LLC, 117312 Moscow, Russia
Federal Research Center “Computer Science and Control” of Russian Academy of Sciences, 119333 Moscow, Russia)
- Vladimir L. Arlazarov
(Smart Engines Service LLC, 117312 Moscow, Russia
Federal Research Center “Computer Science and Control” of Russian Academy of Sciences, 119333 Moscow, Russia)
Abstract
Training deep learning models for classification with limited data and computational resources remains a challenge when the number of classes is large. Metric learning offers an effective solution to this problem. However, it has its own shortcomings due to the known imperfections of widely used loss functions such as contrastive loss and triplet loss, as well as sample mining methods. This paper address these issues by proposing a novel mining method and metric loss function. Firstly, this paper presents an auto-probabilistic mining method designed to automatically select the most informative training samples for Siamese neural networks. Combined with a previously proposed auto-clustering technique, the method improves model training by optimizing the utilization of available data and reducing computational overhead. Secondly, this paper proposes the novel cluster-aware triplet-based metric loss function that addresses the limitations of contrastive and triplet loss, enhancing the overall training process. To evaluate the proposed methods, experiments were conducted with the optical character recognition task using the PHD08 and Omniglot datasets. The proposed loss function with the random-mining method achieved 82.6 % classification accuracy on the PHD08 dataset with full training on the Korean alphabet, surpassing the known baseline. The same experiment, using a reduced training alphabet, set a new baseline of 88.6 % on the PHD08 dataset. The application of the novel mining method further enhanced the accuracy to 90.6 % (+2.0%) and, combined with auto-clustering, achieved 92.3 % (+3.7%) compared with the new baseline. On the Omniglot dataset, the proposed mining method reached 92.32 % , rising to 93.17 % with auto-clustering. These findings highlight the potential effectiveness of the developed loss function and mining method in addressing a wide range of pattern recognition challenges.
Suggested Citation
Arseniy Mokin & Alexander Sheshkus & Vladimir L. Arlazarov, 2025.
"Auto-Probabilistic Mining Method for Siamese Neural Network Training,"
Mathematics, MDPI, vol. 13(8), pages 1-16, April.
Handle:
RePEc:gam:jmathe:v:13:y:2025:i:8:p:1270-:d:1633319
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jmathe:v:13:y:2025:i:8:p:1270-:d:1633319. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.