Author
Listed:
- Liyong Guo
- Erzam Marlisah
Abstract
This year has seen significant advancements in deep learning, and fine-grained few-shot image classification (FGFSIC) has also made substantial progress. FGFSIC faces two key challenges: high intra-class variance and low inter-class variance, which hinder accurate classification with limited data. Despite considerable efforts to extract more discriminative features using powerful networks, few studies have specifically addressed these challenges. This paper proposes a Spatial Multi-Scale Feature Transformer Network to overcome these issues. The approach first modifies the backbone network to extract multi-scale features, with classification results derived from comparing these multi-scale representations. Additionally, a Spatial Feature Transformer network is introduced to adjust the spatial positions of multi-scale features, which helps to reduce intra-class variance. Experiments were conducted on three widely used datasets—CUB-200-2011, Stanford Cars, and Stanford Dogs. The results demonstrate that both components of the proposed model significantly enhance FGFSIC performance, with final accuracies surpassing those of most existing methods. The findings emphasize the effectiveness of the proposed approach in tackling the critical issues of high intra-class variance and low inter-class variance, making it a promising solution for fine-grained image classification tasks, particularly in situations where only limited data is available. This work paves the way for improved performance in real-world applications requiring precise, few-shot learning in fine-grained domains.
Suggested Citation
Liyong Guo & Erzam Marlisah, 2025.
"Spatial multi-scale feature transformer network for fine-grained few-shot image classification,"
Review of Computer Engineering Research, Conscientia Beam, vol. 12(3), pages 195-205.
Handle:
RePEc:pkp:rocere:v:12:y:2025:i:3:p:195-205:id:4439
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:pkp:rocere:v:12:y:2025:i:3:p:195-205:id:4439. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Dim Michael (email available below). General contact details of provider: https://archive.conscientiabeam.com/index.php/76/ .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.